My main research interest is Computational Auditory Scene Analysis (CASA), which aims to build
machine systems that mimic the ability of human listeners to segregate complex mixtures of sound.
I also have interests in reverberation-robust automatic speech recognition, hearing impairment and music technology.
The following book gives an overview of the field:
DeLiang Wang and Guy J. Brown (editors), Computational auditory scene analysis: Principles, Algorithms, and Applications. IEEE Press/Wiley-Interscience, 2006
You can get more information from amazon.co.uk or amazon.com, or go to the web site that accompanies the book.
Selected recent publications:
- T. Jurgens, T. Brand, N. Clark, R. Meddis and G. J. Brown (2013) The robustness of speech representations
obtained from simulated auditory nerve fibers under different noise conditions. Journal of the Acoustical
Society of America, 134 (3), EL232-238.
- E. Kurtic, G. J. Brown and B. Wells (2013) Resources for turn competition in overlapping talk. Speech Communication, 55, pp. 721-743.
- S. Keronen, H. Kallasjoki, U. Remes, G. J. Brown, J. F. Gemmeke and K. J. Palomäki (2013) Mask estimation and imputation methods for missing data speech recognition in a multisource reverberant environments. Computer Speech and Language, 27 (3), pp. 798-819.
- N. R. Clark, G. J. Brown, T. Jurgens and R. Meddis (2012) A frequency-selective feedback model of auditory efferent suppression and its implications for the recognition of speech in noise. Journal of the Acoustical Society of America, in press.
- J. Gorisch, B. Wells and G. J. Brown (2012) Pitch contour matching and interactional alignment across turns: An acoustic investigation. Language and Speech, 55 (1), pp. 57-76.
- G. J. Brown, A. V. Beeston and K. J. Palomäki (2012) Perceptual compensation for the effects of reverberation on consonant identification: A comparison of human and machine performance. Proceedings of Interspeech-2012, in press.
- K. J. Palomaki and G. J. Brown (2011) A computational model of binaural speech recognition: role of across-frequency vs. within-frequency processing and internal noise. Speech Communication 53, pp.924-940 [from ScienceDirect]
- G. J. Brown, R. F. Ferry and R. Meddis (2010) A computer model of auditory efferent suppression: Implications for the coding of speech in noise. Journal of the Acoustical Society of America, 127(2), pp. 943--954 [available from ASA digital library]
- S. N. Wrigley, S. Tucker, G. J. Brown and S. Whittaker (2008) The influence of audio presentation style on multitasking during teleconferences. Accepted for INTERSPEECH-2008 [PDF]
- K. J. Palomäki and G. J. Brown (2008) A computational model of binaural speech intelligibility level difference (Abstract). Journal of the Acoustical Society of America 123 (5), 3715 [available from ASA digital library]
- S. N. Wrigley, S. Tucker, G. J. Brown and S. Whittaker (2008) Effect of sound spatialisation on multitasking in remote meetings (Abstract). Journal of the Acoustical Society of America 123 (5), 3861 [available from ASA digital library]
- G. J. Brown, R. W. Mill and S. Tucker (2008) Auditory-motivated techniques for detection and classification of passive sonar signals (Abstract).Journal of the Acoustical Society of America 123 (5), 3344 [available from ASA digital library]
- R. Meddis, R. Ferry and G. J. Brown (2008) Speech in noise and the medial olivo-cochlear efferent system (Abstract).Journal of the Acoustical Society of America 123 (5), 3051 [available from ASA digital library]
- G. J. Brown and K. J. Palomäki (2008) A reverberation-robust automatic speech recognition system based on temporal masking (Abstract).Journal of the Acoustical Society of America 123 (5), 2978 [available from ASA digital library]
Selected older articles:
Back to top
- S. Harding, J. Barker and G. J. Brown (2006) Mask estimation for missing data speech recognition based on statistics of binaural interaction. IEEE Transactions on Speech and Audio Processing, 14 (1), pp. 58-67 [PDF]
- G. J. Brown and K. J. Palomäki (2005) A computational model of the speech reception threshold for laterally separated speech and noise. Proceedings of Interspeech 2005, Lisbon, 4th-8th September, pp. 1753-1756 [PDF]
- G. J. Brown and D. L. Wang (2005) Separation of speech by computational auditory scene analysis. In Speech enhancement: What's new?, edited by J. Benesty, S. Makino and J. Chen. New York: Springer
- S. Tucker and G. J. Brown (2005) Classification of transient sonar sounds using perceptually motivated features. IEEE Journal of Oceanic Engineering, 30 (3), pp. 588-600 [PDF]
- S. N. Wrigley, G. J. Brown, S. Renals and V. Wan (2005) Speech and crosstalk detection in multi-channel audio. IEEE Transactions on Speech and Audio Processing, 13 (1), 84-91 [PDF]
- K. J. Palomäki, G. J. Brown and D. L. Wang (2004) A binaural processor for missing data speech recognition in the presence of noise and small-room reverberation. Speech Communication, 43 (4), pp. 361-378 [PDF]
- S. N. Wrigley and G. J. Brown (2004) A computational model of auditory selective attention. IEEE Transactions on Neural Networks, 15 (5), pp. 1151-1163 [PDF]
- K. J. Palomäki, G. J. Brown and J. Barker (2004) Techniques for handling convolutional distortion with 'missing data' automatic speech recognition. Speech Communication, 43 (1-2), pp. 123-142 [PDF]
Two funded research projects are current:
- TWO!EARS (Funded by EU-FP7)
- Acoustic analysis of snoring and sleep apnoea (KTP with Passion for Life Healthcare/Innovate UK)
Recently completed project:
Back to top
Some links to online resources related to my research:
Back to top
Modules taught in the 2014-2015 session (only accessible from within the campus). Links are currently being updated for the new session:
Back to top
Current PhD students:
Previous PhD students:
- Tim Metcalf (with Renee Timmers, Music) - perception of emotion in music and speech for cochlear-implanted listeners
- Najwa Alghamdi (with Steve Maddock and Jon Barker) - audio-visual speech training for cochlear-implanted listeners
- Tom White (with Gordon Fraser) - software testing for natural user interfaces
Other collaborators and alumni:
- Amy Beeston - computational models of perceptual compesation for room reverberation
- Jana Eggink (BBC)
- Matthew Robertson (BAE Systems Detica)
- Darryl Godsmark (Intuitive Systems Ltd)
- Jan Gorisch (Laboratoire Parole et Langage, Aix-en-Provence)
- Brian L. Karlsen (Oticon)
- Emina Kurtic (University of Sheffield)
- Simon Makin (Department of Psychology, University of Reading)
- Robert Mill (Institute for Hearing Research, Nottingham)
- Simon Tucker (Department of Computer Science, University of Sheffield)
- Stuart Wrigley (Department of Computer Science, University of Sheffield)
Back to top
- Kalle J. Palomäki (Laboratory of Computer and Information Science, Aalto University)
- Eric Mousset (School of Electrical and Information Engineering, University of Sydney)
I am Deputy Head of Department for the session 2014-2015.
Back to top
Professor Guy J. Brown
Speech and Hearing Research Group
Department of Computer Science
University of Sheffield
211 Portobello Street
Sheffield S1 4DP
Tel: +44 (0)114 222 1821
Fax: +44 (0)114 222 1810
Last modified on 28th August 2014