A computational model of sound recognition used to analyze the capacity and adaptability in learning vowel classes
Abstract
Sound recognition is likely to initiate early in auditory processing and use stored representations (spectrotemporal templates) to compare against spectral information from auditory brainstem responses over time. A computational model of sound recognition is developed using neurobiologically plausible operations. The adaptability and number of templates required for the computational model to correctly recognize 10 Klatt-synthesized vowels is determined to be around 1250 templates when trained with random fundamental frequencies from the male pitch range and randomized variation of the first three formants of each vowel. To investigate the ability to adapt to noise and other unheard vowel utterances, test sets with 1000 randomly generated Klatt vowels in babble at signal-to-noise ratios (SNRs) of 20 dB, 10 dB, 5 dB, 0 dB, and ????5 dB are generated. The vowel recognition rates at each SNR are 99.7%, 99.6%, 97.0%, 77.6%, and 54.0%, respectively. Also, a test set of four vowel recordings from four speakers is tested with no noise, giving 100% recognition rate. These data suggest that storage of auditory representations for speech at the spectrotemporal resolution of the auditory nerve over a typical range of spoken pitch does not require excessive memory resources or computing to implement on parallel computer systems.
References
Deng, L. and Strik, H. (2007). “Structure-based and template-based automatic speech recognition – Comparing parametric and non-parametric approaches”, in Interspeech 2007, pp. 2608-2611.
Fairbanks, G., and Grubb, P. (1961). “A psychophysical investigation of vowel formants”, J. Speech Hear. Res., 4, 203-219.
Hawks, J.W., and Miller, J.D. (1995). “A formant bandwidth estimation procedure for vowel synthesis”, J. Acoust. Soc. Am., 97, 1343-1344.
Hillenbrand, J., and Gayvert, R.T. (1993). “Identification of steady-state vowels synthesized from the Peterson and Barney measurements”, J. Acoust. Soc. Am., 94, 668-674.
Kalinli, O., Seltzer, M.L., Droppo, J., and Acero, A. (2010). “Noise adaptive training for robust automatic speech recognition”, IEEE T. Audio Speech, 18, 1889-1901.
Klatt, D.H. (1980). “Software for a cascade/parallel formant synthesizer”, J. Acoust. Soc. Am., 67, 971-995.
Klatt, D.H., and Klatt, L.C. (1990). “Analysis, synthesis, and perception of voice quality variations among female and male talkers”, J. Acoust. Soc. Am., 87, 820-857.
Krishnamoorthy, K., and Mathew, T. (2009). “The multivariate normal distribution”, in Statistical Tolerance Regions: Theory, Applications, and Computation (John Wiley & Sons, Inc.), pp. 225-247.
McLachlan, N. (2009). “A vomputational model of human pitch strength and height judgments.”, Hear. Res., 249, 23-35.
McLachlan, N., and Wilson, S. (2010). “The central role of recognition in auditory perception: a neurobiological model”, Psychol. Rev., 117, 175-196.
McLachlan, N. (2011). “A neurocognitive model of recognition and pitch segregation”, J. Acoust. Soc. Am., 130, 2845-2854.
Mi, L., Tao, S., Wang, W., Dong, Q., Jin, S.-H., and Liu, C. (2013). “English vowel identification in long-term speech-shaped noise and multi-talker babble for English and Chinese listeners”, J. Acoust. Soc. Am., 133, EL391-EL397.
Moore, B.C.J. (2003). An Introduction to the Psychology of Hearing, 3rd Ed. (Academic Press).
Neel, A.T. (2008). “Vowel space characteristics and vowel identification accuracy”, J. Speech Lang. Hear. Res., 51, 574-585.
Oliphant, T.E. (2007). “Python for scientific computing”, Comput. Sci. Eng., 9, 10-20.
Pearce, D., and Hirsch, H.-G. (2000). “The Aurora experimental framework for the performance evaluation of speech recognition systems under noisy conditions”, in ISCA ITRW ASR-2000 (Paris, France), pp. 181-188.
Peterson, G.E., and Barney, H.L. (1952). “Control methods used in a study of the vowels”, J. Acoust. Soc. Am., 24, 175-184.
Rabiner, L.R. (1989). “A tutorial on hidden Markov models and selected applications in speech recognition”, P. IEEE, 77, 257-286.
Schouten, B., Gerrits, E., and Van Hessen, A. (2003). “The end of categorical perception as we know it”, Speech Commun., 41, 71-80.
Slaney, M. (1993). “An efficient implementation of the Patterson-Holdsworth auditory filter bank”, Apple Computer Technical Report #35.
Viemeister, N.F., and Wakefield, G.H. (1991). “Temporal integration and multiple looks”, J. Acoust. Soc. Am., 90, 858-865.
Downloads
Published
How to Cite
Issue
Section
License
Authors who publish with this journal agree to the following terms:
a. Authors retain copyright* and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
b. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
c. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See The Effect of Open Access).
*From the 2017 issue onward. The Danavox Jubilee Foundation owns the copyright of all articles published in the 1969-2015 issues. However, authors are still allowed to share the work with an acknowledgement of the work's authorship and initial publication in this journal.