Abstract
The development of an audiovisual pronunciation teaching and training method and software system is discussed in this article. The method is designed to help children with speech and hearing disorders gain better control over their speech production. The teaching method is drawn up for progression from individual sound preparation to practice of sounds in sentences for four languages: English, Swedish, Slovenian, and Hungarian. The system is a general language-independent measuring tool and database editor. This database editor makes it possible to construct modules for all participant languages and for different sound groups. Two modules are under development for the system in all languages: one for teaching and training vowels to hearing-impaired children and the other for correction of misarticulated fricative sounds. In the article we present the measuring methods, the used distance score calculations of the visualized speech spectra, and problems in the evaluation of the new multimedia tool.
Similar content being viewed by others
References
CORDIS Community R&D Information Services. (1997). ECSC-EC-EAEC, Luxembourg, Organization of EC for publishing information on R&D in Europe. Media is CD, printout and Internet.
Daniloff, R.G., Wilkox, K., and Stephens, M.I. (1980). An acoustic-articulatory description of children’s defective (s) productions. Journal of Communication Disorders, 13:347–363.
Itakura, F. (1975). Minimum prediction residual principle applied to speech recognition. IEEE Transactions on Acoustics, Speech and Signal Processing, ASSP-23:67–72.
Kent, R.D. and Read, C. (1992). The Acoustic Analysis of Speech, San Diego, CA, Singular Publishing.
Lee, S., Potamianos, A., and Narayanan, S. (1999). Acoustics of children’s speech: Developmental changes of temporal and spectral parameters. JASA, 3:1455–1468.
Povel, D.J. (1991). The visual speech apparatus: Theoretical and practical aspects. Speech Communication, 10:59–80.
Rabiner, L.R. and Levinson, S.E. (1981). Isolated and connected word recognition—theory and selected application. IEEE Trans on Communication, 29:621–659.
Sakoe, H. and Chiba, S. (1978). Dynamic programming algorithm optimatization for spoken word recognition. IEEE Transaction on Acoustics; Speech and Signal Processing, ASSP-26:43–49.
Vicsi, K. (1981). The most relevant acoustical microsegment and its duration necessary for the recognition of unvoiced stops. Acoustica, 48:53–58.
Vicsi, K., Matilla, M., and Berényi, P. (1990). Continuous speech segmentation using different methods. Acoustica, 71:152–156.
Vicsi, K. and Vig, A. (1998). LIAS: Language independent automatic segmentation technique using Sampa labeling of phonemes. First International Conference on Language Resources & Education, Granada, Spain, 1998. 1.317.
Wallace, J.L. et al. (1998). Applications of speech recognition in the primary school classroom ESCA. In Still ’98 Workshop Proceedings, Marholmen, pp. 21–24.
Zwicker, E. (1982). Psychoakustik. Berlin, Springer-Verlag.
Zwicker, E. and Terhardt, E. (1980). Analytical expressions for band rate and critical bandwidth as a function of frequency. J. Soc. Am., 68:1523.
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Vicsi, K., Roach, P., Öster, A. et al. A Multimedia, Multilingual Teaching and Training System for Children with Speech Disorders. International Journal of Speech Technology 3, 289–300 (2000). https://doi.org/10.1023/A:1026563015923
Issue Date:
DOI: https://doi.org/10.1023/A:1026563015923