Publication | Closed Access
Neural network models of sensory integration for improved vowel recognition
79
Citations
40
References
1990
Year
Speech KinematicsNeurolinguisticsSpeech ArticulationAcoustic ModelingSpeech RecognitionSensory IntegrationMultiple SourcesPattern RecognitionPhoneticsRobust Speech RecognitionVoice RecognitionLanguage StudiesAcoustic AnalysisSpeech Signal AnalysisHealth SciencesNeural NetworksSpeech AcquisitionSpeech CommunicationSpeech TechnologySpeech AcousticsLanguage RecognitionSpeech ProcessingSpeech InputAuditory ComputationSpeech PerceptionSpeech InformationLinguisticsSpeaker Recognition
It is demonstrated that multiple sources of speech information can be integrated at a subsymbolic level to improve vowel recognition. Feedforward and recurrent neural networks are trained to estimate the acoustic characteristics of a vocal tract from images of the speaker's mouth. These estimates are then combined with the noise-degraded acoustic information, effectively increasing the signal-to-noise ratio and improving the recognition of these noise-degraded signals. Alternative symbolic strategies such as direct categorization of the visual signals into vowels are also presented. The performances of these neural networks compare favorably with human performance and with other pattern-matching and estimation techniques.< <ETX xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">></ETX>
| Year | Citations | |
|---|---|---|
Page 1
Page 1