Publication | Closed Access
An evaluation of using mutual information for selection of acoustic-features representation of phonemes for speech recognition
10
Citations
5
References
2002
Year
Unknown Venue
EngineeringBiometricsAcoustic Feature SpacePhonologySpeech RecognitionData SciencePattern RecognitionPhoneticsRobust Speech RecognitionBiostatisticsVoice RecognitionLanguage StudiesAcoustic-features RepresentationDistant Speech RecognitionSignal ProcessingSpeech CommunicationSpeech TechnologySpeech ProcessingSpeaker RecognitionSpeech InputMutual InformationSpeech PerceptionLinguisticsPhoneme Set
This paper addresses the problem of finding a subset of the acoustic feature space that best represents the phoneme set used in a speech recognition system. A maximum mutual information approach is presented for selecting acoustic features to be combined together to represent the distinctions among the phonemes. The overall phoneme recognition accuracy is slightly increased for the same length of feature vector for clean speech and at 10 dB compared to FFT-based Mel-frequency cepstrum coefficients (MFCC) by using acoustic features selected based on a maximum mutual information criterion. Using 16 different feature sets, the rank of the feature sets based on mutual information can predict phoneme recognition accuracy with a correlation coefficient of 0.71 compared to a correlation coefficient of 0.28 when using a criterion based on the average pair-wise Kullback-Liebler divergence to rank the feature sets.
| Year | Citations | |
|---|---|---|
Page 1
Page 1