Publication | Open Access
An Acoustic Model Based on Kullback-Leibler Divergence for Posterior Features
42
Citations
10
References
2007
Year
Unknown Venue
EngineeringMachine LearningAcoustic ModelSpoken Language ProcessingAcoustic ModelingSpeech RecognitionData SciencePattern RecognitionHidden Markov ModelPhoneticsAudio AnalysisRobust Speech RecognitionVoice RecognitionAcoustic Signal ProcessingStatisticsHealth SciencesPosterior FeaturesComputer ScienceSignal ProcessingSpeech CommunicationPosterior ProbabilitiesSpeech ProcessingSpeech InputFinite State MachineSpeech Perception
This paper investigates the use of features based on posterior probabilities of subword units such as phonemes. These features are typically transformed when used as inputs for a hidden Markov model with mixture of Gaussians as emission distribution (HMM/GMM). In this work, we introduce a novel acoustic model that avoids the Gaussian assumption and directly uses posterior features without any transformation. This model is described by a finite state machine where each state is characterized by a target distribution and the cost function associated to each state is given by the Kullback-Leibler (KL) divergence between its target distribution and the posterior features. Furthermore, hybrid HMM/ANN system can be seen as a particular case of this KL-based model where state target distributions are predefined. A recursive training algorithm to estimate the state target distributions is also presented.
| Year | Citations | |
|---|---|---|
Page 1
Page 1