2006 IEEE International Conference on Acoustics Speed and Signal Processing Proceedings
DOI: 10.1109/icassp.2006.1660011
|View full text |Cite
|
Sign up to set email alerts
|

Maximum Conditional Mutual Information Weighted Scoring for Speech Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 8 publications
0
4
0
Order By: Relevance
“…This could bring the benefit of using a class specific covariance model while avoiding the overfitting risk associated with it. A similar work using class specific covariance model is [3] In [4] it has been shown that the posterior form of Gaussian HMM can be represented as an HCRF model. For the case of a pooled covariance HMM this simplifies to a CRF or log-linear model.…”
Section: Introductionmentioning
confidence: 99%
“…This could bring the benefit of using a class specific covariance model while avoiding the overfitting risk associated with it. A similar work using class specific covariance model is [3] In [4] it has been shown that the posterior form of Gaussian HMM can be represented as an HCRF model. For the case of a pooled covariance HMM this simplifies to a CRF or log-linear model.…”
Section: Introductionmentioning
confidence: 99%
“…The parameters of the feature transformation matrix are calculated by LDA, but alternatively they could also be trained discriminatively. One such example is [4] where an iterative optimization is used to directly train a reduced dimension feature transformation matrix. Another one is [2]; here the transformation matrix is trained assuming unequal class covariances for Gaussian densities.…”
Section: Introductionmentioning
confidence: 99%
“…Early influential work along these lines involved data-driven methods for robust feature extraction [44] and filterbank design [45]- [47]. More recent methods include: 1) heteroscedastic linear discriminant analysis (HLDA) [48] and neighborhood component analysis [49] to learn informative low dimensional projections of high dimensional acoustic feature vectors; 2) stochastic gradient and second-order methods to tune parameters related to frequency warping and mel-scale filterbanks [50], [51]; 3) maximum-likelihood methods for speaker and environment adaptation [52], [53] that perform linear transformations of the acoustic feature space at test time; and 4) extensions of popular frameworks for discriminative training, such as minimum phone error [54] and maximum mutual information [55], to learn accuracy-improving transformations and projections of the acoustic feature space.…”
Section: Acoustic Feature Adaptationmentioning
confidence: 99%