2010 IEEE International Conference on Acoustics, Speech and Signal Processing 2010
DOI: 10.1109/icassp.2010.5495106
|View full text |Cite
|
Sign up to set email alerts
|

Discriminative training based on an integrated view of MPE and MMI in margin and error space

Abstract: Recent work has demonstrated that the Maximum Mutual Information (MMI) objective function is mathematically equivalent to a simple integral of recognition error, if the latter is expressed as a margin-based Minimum Phone Error (MPE) style error-weighted objective function. This led to the proposal of a general approach to discriminative training based on integrals of MPE-style loss, calculated using "differenced MMI" (dMMI), a finite difference of MMI functionals evaluated at the edges of a margin interval. Th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2012
2012
2024
2024

Publication Types

Select...
5
3

Relationship

3
5

Authors

Journals

citations
Cited by 34 publications
(18 citation statements)
references
References 13 publications
0
18
0
Order By: Relevance
“…The exper imental conditions for the MIT-OCW task are summarized in Ta ble 2. The initial acoustic model was constructed b y using variational Ba y esian triphone clustering [14] and differenced Maximum Mutual Information (dMMI) training [15]. The evaluation set consisted of 8 lectures (6,989 utterances, 72,159 words, and 7.8 hours).…”
Section: )mentioning
confidence: 99%
See 1 more Smart Citation
“…The exper imental conditions for the MIT-OCW task are summarized in Ta ble 2. The initial acoustic model was constructed b y using variational Ba y esian triphone clustering [14] and differenced Maximum Mutual Information (dMMI) training [15]. The evaluation set consisted of 8 lectures (6,989 utterances, 72,159 words, and 7.8 hours).…”
Section: )mentioning
confidence: 99%
“…The exper imental conditions for the CSJ task are summarized in Ta ble 3. The initial acoustic and language models were trained b y discriminative approaches [15,16]. We used CSJ testset 2 as a development set (10 lectures, 794 utterances, 26,798 words, and 2.2 hours) and CSJ test set I as an evaluation set (10 lectures, 977 utterances, 26,329 words, and 2.0 hours).…”
Section: )mentioning
confidence: 99%
“…BMMI modifies the MMI criterion by incorporating margins into the denominator (corresponding to the competitor contribution) of the MMI objective function and a boosting factor, further called margin parameter. Recently, a new discriminative criterion called differenced MMI (dMMI) was proposed to generalize MPE and BMMI [10]. The objective function of dMMI is defined as the difference between two BMMI objective functions with two different margin parameters therefore combining the regularization benefits of BMMI with a loose definition of references [9,11].…”
Section: Introductionmentioning
confidence: 99%
“…p λ is the likelihood of an acoustic model, κ is the acoustic scale, and pL is the likelihood of a language model. Utterances that contain many errors need to be considered intensively and evaluating phoneme accuracies (i.e., evaluating margins [6]) improves performance. In the boosted MMI (bMMI) [22], the standard MMI objective function is modified to include a term that "boosts" the effect of hypotheses with low phoneme accuracy:…”
Section: Discriminative Trainingmentioning
confidence: 99%
“…Over the past ten years model training techniques have migrated from Maximum Likelihood (ML) estimation to discriminative training [2,3,4,5,6]. In addition, various types of feature transformations have been proposed and showed effectiveness [7,8,9,10,11,12].…”
Section: Introductionmentioning
confidence: 99%