ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2022
DOI: 10.1109/icassp43922.2022.9746579
|View full text |Cite
|
Sign up to set email alerts
|

Consistent Training and Decoding for End-to-End Speech Recognition Using Lattice-Free MMI

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(2 citation statements)
references
References 18 publications
0
2
0
Order By: Relevance
“…Table 1 compares the (CER) result of our model on the Aishell-1 test dataset with a few public models include: Espnet [33], WeNet [24], K2 [34] and Neural Transducer+LFMMI [35]. The first three models are all AED model structures, and the last is NT based.…”
Section: Results Of Aishell-1mentioning
confidence: 99%
“…Table 1 compares the (CER) result of our model on the Aishell-1 test dataset with a few public models include: Espnet [33], WeNet [24], K2 [34] and Neural Transducer+LFMMI [35]. The first three models are all AED model structures, and the last is NT based.…”
Section: Results Of Aishell-1mentioning
confidence: 99%
“…We use the same Dev and Test dataset from AISHELL-1 as performance evaluation for private data. Inspecting previous Mandarin speech recognition results [34], RNN-Transducer from ESPNet [35] backbone appears to be a top ASR candidate and is used in our experiments. We follow the benchmark setup to build up our Mandarin ASR with a Conformer encoder, a Transformer decoder, and an LSTM prediction network with 135M trainable model parameters.…”
Section: Continuous Speech Recognition and Resultsmentioning
confidence: 99%