2021
DOI: 10.48550/arxiv.2104.03006
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Librispeech Transducer Model with Internal Language Model Prior Correction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 0 publications
0
4
0
Order By: Relevance
“…Considering the variability of different pronunciations, the window size should be different for different query vector concerns; thus, a parameterized window size calculation method is used rather than a fixed window size. The calculation method is shown in Equation (9).…”
Section: 𝑙 𝑖 = 𝐼 • 𝜎(𝑼 𝑇 𝑔(𝑾(𝑬 𝑥 𝑖 + 𝒖 + 𝒗)))mentioning
confidence: 99%
See 2 more Smart Citations
“…Considering the variability of different pronunciations, the window size should be different for different query vector concerns; thus, a parameterized window size calculation method is used rather than a fixed window size. The calculation method is shown in Equation (9).…”
Section: 𝑙 𝑖 = 𝐼 • 𝜎(𝑼 𝑇 𝑔(𝑾(𝑬 𝑥 𝑖 + 𝒖 + 𝒗)))mentioning
confidence: 99%
“…LAS+Specaugment [27] is a sequence to sequence framework with Specaugment. LSTM Transducer [9] improves external language model and an estimated internal LM. The hybrid model with Transformer rescoring [30] leverages the Transformer to improve hybrid acoustic modeling.…”
Section: Comparison Experimentsmentioning
confidence: 99%
See 1 more Smart Citation
“…In shallow fusion [7,8], a log-linear interpolation between the E2E model score and the LM score is computed at each step of the beam search. To improve shallow fusion, internal LM estimation-based fusion [9,10,11,12,13,14] was proposed to estimate an internal LM (ILM) score and subtract it from the shallow fusion score. However, all these methods require an external LM during inference, increasing decoding time and computational cost.…”
Section: Introductionmentioning
confidence: 99%