ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2022
DOI: 10.1109/icassp43922.2022.9747887
|View full text |Cite
|
Sign up to set email alerts
|

Improving CTC-Based Speech Recognition Via Knowledge Transferring from Pre-Trained Language Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 14 publications
(8 citation statements)
references
References 22 publications
0
8
0
Order By: Relevance
“…However, these models often slow down the decoding speed and usually have a large set of model parameters. On the other hand, a school of research makes the ASR model to learn linguistic information from pre-trained language models in a teacher-student training manner [20,21,7,22,23]. These models still obtain a fast decoding speed, but their improvements are usually incremental.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…However, these models often slow down the decoding speed and usually have a large set of model parameters. On the other hand, a school of research makes the ASR model to learn linguistic information from pre-trained language models in a teacher-student training manner [20,21,7,22,23]. These models still obtain a fast decoding speed, but their improvements are usually incremental.…”
Section: Related Workmentioning
confidence: 99%
“…where the objective function L KT is defined to minimize the cosine embedding loss, and a scaling hyper-parameter k is used to equalize the numerical imbalance between the cosine embedding loss and other losses [7]. The index 0 and N + 1 denote the positions of the special tokens, which are ignored in calculating the training loss.…”
Section: Token-dependent Knowledge Transferring Modulementioning
confidence: 99%
See 3 more Smart Citations