ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2021
DOI: 10.1109/icassp39728.2021.9414518
|View full text |Cite
|
Sign up to set email alerts
|

Making Punctuation Restoration Robust and Fast with Multi-Task Learning and Knowledge Distillation

Abstract: In punctuation restoration, we try to recover the missing punctuation from automatic speech recognition output to improve understandability. Currently, large pre-trained transformers such as BERT set the benchmark on this task but there are two main drawbacks to these models. First, the pre-training data does not match the output data from speech recognition that contains errors. Second, the large number of model parameters increases inference time. To address the former, we use a multi-task learning framework… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 11 publications
(2 citation statements)
references
References 15 publications
0
2
0
Order By: Relevance
“…Currently, knowledge distillation, 23 a popular technology for model enhancement, has been widely used in computer vision, natural language processing, and automatic speech recognition 24,25 . Inspired by its broad applicability, this paper incorporates knowledge distillation into GCN‐based recommendation models to alleviate the above limitations.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Currently, knowledge distillation, 23 a popular technology for model enhancement, has been widely used in computer vision, natural language processing, and automatic speech recognition 24,25 . Inspired by its broad applicability, this paper incorporates knowledge distillation into GCN‐based recommendation models to alleviate the above limitations.…”
Section: Introductionmentioning
confidence: 99%
“…Currently, knowledge distillation, 23 a popular technology for model enhancement, has been widely used in computer vision, natural language processing, and automatic speech recognition. 24,25 Inspired by its broad applicability, this paper incorporates knowledge distillation into GCN-based recommendation models to alleviate the above limitations. The proposed idea is different from existing work as we propose a two-phase knowledge distillation model (TKDM) to improve the effectiveness of GCN-based recommendations.…”
Section: Introductionmentioning
confidence: 99%