ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2021
DOI: 10.1109/icassp39728.2021.9413475
|View full text |Cite
|
Sign up to set email alerts
|

Adaptable Multi-Domain Language Model for Transformer ASR

Abstract: We propose an adapter based multi-domain Transformer based language model (LM) for Transformer ASR. The model consists of a big size common LM and small size adapters. The model can perform multi-domain adaptation with only the small size adapters and its related layers. The proposed model can reuse the full finetuned LM which is fine-tuned using all layers of an original model. The proposed LM can be expanded to new domains by adding about 2% of parameters for a first domain and 13% parameters for after secon… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 21 publications
0
2
0
Order By: Relevance
“…Similar work was performed for other languages like Bengali, Japanese, etc. Also, more speech corpus is collected from the young people for many languages (Zeng et al, 2020;Lee et al, 2021). However, speaker fluctuation, environmental noise, and transmission channel noise all degrade ASR performance.…”
Section: Related Workmentioning
confidence: 99%
“…Similar work was performed for other languages like Bengali, Japanese, etc. Also, more speech corpus is collected from the young people for many languages (Zeng et al, 2020;Lee et al, 2021). However, speaker fluctuation, environmental noise, and transmission channel noise all degrade ASR performance.…”
Section: Related Workmentioning
confidence: 99%
“…Additionally, 13% WER is reduced by LSTM decoder (Zeng et al, 2021). Transformer model encoding and decoding can be carried with self-attention and multi-head attention layer (Lee et al, 2021). For CTC/Attention based End-To-End ASR, the transformer model is used, which result 23.66% of WER (Miao et al, 2020).…”
Section: Related Workmentioning
confidence: 99%