ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019
DOI: 10.1109/icassp.2019.8682478
|View full text |Cite
|
Sign up to set email alerts
|

Exploring Retraining-free Speech Recognition for Intra-sentential Code-switching

Abstract: In this paper, we present our initial efforts for building a codeswitching (CS) speech recognition system leveraging existing acoustic models (AMs) and language models (LMs), i.e., no training required, and specifically targeting intra-sentential switching. To achieve such an ambitious goal, new mechanisms for foreign pronunciation generation and language model (LM) enrichment have been devised. Specifically, we have designed an automatic approach to obtain high quality pronunciation of foreign language (FL) w… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 20 publications
0
2
0
Order By: Relevance
“…They looked at phone-merging techniques to handle the two languages in acoustic modeling, explored further in [9,31], and generating codeswitched text data for language modeling, studied more in [32,39]. Since then, different approaches have been applied to improve codeswitched speech recognition like speech chains [40], transliteration [41], and translation [42]. Authors in [14,27,43] focus on tracking the language switch points, similar to our LID aware training.…”
Section: Relation To Prior Workmentioning
confidence: 99%
“…They looked at phone-merging techniques to handle the two languages in acoustic modeling, explored further in [9,31], and generating codeswitched text data for language modeling, studied more in [32,39]. Since then, different approaches have been applied to improve codeswitched speech recognition like speech chains [40], transliteration [41], and translation [42]. Authors in [14,27,43] focus on tracking the language switch points, similar to our LID aware training.…”
Section: Relation To Prior Workmentioning
confidence: 99%
“…We also apply the multi-graph decoding strategy [17], which contains code-switch and English 1 and the code-switch dataset as Table 2, and the English is trained by the transcript of librispeech. Besides, we implement the methods that optimizes the NGram language model in [7] and [18] as the baseline of the first-pass decoding, and these two methods together achieve 7% relative WER reduction. The number of NBEST is set to 128.…”
Section: Setupmentioning
confidence: 99%