2020
DOI: 10.48550/arxiv.2005.08213
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Speech to Text Adaptation: Towards an Efficient Cross-Modal Distillation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(8 citation statements)
references
References 0 publications
0
8
0
Order By: Relevance
“…Lastly, while transcriptions are provided in ATIS and SNIPS, they are not normalized for ASR. Text normalization is applied with an open-source software 10 . For ATIS, utterances are ignored if they contain words with multiple slot labels [59].…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Lastly, while transcriptions are provided in ATIS and SNIPS, they are not normalized for ASR. Text normalization is applied with an open-source software 10 . For ATIS, utterances are ignored if they contain words with multiple slot labels [59].…”
Section: Methodsmentioning
confidence: 99%
“…Although there is much work on E2E speech enhancement [57], we found that merely augmenting the training data with a diverse set of environmental noises works well. We followed the noise augmentation protocol described in [50], where for each training sample, five noise files are randomly sampled and added to the clean file with SNR levels of [0, 10,20,30,40]dB, resulting in a five-fold data augmentation. Table 3 shows our proposed models trained with noise augmentation.…”
Section: Environmental Noise Augmentationmentioning
confidence: 99%
“…Several works use cross-modal distillation approach on SLU [13,14] to exploit textual knowledge. Cho et al [13] use knowledge distillation from a fine-tuned text BERT to an SLU model by making predicted logits for intent classification close to each other in fine-tuning.…”
Section: Knowledge Distillation For Slumentioning
confidence: 99%
“…Several works use cross-modal distillation approach on SLU [13,14] to exploit textual knowledge. Cho et al [13] use knowledge distillation from a fine-tuned text BERT to an SLU model by making predicted logits for intent classification close to each other in fine-tuning. Denisov and Vu [14] match an utterance embedding and a sentence embeddings of ASR pairs using knowledge distillation as a pre-training.…”
Section: Knowledge Distillation For Slumentioning
confidence: 99%
See 1 more Smart Citation