Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.656
|View full text |Cite
|
Sign up to set email alerts
|

Code-switched inspired losses for spoken dialog representations

Abstract: Spoken dialog systems need to be able to handle both multiple languages and multilinguality inside a conversation (e.g in case of codeswitching). In this work, we introduce new pretraining losses tailored to learn multilingual spoken dialog representations. The goal of these losses is to expose the model to codeswitched language. To scale up training, we automatically build a pretraining corpus composed of multilingual conversations in five different languages (French, Italian, English, German and Spanish) fro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3

Relationship

2
1

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 65 publications
0
3
0
Order By: Relevance
“…For future work, we plan to study OOD in sequence labelling tasks (Witon* et al, 2018;Colombo* et al, 2020;Chapuis* et al, 2020a;Colombo et al, 2021a), sequence generation (Colombo* et al, 2019;Jalalzai* et al, 2020;Modi et al, 2020;Colombo et al, 2021e) and fair classification (Colombo et al, 2021d;Pichler et al, 2022) and multimodal scenario (Garcia* et al, 2019;Dinkar* et al, 2020) as well as automatic evaluation (Colombo et al, 2021c;Colombo, 2021a;Staerman et al, 2021b).…”
Section: G Futures Applicationsmentioning
confidence: 99%
“…For future work, we plan to study OOD in sequence labelling tasks (Witon* et al, 2018;Colombo* et al, 2020;Chapuis* et al, 2020a;Colombo et al, 2021a), sequence generation (Colombo* et al, 2019;Jalalzai* et al, 2020;Modi et al, 2020;Colombo et al, 2021e) and fair classification (Colombo et al, 2021d;Pichler et al, 2022) and multimodal scenario (Garcia* et al, 2019;Dinkar* et al, 2020) as well as automatic evaluation (Colombo et al, 2021c;Colombo, 2021a;Staerman et al, 2021b).…”
Section: G Futures Applicationsmentioning
confidence: 99%
“…Update ϕ, ψ using (1). As future work we plan to disentangled more complex labels such as dialog acts (Colombo et al, , 2021a, emotions (Witon et al, 2018) and linguistic phenomena such as disfluencies (Dinkar et al, 2020) and other spoken language phe-…”
Section: D4 Related Work General Algorithmmentioning
confidence: 99%
“…Update φ, ψ using (1). As future work we plan to disentangled more complex labels such as dialog acts (Colombo et al, , 2021a, emotions (Witon et al, 2018) and linguistic phenomena such as disfluencies (Dinkar et al, 2020) and other spoken language phenomenon . Future research also include extending these losses to data augmentation (Dhole et al, 2021;Colombo et al, 2021e) and sentence generation (Colombo et al, 2021c,f) and study the trade-off using rankings (Colombo et al, 2022) or anomaly detection (Staerman et al, 2019(Staerman et al, , 2020(Staerman et al, , 2021b(Staerman et al, , 2022.…”
Section: D4 Related Work General Algorithmmentioning
confidence: 99%