Findings of the Association for Computational Linguistics: ACL 2022 2022
DOI: 10.18653/v1/2022.findings-acl.160
|View full text |Cite
|
Sign up to set email alerts
|

Data Augmentation and Learned Layer Aggregation for Improved Multilingual Language Understanding in Dialogue

Abstract: Scaling dialogue systems to a multitude of domains, tasks and languages relies on costly and time-consuming data annotation for different domain-task-language configurations. The annotation efforts might be substantially reduced by the methods that generalise well in zero-and few-shot scenarios, and also effectively leverage external unannotated data sources (e.g., Web-scale corpora). We propose two methods to this aim, offering improved dialogue natural language understanding (NLU) across multiple languages: … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 79 publications
0
2
0
Order By: Relevance
“…The goal in Step 1 is to ensure high recall (i.e., to avoid too aggressive filtering), which is why we opt for the earlier stopping. As a baseline, we fine-tune XLM-R for the token classification task, as the standard SL task format (Xu et al, 2020;Razumovskaia et al, 2022b). Detailed training hyperparameters are provided in Appendix B.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The goal in Step 1 is to ensure high recall (i.e., to avoid too aggressive filtering), which is why we opt for the earlier stopping. As a baseline, we fine-tune XLM-R for the token classification task, as the standard SL task format (Xu et al, 2020;Razumovskaia et al, 2022b). Detailed training hyperparameters are provided in Appendix B.…”
Section: Methodsmentioning
confidence: 99%
“…The current approach to mitigate the issue is the standard cross-lingual transfer. The main 'transfer' assumption is that a suitable large English annotated dataset is always available for a particular task and domain: (i) the systems are then trained on the English data and then directly deployed to the target language (i.e., zero-shot transfer), or (ii) further adapted to the target language relying on a small set of target language examples (Xu et al, 2020;Razumovskaia et al, 2022b) which are combined with the large English dataset (i.e., few-shot transfer). However, this assumption might often be unrealistic in the context of TOD due to a large number of potential tasks and domains that should be supported by TOD systems .…”
Section: Introductionmentioning
confidence: 99%
“…Firstly, several automated approaches to reduce human effort were proposed, predominantly focusing on data augmentation. These approaches, previously applied to multilingual dialogue, include (i) word or span substitution, creating code-switched data between source and target languages Krishnan et al, 2021;Qin et al, 2021) or synonymous span substitution in a target language (Louvan & Magnini, 2020b)2020; and (ii) semi-supervised training with target language sentence retrieval (Razumovskaia et al, 2022).…”
Section: Outlook For Multilingual Tod Datasetsmentioning
confidence: 99%
“…Due to the lack of OOS annotations in openworld settings, previous research usually detects OOS samples indirectly such as resorting to inscope (INS) samples. Recently, data augmentation methods (Ng et al, 2020;Razumovskaia et al, 2022) have made it possible to detect OOS directly using a binary classifier.…”
Section: Introductionmentioning
confidence: 99%