Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.669
|View full text |Cite
|
Sign up to set email alerts
|

Rethinking Data Augmentation for Low-Resource Neural Machine Translation: A Multi-Task Learning Approach

Abstract: In the context of neural machine translation, data augmentation (DA) techniques may be used for generating additional training samples when the available parallel data are scarce. Many DA approaches aim at expanding the support of the empirical data distribution by generating new sentence pairs that contain infrequent words, thus making it closer to the true data distribution of parallel sentences. In this paper, we propose to follow a completely different approach and present a multi-task DA approach in which… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 14 publications
(12 citation statements)
references
References 36 publications
1
7
0
Order By: Relevance
“…Our framework -which extends preliminary work 1 reported in a conference paper by the same authors [13]does not require elaborate preprocessing steps, training additional systems, or data besides the available training parallel corpora. Experiments with ten low-resource translation tasks show that it systematically outperforms state-of-the art methods aimed at extending the support of the empirical data distribution.…”
Section: Introductionmentioning
confidence: 61%
See 3 more Smart Citations
“…Our framework -which extends preliminary work 1 reported in a conference paper by the same authors [13]does not require elaborate preprocessing steps, training additional systems, or data besides the available training parallel corpora. Experiments with ten low-resource translation tasks show that it systematically outperforms state-of-the art methods aimed at extending the support of the empirical data distribution.…”
Section: Introductionmentioning
confidence: 61%
“…4. This differs from previous work[13] in which transformations were applied to the original training samples during the pre-processing of the corpus, and therefore before training.…”
mentioning
confidence: 87%
See 2 more Smart Citations
“…The highest BLEU of 45.98 was obtained for word-level tokenization, showing that selecting the right pre-processing methods may improve the performance of the MT models. [25] created an NMT between English and Swahili using news data. Authors obtained BLEU of 27.42 for English-to-Swahili translation when using the data that included part-of-speech tags outperforming Google translate.…”
Section: Related Workmentioning
confidence: 99%