Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1045
|View full text |Cite
|
Sign up to set email alerts
|

Understanding Back-Translation at Scale

Abstract: An effective method to improve neural machine translation with monolingual data is to augment the parallel training corpus with back-translations of target language sentences. This work broadens the understanding of back-translation and investigates a number of methods to generate synthetic source sentences. We find that in all but resource poor settings back-translations obtained via sampling or noised beam outputs are most effective. Our analysis shows that sampling or noisy synthetic data gives a much stron… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

18
646
3
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 775 publications
(669 citation statements)
references
References 37 publications
18
646
3
2
Order By: Relevance
“…Back-translation, either generating translations using beam search (i.e., SEARCH) or using sampling (i.e., SAMPLE), does lead to significant improvements over using only the authentic bilingual corpus (i.e., NONE). We find that SAM-PLE is more effective than SEARCH, which confirms the finding of Edunov et al (2018). Using uncertainty-based confidence (i.e., "U") signifi-cantly improves over both SEARCH and SAMPLE on the combination of all test sets (p < 0.01).…”
Section: Resultssupporting
confidence: 81%
See 1 more Smart Citation
“…Back-translation, either generating translations using beam search (i.e., SEARCH) or using sampling (i.e., SAMPLE), does lead to significant improvements over using only the authentic bilingual corpus (i.e., NONE). We find that SAM-PLE is more effective than SEARCH, which confirms the finding of Edunov et al (2018). Using uncertainty-based confidence (i.e., "U") signifi-cantly improves over both SEARCH and SAMPLE on the combination of all test sets (p < 0.01).…”
Section: Resultssupporting
confidence: 81%
“…SEARCH: the translations of the monolingual corpus are generated by beam search (Sennrich et al, 2016a). SAMPLE: the translations of the monolingual corpus are generated by sampling (Edunov et al, 2018). "CE": confidence estimation method.…”
Section: Setupmentioning
confidence: 99%
“…According to the evaluation results reported in (Arase and Tsujii, 2017), the precision and recall of alignments are 83.6% and 78.9%, which are 89% and 92% of those of humans, respec- tively. Although alignment errors occur, previous studies show that neural networks are relatively robust against noise in a training corpus and still benefit from extra supervisions as demonstrated in (Edunov et al, 2018;Prabhumoye et al, 2018). We collect all the spans of phrases in a sentential paraphrase pair and their alignments as pairs of phrase spans.…”
Section: Phrase Alignment For Paraphrasesmentioning
confidence: 99%
“…For the purpose of the task, we extended the Marian toolkit with fp16 training, BERT-models (Devlin et al, 2018) and multi-task training. Similar to Edunov et al (2018) we use mixed-precision training with fp16, an optimizer delay of 16 before updating the gradients. We train on 8 Voltas with 16GB each.…”
Section: Model and Trainingmentioning
confidence: 99%
“…We mostly reproduce the results from Edunov et al (2018) and back-translate the entire German News-Crawl data with noisy back-translation. Similar to Edunov et al (2018)'s best method, we use output sampling as the noising approach. This has been implemented in Marian with the Gumbel softmax trick.…”
Section: Noisy Back-translationmentioning
confidence: 99%