Proceedings of the 2019 Conference of the North 2019
DOI: 10.18653/v1/n19-1120
|View full text |Cite
|
Sign up to set email alerts
|

Extract and Edit: An Alternative to Back-Translation for Unsupervised Neural Machine Translation

Abstract: The overreliance on large parallel corpora significantly limits the applicability of machine translation systems to the majority of language pairs. Back-translation has been dominantly used in previous approaches for unsupervised neural machine translation, where pseudo sentence pairs are generated to train the models with a reconstruction loss. However, the pseudo sentences are usually of low quality as translation errors accumulate during training. To avoid this fundamental issue, we propose an alternative b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
27
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 33 publications
(27 citation statements)
references
References 30 publications
0
27
0
Order By: Relevance
“…Our method is compared with several previous unsupervised NMT systems (Artetxe et al, 2018b;4 https://github.com/artetxem/vecmap Lample et al, 2018a,b;Yang et al, 2018;Wu et al, 2019;Song et al, 2019). Although, Song et al (2019) have achieved comparable results with supervised NMT systems with larger monolingual data (Wikipedia data) and bigger model 5 , we still list the results that obtained with the same data and model as ours for fair comparison.…”
Section: Overall Resultsmentioning
confidence: 99%
“…Our method is compared with several previous unsupervised NMT systems (Artetxe et al, 2018b;4 https://github.com/artetxem/vecmap Lample et al, 2018a,b;Yang et al, 2018;Wu et al, 2019;Song et al, 2019). Although, Song et al (2019) have achieved comparable results with supervised NMT systems with larger monolingual data (Wikipedia data) and bigger model 5 , we still list the results that obtained with the same data and model as ours for fair comparison.…”
Section: Overall Resultsmentioning
confidence: 99%
“…But they use much more monolingual data from Wikipedia than previous work and this paper. We must also mention the work of Wu et al (2019). They similarly use retrieval and rewriting framework for unsupervised MT.…”
Section: Related Workmentioning
confidence: 99%
“…Artetxe et al (2018a) combined back-translation and MERT (Och, 2003) to iteratively refine the SMT model. Wu et al (2019) proposed to discard back-translation. Instead, they extracted and edited the nearest sentences in the target language to construct pseudo-parallel data, which was used as a supervision signal.…”
Section: Unsupervised Machine Translationmentioning
confidence: 99%
“…Building an effective model on low resource data or even in an unsupervised way is always an interesting and challenging research topic (Gu et al, 2018;Radford et al, 2016;Lee et al, 2019). Recently, unsupervised MT (Artetxe et al, 2018b,a;Conneau et al, 2018;Lample et al, 2018b;Wu et al, 2019), which can immensely reduce the reliance on parallel corpora, has been gaining more and more interest.…”
Section: Introductionmentioning
confidence: 99%