2020
DOI: 10.48550/arxiv.2004.10581
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

When and Why is Unsupervised Neural Machine Translation Useless?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
8
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(9 citation statements)
references
References 34 publications
1
8
0
Order By: Relevance
“…Notice that even in this data-starved setting, we still outperform the competing unsupervised models. Once we reach only 100,000 lines, performance degrades below mBART but still outperforms the bilingual UNMT approach of Kim et al (2020), revealing the power of multilinguality in low-resource settings.…”
Section: Our Approach Is Robust Under Multiple Domainsmentioning
confidence: 96%
See 3 more Smart Citations
“…Notice that even in this data-starved setting, we still outperform the competing unsupervised models. Once we reach only 100,000 lines, performance degrades below mBART but still outperforms the bilingual UNMT approach of Kim et al (2020), revealing the power of multilinguality in low-resource settings.…”
Section: Our Approach Is Robust Under Multiple Domainsmentioning
confidence: 96%
“…Unsupervised baselines: For the bilingual unsupervised baselines, we include the results of Kim et al (2020) 9 for EnØGu and EnØKk and of for EnØSi. We also report other multilingual unsupervised baselines.…”
Section: Baselinesmentioning
confidence: 99%
See 2 more Smart Citations
“…Another thread of research has pursued learning MT models directly from monolingual data (Artetxe et al, 2017;Lample et al, 2018a,b;Song et al, 2019;Lewis et al, 2019). While unsupervised MT approaches have recently started get-ting close to the quality of fully supervised systems, these approaches are typically brittle, and rely on the availability of large amounts of domain matched monolingual datasets across the source and target languages (Marchisio et al, 2020;Kim et al, 2020); a luxury not available for real-world low resource languages.…”
Section: Introductionmentioning
confidence: 99%