2020
DOI: 10.48550/arxiv.2010.03142
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Pre-training Multilingual Neural Machine Translation by Leveraging Alignment Information

Abstract: We investigate the following question for machine translation (MT): can we develop a single universal MT model to serve as the common seed and obtain derivative and improved models on arbitrary language pairs? We propose mRASP, an approach to pre-train a universal multilingual neural machine translation model. Our key idea in mRASP is its novel technique of random aligned substitution, which brings words and phrases with similar meanings across multiple languages closer in the representation space. We pre-trai… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
7

Relationship

2
5

Authors

Journals

citations
Cited by 38 publications
(7 citation statements)
references
References 16 publications
0
7
0
Order By: Relevance
“…Some explores data symmetry (Freitag and Firat, 2020;Birch et al, 2008;Lin et al, 2019). Zero-shot translation in severely low resource settings exploits the massive multilinguality, cross-lingual transfer, pretraining, iterative backtranslation and freezing subnetworks (Lauscher et al, 2020;Nooralahzadeh et al, 2020;Pfeiffer et al, 2020;Baziotis et al, 2020;Chronopoulou et al, 2020;Lin et al, 2020;Thompson et al, 2018;Luong et al, 2014;Dou et al, 2020).…”
Section: Machine Polyglotism and Pretrainingmentioning
confidence: 99%
See 2 more Smart Citations
“…Some explores data symmetry (Freitag and Firat, 2020;Birch et al, 2008;Lin et al, 2019). Zero-shot translation in severely low resource settings exploits the massive multilinguality, cross-lingual transfer, pretraining, iterative backtranslation and freezing subnetworks (Lauscher et al, 2020;Nooralahzadeh et al, 2020;Pfeiffer et al, 2020;Baziotis et al, 2020;Chronopoulou et al, 2020;Lin et al, 2020;Thompson et al, 2018;Luong et al, 2014;Dou et al, 2020).…”
Section: Machine Polyglotism and Pretrainingmentioning
confidence: 99%
“…We find that using many languages that are distant to the target low resource language may produce marginal improvements, if not negative impact. Indeed, existing literature on zero-shot translation also suffers from the limitation of linguistic distance between the source languages and the target language (Lauscher et al, 2020;Lin et al, 2020;Pfeiffer et al, 2020). We therefore rank and select the top few source languages that are closer to the target low resource language using the two metrics below.…”
Section: Ranking Source Languagesmentioning
confidence: 99%
See 1 more Smart Citation
“…Some use data selection for active learning (Eck et al, 2005). Some use as few as ∼4,000 lines (Lin et al, 2020;Qi et al, 2018) and ∼1,000 lines (Zhou and Waibel, 2021) of data. Some do not use low resource data (Neubig and Hu, 2018;Karakanta et al, 2018).…”
Section: Severely Low Resource Text-based Translationmentioning
confidence: 99%
“…Given a closed text that has many existing translations in different languages, we are interested in translating it into a severely low resource language well. Researchers recently have shown achievements in translation using very small seed parallel corpora in low resource languages (Lin et al, 2020;Qi et al, 2018;Zhou et al, 2018a). Construction methods of such seed corpora are therefore pivotal in translation performance.…”
Section: Introductionmentioning
confidence: 99%