Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1080
|View full text |Cite
|
Sign up to set email alerts
|

Pivot-based Transfer Learning for Neural Machine Translation between Non-English Languages

Abstract: We present effective pre-training strategies for neural machine translation (NMT) using parallel corpora involving a pivot language, i.e., source-pivot and pivot-target, leading to a significant improvement in source→target translation. We propose three methods to increase the relation among source, pivot, and target languages in the pre-training: 1) step-wise training of a single model for different language pairs, 2) additional adapter component to smoothly connect pre-trained encoder and decoder, and 3) cro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
36
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 44 publications
(37 citation statements)
references
References 26 publications
0
36
0
Order By: Relevance
“…Combining Pre-trained Encoders and Decoders. Kim et al [72] combined S-P encoder with P-T decoder to create the S-T model. They improved the simple initialization using some finetuning objectives and/or source-pivot adaptation to ensure that source and pivot representations are aligned.…”
Section: Zero-resource Translationmentioning
confidence: 99%
“…Combining Pre-trained Encoders and Decoders. Kim et al [72] combined S-P encoder with P-T decoder to create the S-T model. They improved the simple initialization using some finetuning objectives and/or source-pivot adaptation to ensure that source and pivot representations are aligned.…”
Section: Zero-resource Translationmentioning
confidence: 99%
“…Back-translation is simple and easy to achieve without modifying the architecture of the machine translation models. Back-translation has been studied in both SMT [111][112][113] and NMT [23,80,110,[114][115][116].…”
Section: Pivot Translationmentioning
confidence: 99%
“…However, CBMT systems suffer from the lack of parallel corpora for under-resourced languages to train machine translation systems. A number of the methods have been proposed to address the non-availability of parallel corpora for under-resourced languages, such as pivot-based approaches [23][24][25], zero-shot translation [26][27][28][29][30] and unsupervised methods [31][32][33], which are described in detail in following sections. A large array of techniques have been applied to overcome the data sparsity problem in MT, and virtually all of them seem to be based on the field of transfer learning from high-resource languages in recent years.…”
Section: Introductionmentioning
confidence: 99%
“…However, achieving satisfactory performance in low-resource settings turns out to be challenging for NMT systems (Koehn and Knowles, 2017). Recent research has mainly focused on creating and cleaning parallel (Ramasamy et al, 2014;Islam, 2018) and comparable data (Tiedemann, 2012), utilizing bilingual lexicon induction (Conneau et al, 2017;Artetxe et al, 2018;Joty, 2019, 2020;, fine-grained hyperparameter tuning (Sennrich and Zhang, 2019), and using other language pairs as pivot (Cheng et al, 2017;Kim et al, 2019).…”
Section: Related Workmentioning
confidence: 99%