Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1430
|View full text |Cite
|
Sign up to set email alerts
|

Exploiting Monolingual Data at Scale for Neural Machine Translation

Abstract: While target-side monolingual data has been proven to be very useful to improve neural machine translation (briefly, NMT) through back translation, source-side monolingual data is not well investigated. In this work, we study how to use both the source-side and targetside monolingual data for NMT, and propose an effective strategy leveraging both of them. First, we generate synthetic bitext by translating monolingual data from the two domains into the other domain using the models pretrained on genuine bitext.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
22
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
7
1
1

Relationship

2
7

Authors

Journals

citations
Cited by 39 publications
(29 citation statements)
references
References 25 publications
1
22
0
Order By: Relevance
“…Self learning (Zhang and Zong, 2016) leverages the source-side monolingual data. Dual learning paradigms utilize monolingual data in both source and target language (He et al, 2016;Wu et al, 2019). While these approaches can effectively improve the NMT performance, they have two limitations.…”
Section: Introductionmentioning
confidence: 99%
“…Self learning (Zhang and Zong, 2016) leverages the source-side monolingual data. Dual learning paradigms utilize monolingual data in both source and target language (He et al, 2016;Wu et al, 2019). While these approaches can effectively improve the NMT performance, they have two limitations.…”
Section: Introductionmentioning
confidence: 99%
“…Supervised Learning on Parallel Data First, we evaluate our model's performance when trained with parallel data on standard WMT datasets. Table 2 shows that our model consistently outperforms both VNMT and DCVAE models-which 2016a; Zhang and Zong, 2016;Wu et al, 2019). We use the joint training objective described in Equation 14.…”
Section: Translation Qualitymentioning
confidence: 88%
“…Back-translation (Sennrich, Haddow, and Birch 2016a), which generates a synthetic training corpus by translating the target-side monolingual sentences with a backward target-to-source model, is widely adopted due to its simplicity and effectiveness. (Wu et al 2019) goes beyond back-translation and leverages both source side and target side monolingual data. Dual learning (He et al 2016;Wang et al 2019) is another way to leverage monolingual data, where the source sentence is first forward translated to the target space and then back translated to the source space.…”
Section: Neural Machine Translationmentioning
confidence: 99%