2020
DOI: 10.2197/ipsjjip.28.413
|View full text |Cite
|
Sign up to set email alerts
|

A Survey of Domain Adaptation for Machine Translation

Abstract: Neural machine translation (NMT) is a deep learning based approach for machine translation, which yields the state-of-the-art translation performance in scenarios where large-scale parallel corpora are available. Although the high-quality and domain-specific translation is crucial in the real world, domain-specific corpora are usually scarce or nonexistent, and thus vanilla NMT performs poorly in such scenarios. Domain adaptation that leverages both out-of-domain parallel corpora as well as monolingual corpora… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

1
120
0
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 115 publications
(124 citation statements)
references
References 50 publications
(36 reference statements)
1
120
0
1
Order By: Relevance
“…Other surveys Comprehensive reviews on DA exist, each with a different focus: visual applications (Csurka, 2017;Patel et al, 2015;Wilson and Cook, 2020), machine translation (MT) (Chu and Wang, 2018), pre-neural DA methods in NLP (Jiang, 2008;Margolis, 2011). Seminal surveys in machine learning on transfer learning include Pan and Yang (2009), Weiss et al (2016), and.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Other surveys Comprehensive reviews on DA exist, each with a different focus: visual applications (Csurka, 2017;Patel et al, 2015;Wilson and Cook, 2020), machine translation (MT) (Chu and Wang, 2018), pre-neural DA methods in NLP (Jiang, 2008;Margolis, 2011). Seminal surveys in machine learning on transfer learning include Pan and Yang (2009), Weiss et al (2016), and.…”
Section: Introductionmentioning
confidence: 99%
“…We take inspiration of the data-centric and model-centric terms fromChu and Wang (2018) in MT, and add hybrid.3 We disregard methods which are task-specific (like leveraging a sentiment thesaurus).…”
mentioning
confidence: 99%
“…domain transfer (Shimodaira, 2000;Subbaswamy and Saria, 2020). Accordingly, substantial effort has been devoted to developing computational methods for domain adaptation (Imran et al, 2016;Chu and Wang, 2018). Outcomes from this work often provide a solid foundation for use across multiple natural language processing tasks (Daume III and Marcu, 2006).…”
Section: Introductionmentioning
confidence: 99%
“…During training, we evaluate the performance of the model after every epoch using a development set from the Biomedical domain. Our system is prone to over-fitting as the Biomedical (2014 and2018) training data sets that we use are significantly smaller (see Table 1) as compared to News. Generally over-fitting means that the model performs excellent on the training data, but worse on (Koehn, 2017) any other unseen data.…”
Section: Transfer Learningmentioning
confidence: 99%
“…Data-driven machine translation models assume the training data and test data have the same distribution and feature space (Koehn, 2009), which is rare in real-world applications (Olive et al, 2011). In statistical machine translation, a standard solution is to apply domain adaptation (Xu et al, 2007;Foster and Kuhn, 2007;Chu and Wang, 2018). For example, interpolating phrase or word probabilities in a sentence learned on in-domain and outof-domain data and then computing their product.…”
Section: Introductionmentioning
confidence: 99%