Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1) 2019
DOI: 10.18653/v1/w19-5338
|View full text |Cite
|
Sign up to set email alerts
|

The RWTH Aachen University Machine Translation Systems for WMT 2019

Abstract: This paper describes the neural machine translation systems developed at the RWTH Aachen University for the De→En, Zh→En and Kk→En news translation tasks of the Fourth Conference on Machine Translation (WMT19). For all tasks, the final submitted system is based on the Transformer architecture. We focus on improving data filtering and fine-tuning as well as systematically evaluating interesting approaches like unigram language model segmentation and transfer learning. For the De→En task, none of the tested meth… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 15 publications
(11 reference statements)
0
1
0
Order By: Relevance
“…RUG_ENKK_BPE (Toral et al, 2019) uses data from an additional language (Russian), given the scarcity of English-Kazakh data and synthetic data (for both source and target languages) filtered using language-independent sentence similarity. (Rosendahl et al, 2019) The systems by RWTH AACHEN are all based on Transformer architecture and aside from careful corpus filtering and fine tuning, they experiment with different types of subword units. For English-German, no gains over the last year setup are observed.…”
Section: Rugmentioning
confidence: 99%
“…RUG_ENKK_BPE (Toral et al, 2019) uses data from an additional language (Russian), given the scarcity of English-Kazakh data and synthetic data (for both source and target languages) filtered using language-independent sentence similarity. (Rosendahl et al, 2019) The systems by RWTH AACHEN are all based on Transformer architecture and aside from careful corpus filtering and fine tuning, they experiment with different types of subword units. For English-German, no gains over the last year setup are observed.…”
Section: Rugmentioning
confidence: 99%