Proceedings of the Third Conference on Machine Translation: Shared Task Papers 2018
DOI: 10.18653/v1/w18-6427
|View full text |Cite
|
Sign up to set email alerts
|

The University of Cambridge’s Machine Translation Systems for WMT18

Abstract: The University of Cambridge submission to the WMT18 news translation task focuses on the combination of diverse models of translation. We compare recurrent, convolutional, and self-attention-based neural models on German-English, English-German, and Chinese-English. Our final system combines all neural models together with a phrase-based SMT system in an MBR-based scheme. We report small but consistent gains on top of strong Transformer ensembles.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
40
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
5

Relationship

2
8

Authors

Journals

citations
Cited by 18 publications
(40 citation statements)
references
References 36 publications
0
40
0
Order By: Relevance
“…Especially, for newstest2016 and newstest2018, we achieve 1.0 BLEU score improvement over the MS-Marian and set new records for these tasks. We also list the WMT18 top-2 systems for De!En translation in Table 4: RWTH (Graça et al, 2018) and UCAM (Stahlberg et al, 2018) systems, which are both ensemble models. Similarly, our single model surpasses these ensemble systems by a large margin.…”
Section: Resultsmentioning
confidence: 99%
“…Especially, for newstest2016 and newstest2018, we achieve 1.0 BLEU score improvement over the MS-Marian and set new records for these tasks. We also list the WMT18 top-2 systems for De!En translation in Table 4: RWTH (Graça et al, 2018) and UCAM (Stahlberg et al, 2018) systems, which are both ensemble models. Similarly, our single model surpasses these ensemble systems by a large margin.…”
Section: Resultsmentioning
confidence: 99%
“…As our APE model seems agnostic to the model which produced the RTT, we applied it to the best submissions of the recent WMT18 evaluation campaign, applying to German-original half of the test set only. Table 4 shows the results for the 2 top submissions of Microsoft (Junczys-Dowmunt, 2018) and Cambridge (Stahlberg et al, 2018). Both systems improved by up to 0.8 points in BLEU.…”
Section: English→germanmentioning
confidence: 99%
“…Our LMs are Transformer (Vaswani et al, 2017) decoders (transformer big) trained using the Tensor2Tensor library (Vaswani et al, 2018). We delay SGD updates (Stahlberg et al, 2018a;Saunders et al, 2018) with factor 2 to simulate 500K training steps with 8 GPUs on 4 physical GPUs. Training batches contain about 4K source and target tokens.…”
Section: Methodsmentioning
confidence: 99%