Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1) 2019
DOI: 10.18653/v1/w19-5337
|View full text |Cite
|
Sign up to set email alerts
|

English-Czech Systems in WMT19: Document-Level Transformer

Abstract: We describe our NMT systems submitted to the WMT19 shared task in English→Czech news translation. Our systems are based on the Transformer model implemented in either Tensor2Tensor (T2T) or Marian framework.We aimed at improving the adequacy and coherence of translated documents by enlarging the context of the source and target. Instead of translating each sentence independently, we split the document into possibly overlapping multi-sentence segments. In case of the T2T implementation, this "documentlevel"-tra… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3

Relationship

2
5

Authors

Journals

citations
Cited by 19 publications
(22 citation statements)
references
References 8 publications
0
14
0
Order By: Relevance
“…(no associated paper) CAIRE CUNI Charles University (Popel et al, 2019;Kocmi and Bojar, 2019) and (Kvapilíková et al, 2019)…”
Section: Afrlmentioning
confidence: 99%
See 2 more Smart Citations
“…(no associated paper) CAIRE CUNI Charles University (Popel et al, 2019;Kocmi and Bojar, 2019) and (Kvapilíková et al, 2019)…”
Section: Afrlmentioning
confidence: 99%
“…and CUNI-TRANSFORMER-T2T2019 (Popel et al, 2019) are trained in the T2T framework following the last year submission (Popel, 2018), but training on WMT19 document-level parallel and monoliongual data. During decoding, each document is split into overlapping multi-sentence segments, where only the "middle" sentences in each segment are used for the final translation.…”
Section: Cuni-doctransformer-t2t2019mentioning
confidence: 99%
See 1 more Smart Citation
“…We use the Transformer architecture by Vaswani et al (2017) implemented in Marian framework (Junczys-Dowmunt et al, 2018) to train an NMT model on the synthetic corpus produced by the PBMT model. The model setup, training and decoding hyperparameters are identical to the CUNI Marian systems in English-to-Czech news translation task in WMT19 (Popel et al, 2019), but in this case, due to smaller and noisier training data, we set the dropout between Transformer layers to 0.3. We use 8 Quadro P5000 GPUs with 16GB memory.…”
Section: Model and Trainingmentioning
confidence: 99%
“…Our other comparison system, Benchmark-TransferEN, was first trained as an English-to-Czech NMT system (see CUNI Transformer Marian for the English-to-Czech news translation task in WMT19 by Popel et al (2019)) and then finetuned for 6 days on the SynthCorpus-noCzechreordered-NER. The vocabulary remained unchanged, it was trained on the English-Czech training corpus.…”
Section: Benchmarksmentioning
confidence: 99%