Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1) 2019
DOI: 10.18653/v1/w19-5305
|View full text |Cite
|
Sign up to set email alerts
|

GTCOM Neural Machine Translation Systems for WMT19

Abstract: This paper describes the Global Tone Communication Co., Ltd.'s submission of the WMT19 shared news translation task. We participate in six directions: English to (Gujarati, Lithuanian and Finnish) and (Gujarati, Lithuanian and Finnish) to English. Further, we get the best BLEU scores in the directions of English to Gujarati and Lithuanian to English (28.2 and 36.3 respectively) among all the participants. The submitted systems mainly focus on backtranslation, knowledge distillation and reranking to build a com… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 13 publications
0
3
0
Order By: Relevance
“…2.5.16 GTCOM (Bei et al, 2019) GTCOM's systems (sysNameGTCOM-Primary) mainly focus on backtranslation, knowledge distillation and reranking to build a competitive model with transformer architecture. Also, the language model is applied to filter monolingual data, backtranslated data and parallel data.…”
Section: Frank-s-mtmentioning
confidence: 99%
“…2.5.16 GTCOM (Bei et al, 2019) GTCOM's systems (sysNameGTCOM-Primary) mainly focus on backtranslation, knowledge distillation and reranking to build a competitive model with transformer architecture. Also, the language model is applied to filter monolingual data, backtranslated data and parallel data.…”
Section: Frank-s-mtmentioning
confidence: 99%
“…Comparing Row-11 with its corresponding full-BT baseline in Row-8, we see that this helps for gu→en, giving a further boost in performance of +0.8 to get a final BLEU score of 20.8. To the best of our knowledge this outperforms the bilingual SoTA performance for gu→en (Bei et al, 2019). To summarize, except for hi→en, Iterative-BT helps improve Hinted BT significantly.…”
Section: Iterative Hintedbtmentioning
confidence: 61%
“…We retained the top 80% of sentence pairs based on the alignment score(a score generated by the word alignment model that measures the quality of word alignment between source and target sentences), encompassing all directions. Subsequently, we trained the Transformer model for all languages using Fairseq, following a similar approach as outlined in the study conducted by Bei et al (Bei et al, 2019). The scores were calculated as follows:…”
Section: Bitext Datamentioning
confidence: 99%