2017
DOI: 10.1515/pralin-2017-0035
|View full text |Cite
|
Sign up to set email alerts
|

NMTPY: A Flexible Toolkit for Advanced Neural Machine Translation Systems

Abstract: In this paper, we present nmtpy, a flexible Python toolkit based on Theano for training Neural Machine Translation and other neural sequence-to-sequence architectures. nmtpy decouples the specification of a network from the training and inference utilities to simplify the addition of a new architecture and reduce the amount of boilerplate code to be written. nmtpy has been used for LIUM's topranked submissions to WMT Multimodal Machine Translation and News Translation tasks in 2016 and 2017.

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
40
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
4
1

Relationship

2
8

Authors

Journals

citations
Cited by 57 publications
(43 citation statements)
references
References 21 publications
0
40
0
Order By: Relevance
“…We decode hypotheses using a beam size of 10. The experiments are conducted using nmtpytorch 1 [23].…”
Section: Resultsmentioning
confidence: 99%
“…We decode hypotheses using a beam size of 10. The experiments are conducted using nmtpytorch 1 [23].…”
Section: Resultsmentioning
confidence: 99%
“…The model hyper parameters are the same ones as in [10]. All models are implemented using the nmtpytorch framework [21]. For each experiment, we train three models with different random seeds and report the average results.…”
Section: Model Implementationmentioning
confidence: 99%
“…By using ensembling 3 networks with different configs and rescoring using a model trained with reversed target sentences, we managed to reach 26.96 BLEU score for the development set, which yields 2.8 point of improvement compared to the baseline model. Details about the effect of each technique is described in Pham et al (2017) 3.3 LIMSI LIMSI's intput to this system combination consists of two NMT systems, both trained with the NMTPY framework (Caglayan et al, 2017) on bitext, then on synthetic parallel data. All of them were rescored with a Nematus system (Sennrich et al, 2017b).…”
Section: Kitmentioning
confidence: 99%