Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2016
DOI: 10.18653/v1/n16-1006
|View full text |Cite
|
Sign up to set email alerts
|

An Empirical Evaluation of Noise Contrastive Estimation for the Neural Network Joint Model of Translation

Abstract: The neural network joint model of translation or NNJM (Devlin et al., 2014) combines source and target context to produce a powerful translation feature. However, its softmax layer necessitates a sum over the entire output vocabulary, which results in very slow maximum likelihood (MLE) training. This has led some groups to train using Noise Contrastive Estimation (NCE), which side-steps this sum. We carry out the first direct comparison of MLE and NCE training objectives for the NNJM, showing that NCE is signi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2017
2017
2018
2018

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 16 publications
0
1
0
Order By: Relevance
“…U i = M i e P A, SI, M i f , CH . Observing that the alignment A and source-dependency tree T f are fixed once the bilingual corpus is given, we specify the NB(U i ) as the set of DBiCUs: each DBiCU is with form of M e P A, SI, M i f , CH , satisfying |M e |=|M i e |; and it is generated by the IBM Model 1 distribution (Brown et al 1993), inspired by (Cherry 2016).…”
Section: Dbicsnnlm Trainingmentioning
confidence: 99%
“…U i = M i e P A, SI, M i f , CH . Observing that the alignment A and source-dependency tree T f are fixed once the bilingual corpus is given, we specify the NB(U i ) as the set of DBiCUs: each DBiCU is with form of M e P A, SI, M i f , CH , satisfying |M e |=|M i e |; and it is generated by the IBM Model 1 distribution (Brown et al 1993), inspired by (Cherry 2016).…”
Section: Dbicsnnlm Trainingmentioning
confidence: 99%