Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) 2016
DOI: 10.18653/v1/p16-2049
|View full text |Cite
|
Sign up to set email alerts
|

Syntactically Guided Neural Machine Translation

Abstract: We investigate the use of hierarchical phrase-based SMT lattices in end-to-end neural machine translation (NMT). Weight pushing transforms the Hiero scores for complete translation hypotheses, with the full translation grammar score and full ngram language model score, into posteriors compatible with NMT predictive probabilities. With a slightly modified NMT beam-search decoder we find gains over both Hiero and NMT decoding alone, with practical advantages in extending NMT to very large input and output vocabu… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
44
0
1

Year Published

2017
2017
2021
2021

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 57 publications
(47 citation statements)
references
References 23 publications
(27 reference statements)
2
44
0
1
Order By: Relevance
“…Although Bahdanau et al (2015) used a bidirectional recurrent neural network (RNN) (Schuster and Paliwal, 1997) to consider preceding and following words jointly, these sequential representations are insufficient to fully capture the semantics of a sentence, due to the fact that they do not account for the syntactic interpretations of sentence structure (Eriguchi et al, 2016;Tai et al, 2015). By incorporating additional features into a sequential model, and Stahlberg et al (2016) suggest that a greater amount of linguistic information can improve the translation performance.…”
Section: Introductionmentioning
confidence: 99%
“…Although Bahdanau et al (2015) used a bidirectional recurrent neural network (RNN) (Schuster and Paliwal, 1997) to consider preceding and following words jointly, these sequential representations are insufficient to fully capture the semantics of a sentence, due to the fact that they do not account for the syntactic interpretations of sentence structure (Eriguchi et al, 2016;Tai et al, 2015). By incorporating additional features into a sequential model, and Stahlberg et al (2016) suggest that a greater amount of linguistic information can improve the translation performance.…”
Section: Introductionmentioning
confidence: 99%
“…Some of these solutions lead to improvements in performance, but they all require time-intensive training of the NMT models to use an enriched input representation or to optimize the parameters of the model. (Stahlberg et al, 2016) proposed an approach which can be used at decoding time. A hierarchical PBSMT system is used to generate the translation lattices, which are then re-scored by the NMT decoder.…”
Section: Related Workmentioning
confidence: 99%
“…Regarding handling OOV words, Jean et al (2015) presented an efficient training method to support a larger vocabulary, which helps alleviate the OOV problem significantly. Stahlberg et al (2016) used SMT to produce candidate results in the form of lattice and NMT to re-score the results. As SMT uses a larger vocabulary than NMT, some OOV words can be retained.…”
Section: Related Workmentioning
confidence: 99%