Proceedings of the 15th Conference of the European Chapter of The Association for Computational Linguistics: Volume 2 2017
DOI: 10.18653/v1/e17-2058
|View full text |Cite
|
Sign up to set email alerts
|

Neural Machine Translation by Minimising the Bayes-risk with Respect to Syntactic Translation Lattices

Abstract: We present a novel scheme to combine neural machine translation (NMT) with traditional statistical machine translation (SMT). Our approach borrows ideas from linearised lattice minimum Bayes-risk decoding for SMT. The NMT score is combined with the Bayes-risk of the translation according the SMT lattice. This makes our approach much more flexible than n-best list or lattice rescoring as the neural decoder is not restricted to the SMT search space. We show an efficient and simple way to integrate risk estimatio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
41
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
4

Relationship

3
5

Authors

Journals

citations
Cited by 35 publications
(41 citation statements)
references
References 27 publications
0
41
0
Order By: Relevance
“…Finally, we map the word-level FSTs to the subword-level by composition with a mapping transducer T that applies byte pair encoding (Sennrich et al, 2016c, BPE) to the full words. Word-to-BPE mapping transducers have been used in prior work to combine word-level models with subword-level neural sequence models (Stahlberg et al, , 2017b(Stahlberg et al, , 2018b(Stahlberg et al, , 2017a.…”
Section: Fst-based Grammatical Error Correctionmentioning
confidence: 99%
“…Finally, we map the word-level FSTs to the subword-level by composition with a mapping transducer T that applies byte pair encoding (Sennrich et al, 2016c, BPE) to the full words. Word-to-BPE mapping transducers have been used in prior work to combine word-level models with subword-level neural sequence models (Stahlberg et al, , 2017b(Stahlberg et al, , 2018b(Stahlberg et al, , 2017a.…”
Section: Fst-based Grammatical Error Correctionmentioning
confidence: 99%
“…Restricts the search space to a bag of words with or without repetition (Hasler et al, 2017 • consume(token) Update the internal predictor state by adding token to the current history.…”
Section: Predictorsmentioning
confidence: 99%
“…• nmt,ngramc,wc: MBR-based NMT following Stahlberg et al (2017) with n-gram posteriors extracted from an SMT lattice (ngramc) and a simple word penalty (wc).…”
Section: Example Predictor Constellationsmentioning
confidence: 99%
See 1 more Smart Citation
“…• Even though the performance gap between NMT and traditional statistical machine translation (SMT) is growing rapidly on the task at hand, SMT can still improve very strong NMT ensembles. To combine NMT and SMT we follow Stahlberg et al (2017aStahlberg et al ( , 2018b and build a specialized n-gram LM for each sentence that computes the risk of hypotheses relative to SMT lattices.…”
Section: Introductionmentioning
confidence: 99%