2017
DOI: 10.1109/taslp.2017.2751420
|View full text |Cite
|
Sign up to set email alerts
|

A Context-Aware Recurrent Encoder for Neural Machine Translation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
26
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 62 publications
(26 citation statements)
references
References 14 publications
0
26
0
Order By: Relevance
“…In particular, our single model yields a detokenized BLEU score of 21.99. In order to show that the proposed model can be orthogonal to previous methods that improve LSTM/GRU-based NMT, we integrate a singlelayer context-aware (CA) encoder (Zhang et al, 2017b) into our system. The ATR+CA system further reaches 22.7 BLEU, outperforming the winner system (Buck et al, 2014) by a substantial improvement of 2 BLEU points.…”
Section: Results On English-german Translationmentioning
confidence: 99%
See 1 more Smart Citation
“…In particular, our single model yields a detokenized BLEU score of 21.99. In order to show that the proposed model can be orthogonal to previous methods that improve LSTM/GRU-based NMT, we integrate a singlelayer context-aware (CA) encoder (Zhang et al, 2017b) into our system. The ATR+CA system further reaches 22.7 BLEU, outperforming the winner system (Buck et al, 2014) by a substantial improvement of 2 BLEU points.…”
Section: Results On English-german Translationmentioning
confidence: 99%
“…"RL" and "WPM" is the reinforcement learning optimization and word piece model used in . "CA" is the context-aware recurrent encoder (Zhang et al, 2017b). "LAU" and "F-F" denote the linear associative unit and the fast-forward architecture proposed by Wang et al (2017a) and Zhou et al (2016) respectively.…”
Section: Trainingmentioning
confidence: 99%
“…W ORD embedding is a real-valued vector representation of words by embedding both semantic and syntactic meanings obtained from unlabeled large corpus. It is a powerful tool widely used in modern natural language processing (NLP) tasks, including semantic analysis [1], information retrieval [2], dependency parsing [3], [4], [5], question answering [6], [7] and machine translation [6], [8], [9]. Learning a high quality representation is extremely important for these tasks, yet the question "what is a good word embedding model" remains an open problem.…”
Section: Introductionmentioning
confidence: 99%
“…Indeed, the model proposed by Choi et al (2017) attempts to improve NMT by integrating context vectors associated to source words into the generation process during decoding. The model proposed by Zhang et al (2017) is aware of previous attended words on the source side in order to better predict which words will be attended in future. The self-attentive residual decoder designed by Werlen et al (2018) leverages the contextual information from previously translated words on the target side.…”
Section: Resultsmentioning
confidence: 99%