Proceedings of SSST-8, Eighth Workshop on Syntax, Semantics and Structure in Statistical Translation 2014
DOI: 10.3115/v1/w14-4012
|View full text |Cite
|
Sign up to set email alerts
|

On the Properties of Neural Machine Translation: Encoder–Decoder Approaches

Abstract: Neural machine translation is a relatively new approach to statistical machine translation based purely on neural networks. The neural machine translation models often consist of an encoder and a decoder. The encoder extracts a fixed-length representation from a variable-length input sentence, and the decoder generates a correct translation from this representation. In this paper, we focus on analyzing the properties of the neural machine translation using two models; RNN Encoder-Decoder and a newly proposed g… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

11
2,471
0
9

Year Published

2017
2017
2023
2023

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 4,682 publications
(2,514 citation statements)
references
References 6 publications
11
2,471
0
9
Order By: Relevance
“…Our model makes extensive use of RNN encoders to transform sequences into fixed length vectors. For our purposes, an RNN encoder consists of GRU units (Cho et al, 2014) defined as…”
Section: Preliminaries and Notationmentioning
confidence: 99%
“…Our model makes extensive use of RNN encoders to transform sequences into fixed length vectors. For our purposes, an RNN encoder consists of GRU units (Cho et al, 2014) defined as…”
Section: Preliminaries and Notationmentioning
confidence: 99%
“…The variations we tried includei) using only LSTM/CNN plus fully connected layers, and also the combination of these architectures with the initial LSTM's output for each word fed to a CNN, or vice versa. ii) Using Simple RNN, Bidirectional LSTM ( (Schuster and Paliwal, 1997), (Godin et al, 2015)), Gated Recurrent Units (GRU) (Cho et al, 2014) instead of LSTM. iii) Using (global) max pooling versus (global) average pooling for CNNs.…”
Section: Approach 3: Sequence Modeling Using Cnns and Lstmsmentioning
confidence: 99%
“…We propose a Natural Language Question Generation (NLQG) model that first encodes the input sequence using some distributed representation and then decodes the output sequence from this encoded representation. Specifically, we use a RNN based encoder and decoder recently proposed for language processing tasks by number of groups (Cho et al, 2014;Sutskever et al, 2014). We now formally define the encoder and decoder models.…”
Section: Rnn Based Natural Language Question Generatormentioning
confidence: 99%