Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2014
DOI: 10.3115/v1/d14-1179
|View full text |Cite
|
Sign up to set email alerts
|

Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation

Abstract: In this paper, we propose a novel neural network model called RNN EncoderDecoder that consists of two recurrent neural networks (RNN). One RNN encodes a sequence of symbols into a fixedlength vector representation, and the other decodes the representation into another sequence of symbols. The encoder and decoder of the proposed model are jointly trained to maximize the conditional probability of a target sequence given a source sequence. The performance of a statistical machine translation system is empiricall… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

14
9,886
1
103

Year Published

2016
2016
2022
2022

Publication Types

Select...
6
2
2

Relationship

0
10

Authors

Journals

citations
Cited by 16,798 publications
(10,072 citation statements)
references
References 13 publications
14
9,886
1
103
Order By: Relevance
“…In order to train an NMT (Cho et al, 2014;Sutskever et al, 2014;Bahdanau et al, 2015) model for a languagepair, the size of vocabularies for source and target languages should be constant. But in reality, the vocabulary of a natural language is open.…”
Section: Introductionmentioning
confidence: 99%
“…In order to train an NMT (Cho et al, 2014;Sutskever et al, 2014;Bahdanau et al, 2015) model for a languagepair, the size of vocabularies for source and target languages should be constant. But in reality, the vocabulary of a natural language is open.…”
Section: Introductionmentioning
confidence: 99%
“…As Tilk et al, we use gated recurrent units (GRU) [8] for the RNN layers. Introduced as a simpler variate of long short-term memory (LSTM) units [11], GRUs make computation simpler by having fewer parameters.…”
Section: Our Modelmentioning
confidence: 99%
“…Favour underlying semantic and syntactic information in natural language texts, and save researchers the efforts of feature engineering [14,15]. Recently, they have achieved significant improvements in various natural language processing tasks, such as Machine Translation [2,3], Question Answering [14], Sentiment Analysis [6,11,15,18], etc. However, applying deep neural networks on target-specific Stance Detection has not been successful, as their performances have, up to now, been slightly worse than traditional machine learning algorithms with manual feature engineering, such as Support Vector Machines (SVM) [8].…”
Section: Againstmentioning
confidence: 99%