Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing 2015
DOI: 10.18653/v1/d15-1092
|View full text |Cite
|
Sign up to set email alerts
|

Sentence Modeling with Gated Recursive Neural Network

Abstract: Recently, neural network based sentence modeling methods have achieved great progress. Among these methods, the recursive neural networks (RecNNs) can effectively model the combination of the words in sentence. However, RecNNs need a given external topological structure, like syntactic tree. In this paper, we propose a gated recursive neural network (GRNN) to model sentences, which employs a full binary tree (FBT) structure to control the combinations in recursive structure. By introducing two kinds of gates, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
25
0
1

Year Published

2015
2015
2020
2020

Publication Types

Select...
4
3
3

Relationship

1
9

Authors

Journals

citations
Cited by 32 publications
(26 citation statements)
references
References 13 publications
(15 reference statements)
0
25
0
1
Order By: Relevance
“…The most primitive embedding representation is the One-Hot encoding, which is simple but has two major shortages: (1) It supposes the words are independent from each other both semantically and grammatically. Hence, it cannot reflect the correlation between two words even they actually have; (2) As the lexicon grows, the dimensions of the vector increase drastically so that a ''dimensional disaster'' will burst, that will lead to a high cost for subsequent computations. To overcome these drawbacks, Hinton propose a distributed representation of word vectors [11], exploiting a fixed-length vector to represent words.…”
Section: Related Workmentioning
confidence: 99%
“…The most primitive embedding representation is the One-Hot encoding, which is simple but has two major shortages: (1) It supposes the words are independent from each other both semantically and grammatically. Hence, it cannot reflect the correlation between two words even they actually have; (2) As the lexicon grows, the dimensions of the vector increase drastically so that a ''dimensional disaster'' will burst, that will lead to a high cost for subsequent computations. To overcome these drawbacks, Hinton propose a distributed representation of word vectors [11], exploiting a fixed-length vector to represent words.…”
Section: Related Workmentioning
confidence: 99%
“…PV slightly surpasses MTLE on IMDB (91.7 against 91.3), as sentences from IMDB are much longer than SST and MDSD, which require stronger capabilities of long-term dependency learning. In this paper, we mainly focus the idea and effects of integrating label embedding with multi-task learning, so we just apply (Graves 2013) to realize Le I and Le L , which can be further implemented by other more effective sentence learning models (Liu et al 2015a;Chen et al 2015) and produce better performances.…”
Section: Comparisons With State-of-the-art Modelsmentioning
confidence: 99%
“…RTfpass Chen et al, 2015] and general-purpose sentence representations [Le and Mikolov, 2014;Kiros et al, 2015;Kenter et al, 2016;Wieting et al, 2016]. Our work addresses the problem of learning general-purpose sentence representations that capture textual semantics and perform robustly across tasks.…”
Section: Modelsmentioning
confidence: 99%