Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing 2015
DOI: 10.18653/v1/d15-1278
|View full text |Cite
|
Sign up to set email alerts
|

When Are Tree Structures Necessary for Deep Learning of Representations?

Abstract: Recursive neural models, which use syntactic parse trees to recursively generate representations bottom-up, are a popular architecture. But there have not been rigorous evaluations showing for exactly which tasks this syntax-based method is appropriate. In this paper we benchmark recursive neural models against sequential recurrent neural models (simple recurrent and LSTM models), enforcing apples-to-apples comparison as much as possible. We investigate 4 tasks: (1) sentiment classification at the sentence lev… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

4
119
0
1

Year Published

2015
2015
2021
2021

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 156 publications
(124 citation statements)
references
References 25 publications
4
119
0
1
Order By: Relevance
“…This has allowed for the simplified creation of large labeled datasets, ideal for the application of unsupervised learning methods. Approaches have continued to evolve, both in terms of problem formulation (Chen et al, 2015) and as the full weight of modern machine learning techniques have been brought to bear (Socher et al, 2013;Tai et al, 2015;Li et al, 2015).…”
Section: Study 1: Sentiment Analysismentioning
confidence: 99%
“…This has allowed for the simplified creation of large labeled datasets, ideal for the application of unsupervised learning methods. Approaches have continued to evolve, both in terms of problem formulation (Chen et al, 2015) and as the full weight of modern machine learning techniques have been brought to bear (Socher et al, 2013;Tai et al, 2015;Li et al, 2015).…”
Section: Study 1: Sentiment Analysismentioning
confidence: 99%
“…Neural network classifiers are popular for relation extraction recently. Many of them focus on fully supervised settings, recurrent neural networks (RNN) and convolutional neural networks (CNN) (Vu et al, 2016;Zeng et al, 2015;Xu et al, 2015a;Xu et al, 2015b;Zhang and Wang, 2015), sequence models and tree models are investigated (Li et al, 2015;dos Santos et al, 2015). One similar network structure to our model is proposed in (Miwa and Bansal, 2016).…”
Section: Related Workmentioning
confidence: 99%
“…Compared with previous neural models, we keep the advantage of convolutional neural network (Nguyen and Grishman, 2015) in capturing local contexts. Besides, we also incorporate a Bi-directional LSTM to model the preceding and following information of a word as it has been commonly accepted that LSTM is good at capturing long-term dependencies in a sequence (Tang et al, 2015b;Li et al, 2015a).…”
Section: Related Workmentioning
confidence: 99%