Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing 2015
DOI: 10.18653/v1/d15-1041
|View full text |Cite
|
Sign up to set email alerts
|

Improved Transition-based Parsing by Modeling Characters instead of Words with LSTMs

Abstract: We present extensions to a continuousstate dependency parsing method that makes it applicable to morphologically rich languages. Starting with a highperformance transition-based parser that uses long short-term memory (LSTM) recurrent neural networks to learn representations of the parser state, we replace lookup-based word representations with representations constructed from the orthographic representations of the words, also using LSTMs. This allows statistical sharing across word forms that are similar on … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
181
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 224 publications
(186 citation statements)
references
References 22 publications
0
181
0
Order By: Relevance
“…Ling et al (2015a) used a bidirectional long shortterm memory (LSTM) RNN on characters to embed arbitrary word types, showing strong performance for language modeling and POS tagging. Ballesteros et al (2015) used this model to represent words for dependency parsing. Several have used character-level RNN architectures for machine translation, whether for representing source or target words (Ling et al, 2015b;Luong and Manning, 2016), or for generating entire translations character-by-character (Chung et al, 2016).…”
Section: Related Workmentioning
confidence: 99%
“…Ling et al (2015a) used a bidirectional long shortterm memory (LSTM) RNN on characters to embed arbitrary word types, showing strong performance for language modeling and POS tagging. Ballesteros et al (2015) used this model to represent words for dependency parsing. Several have used character-level RNN architectures for machine translation, whether for representing source or target words (Ling et al, 2015b;Luong and Manning, 2016), or for generating entire translations character-by-character (Chung et al, 2016).…”
Section: Related Workmentioning
confidence: 99%
“…Generally, given a document and a fixed number of classes, the classification model has to predict the class that is most relevant to that document. Several recent studies have discovered that character-based representation provides straightforward and powerful models for relation extraction [28], sentiment classification [33], and transition based parsing [5]. Lodhi et.…”
Section: S:2 Connecting To Previous Studiesmentioning
confidence: 99%
“…Several recent studies have discovered that character-based representation provides straightforward and powerful models for relation extraction [28], sentiment classification [33], and transition based parsing [5]. We downloaded the processed WebKB datasets (removed stop/short words, stemming, etc.)…”
Section: S:32 19 Datasets Used In Evaluationsmentioning
confidence: 99%
“…Based on Dyer et al (2015). Ballesteros et al (2015) further propose to use LSTM to model the relation among characters, leading to better parsing performances on morphology-rich languages. By modeling more history, the parser gives significant better accuracies compared to the greedy neural parser of Chen and Manning.…”
Section: Parsing By Neural Networkmentioning
confidence: 99%