Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) 2014
DOI: 10.3115/v1/p14-2131
|View full text |Cite
|
Sign up to set email alerts
|

Tailoring Continuous Word Representations for Dependency Parsing

Abstract: Word representations have proven useful for many NLP tasks, e.g., Brown clusters as features in dependency parsing (Koo et al., 2008). In this paper, we investigate the use of continuous word representations as features for dependency parsing. We compare several popular embeddings to Brown clusters, via multiple types of features, in both news and web domains. We find that all embeddings yield significant parsing gains, including some recent ones that can be trained in a fraction of the time of others. Explici… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

9
232
3
1

Year Published

2015
2015
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 228 publications
(256 citation statements)
references
References 17 publications
9
232
3
1
Order By: Relevance
“…For syntactic tasks, it has been shown that increasing the window size can adversely impact in the quality of the embeddings (Bansal et al, 2014;Lin et al, 2015).…”
Section: Attention-based Continuousmentioning
confidence: 99%
See 1 more Smart Citation
“…For syntactic tasks, it has been shown that increasing the window size can adversely impact in the quality of the embeddings (Bansal et al, 2014;Lin et al, 2015).…”
Section: Attention-based Continuousmentioning
confidence: 99%
“…While these models are adequate for learning semantic features, one of the problems of this model is the lack of sensitivity for word order, which limits their ability of learn syntactically motivated embeddings (Ling et al, 2015a;Bansal et al, 2014). While models have been proposed to address this problem, the complexity of these models ("Structured skip-n-gram" and "CWindow") grows linearly as size of the window of words considered increases, as a new set of parameters is created for each relative position.…”
Section: Introductionmentioning
confidence: 99%
“…This hypothesis supports unsupervised learning of meaningful word representations from large corpora (Curran, 2003;Ó Séaghdha and Korhonen, 2014;Mikolov et al, 2013;Pennington et al, 2014). Word vectors trained using these methods have proven useful for many downstream tasks including machine translation (Zou et al, 2013) and dependency parsing (Bansal et al, 2014).…”
Section: Introductionmentioning
confidence: 99%
“…word2vec and GloVe representations have provided state-of-the-art results on various word similarity and analogy detection task (Mikolov et al, 2013c;Mikolov et al, 2013b;Pennington et al, 2014). Word embedding based models are also used for other NLP tasks such as dependency parsing, semantic role labeling, POS tagging, NER, question-answering (Bansal et al, 2014;Collobert et al, 2011;Weston et al, 2015) and our work on LSSD is a novel application of word embeddings.…”
Section: Related Workmentioning
confidence: 99%