Second Meeting of the North American Chapter of the Association for Computational Linguistics on Language Technologies 2001 - 2001
DOI: 10.3115/1073336.1073359
|View full text |Cite
|
Sign up to set email alerts
|

Applying co-training methods to statistical parsing

Abstract: We propose a novel Co-Training method for statistical parsing. The algorithm takes as input a small corpus (9695 sentences) annotated with parse trees, a dictionary of possible lexicalized structures for each word in the training set and a large pool of unlabeled text. The algorithm iteratively labels the entire data set with parse trees. Using empirical results based on parsing the Wall Street Journal corpus we show that training a statistical parser on the combined labeled and unlabeled data strongly outperf… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
57
0
2

Year Published

2003
2003
2017
2017

Publication Types

Select...
6
2
2

Relationship

0
10

Authors

Journals

citations
Cited by 88 publications
(60 citation statements)
references
References 24 publications
0
57
0
2
Order By: Relevance
“…However, although co-training has been used in many domains such as statistical parsing and noun phrase identification [22], [29], [33], [38], in most scenarios the requirement of sufficient and redundant views, or even the requirement of sufficient redundancy, could not be met. Therefore, researchers attempt to develop variants of the co-training algorithm for relaxing such a requirement.…”
Section: Semi-supervised Learningmentioning
confidence: 99%
“…However, although co-training has been used in many domains such as statistical parsing and noun phrase identification [22], [29], [33], [38], in most scenarios the requirement of sufficient and redundant views, or even the requirement of sufficient redundancy, could not be met. Therefore, researchers attempt to develop variants of the co-training algorithm for relaxing such a requirement.…”
Section: Semi-supervised Learningmentioning
confidence: 99%
“…Although co-training has already been successfully applied to some fields [20][21] [22], the requirement on two sufficient and redundant attribute subsets might be too strong to be met in many activity recognition systems.…”
Section: Introductionmentioning
confidence: 99%
“…Co-training (Blum and Mitchell, 1998), and several variants of co-training, have been applied to a number of NLP problems, including word sense disambiguation (Yarowsky, 1995), named entity recognition (Collins and Singer, 1999), noun phrase bracketing (Pierce and Cardie, 2001) and statistical parsing (Sarkar, 2001;Steedman et al, 2003). In each case, co-training was used successfully to bootstrap a model from only a small amount of labelled data and a much larger pool of unlabelled data.…”
Section: Introductionmentioning
confidence: 99%