Proceedings of the Main Conference on Human Language Technology Conference of the North American Chapter of the Association of 2006
DOI: 10.3115/1220835.1220851
|View full text |Cite
|
Sign up to set email alerts
|

An empirical study of the behavior of active learning for word sense disambiguation

Abstract: This paper shows that two uncertaintybased active learning methods, combined with a maximum entropy model, work well on learning English verb senses. Data analysis on the learning process, based on both instance and feature levels, suggests that a careful treatment of feature extraction is important for the active learning to be useful for WSD. The overfitting phenomena that occurred during the active learning process are identified as classic overfitting in machine learning based on the data analysis.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
34
0

Year Published

2007
2007
2023
2023

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 46 publications
(35 citation statements)
references
References 8 publications
0
34
0
Order By: Relevance
“…A trend of the last ten years (Abe and Mamitsuka 1998;Banko and Brill 2001;Chen et al 2006;Dagan and Engelson 1995;Hwa 2004;Lewis and Gale 1994;McCallum and Nigam 1998;Melville and Mooney 2004;Roy and McCallum 2001;Tang et al 2002) has been to employ heuristic methods of active learning with no explicitly defined objective function. Uncertainty sampling (Lewis and Gale 1994), query by committee (Seung et al 1992), 1 and variants have proven particularly attractive because of their portability across a wide spectrum of machine learning algorithms.…”
Section: Background and Related Workmentioning
confidence: 99%
“…A trend of the last ten years (Abe and Mamitsuka 1998;Banko and Brill 2001;Chen et al 2006;Dagan and Engelson 1995;Hwa 2004;Lewis and Gale 1994;McCallum and Nigam 1998;Melville and Mooney 2004;Roy and McCallum 2001;Tang et al 2002) has been to employ heuristic methods of active learning with no explicitly defined objective function. Uncertainty sampling (Lewis and Gale 1994), query by committee (Seung et al 1992), 1 and variants have proven particularly attractive because of their portability across a wide spectrum of machine learning algorithms.…”
Section: Background and Related Workmentioning
confidence: 99%
“…In this article, we are interested in uncertainty sampling schemes [Lewis and Gale 1994] for pool-based active learning, which in recent years has been widely studied in tasks such as word sense disambiguation [Chen et al 2006;Chan and Ng 2007], Text Classification (TC) [Lewis and Gale 1994;Zhu et al 2008b], statistical syntactic parsing [Tang et al 2002], and named entity recognition [Shen et al 2004].…”
Section: Active Learning Processmentioning
confidence: 99%
“…The main difference among the various pool-based active learning algorithms is the method of assessing the uncertainty of each unlabeled example in the pool. In the case of probabilistic models, the uncertainty of the classifier is commonly estimated using the entropy of its output [Tang et al 2002;Chen et al 2006;Zhu and Hovy 2007]. For active learning with nonprobabilistic models such as support vector machines [Tong and Koller 2001;Schohn and Cohn 2000], the classification margin is used.…”
Section: Active Learning Processmentioning
confidence: 99%
See 2 more Smart Citations