Proceedings of the Human Language Technology Conference of the NAACL, Companion Volume: Short Papers on XX - NAACL '06 2006
DOI: 10.3115/1614049.1614067
|View full text |Cite
|
Sign up to set email alerts
|

MMR-based active machine learning for bio named entity recognition

Abstract: This paper presents a new active learning paradigm which considers not only the uncertainty of the classifier but also the diversity of the corpus. The two measures for uncertainty and diversity were combined using the MMR (Maximal Marginal Relevance) method to give the sampling scores in our active learning strategy. We incorporated MMR-based active machinelearning idea into the biomedical namedentity recognition system. Our experimental results indicated that our strategies for active-learning based sample s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
41
0

Year Published

2007
2007
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 39 publications
(42 citation statements)
references
References 11 publications
1
41
0
Order By: Relevance
“…Note, however, that the number of possible labelings grows exponentially with the length of x. To make this feasible, previous work (Kim et al, 2006) has employed an approximation we call N-best sequence entropy (NSE):…”
Section: Uncertainty Samplingmentioning
confidence: 99%
See 1 more Smart Citation
“…Note, however, that the number of possible labelings grows exponentially with the length of x. To make this feasible, previous work (Kim et al, 2006) has employed an approximation we call N-best sequence entropy (NSE):…”
Section: Uncertainty Samplingmentioning
confidence: 99%
“…A few methods have been proposed, based mostly on the conventions of uncertainty sampling, where the learner queries the instance about which it has the least certainty (Scheffer et al, 2001;Culotta and McCallum, 2005;Kim et al, 2006), or query-by-committee, where a "committee" of models selects the instance about which its members most disagree (Dagan and Engelson, 1995). We provide more detail on these and the new strategies we propose in Section 3.…”
Section: Introductionmentioning
confidence: 99%
“…This criterion assumes that the most uncertain sentence is the most useful instance for learning an NER model. There are several ways to implement this, such as least confidence (Culotta and McCallum, 2005)-the lower the probability of a sequence of labels, the less confidence the model, and entropy (Kim et al, 2006) that can measure the uncertainty of a probability distribution. Some other criteria are a diversity measurement (Kim et al, 2006) and a density criterion (Settles and Craven, 2008).…”
Section: Comparison With Baselinesmentioning
confidence: 99%
“…We experimented with the following criteria: least confidence (Culotta and McCallum, 2005), normalized entropy (Kim et al, 2006), MMR (Maximal Marginal Relevance) (Kim et al, 2006), density (Settles and Craven, 2008) when using feature vectors and word embeddings, and the combination of least confidence and density criterion. Equation 8 describes the combination criterion used in our experiments.…”
Section: Active Learning Criteriamentioning
confidence: 99%
“…Kim et al (2006) propose using entropy as a confidence estimator in active learning in CRFs, where examples with the most uncertainty are selected for presentation to humans labelers. In practice, they approximate the entropy of the labels given the N-best labels.…”
Section: Confidence Estimationmentioning
confidence: 99%