2009
DOI: 10.1016/j.jcss.2008.07.003
|View full text |Cite
|
Sign up to set email alerts
|

Agnostic active learning

Abstract: We state and analyze the first active learning algorithm that finds an -optimal hypothesis in any hypothesis class, when the underlying distribution has arbitrary forms of noise. The algorithm, A 2 (for Agnostic Active), relies only upon the assumption that it has access to a stream of unlabeled examples drawn i.i.d. from a fixed distribution. We show that A 2 achieves an exponential improvement (i.e., requires only O (ln 1 ) samples to find an -optimal classifier) over the usual sample complexity of supervise… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

4
249
1

Year Published

2010
2010
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 218 publications
(255 citation statements)
references
References 25 publications
4
249
1
Order By: Relevance
“…As was shown by [8], active learning in this setting is particularly challenging, and a long line of work, from theoretical to practical, has specifically addressed this setting [3], [9], [10], [4], [15], [2].…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…As was shown by [8], active learning in this setting is particularly challenging, and a long line of work, from theoretical to practical, has specifically addressed this setting [3], [9], [10], [4], [15], [2].…”
Section: Related Workmentioning
confidence: 99%
“…There is a great deal of literature on active learning [7], [9], [1], [13]. The problem is tractable when the oracle always outputs the ground truth labels [7], in which case, a generalized binary search-style approach results in an estimation error that decays exponentially with the number of labels.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The labels for messages, which on the average appear on the wrong side of the boundary, are flipped and a final SVM model is trained using the modified data. We note that RHT expands on agnostic active (A2) learning [44] [45] that maintains both a current version space and a region of uncertainty. Two data sets, TREC 2007 and CEAS 2008 were used for comparing the performance of RHT-SVM to the performance of Reject On Negative Impact (RONI) [5] as well as the performance of an SVM trained on the tainted training data.…”
Section: Semantics and Stratificationmentioning
confidence: 99%
“…While margin-based active learning remains by far the most popular formalism despite a lack of strong performance guarantees, there have been several recent works examining active learning based upon the PAC learning model [74] for realizable concept classes [7,16,25,27,37] and the agnostic learning model [42] for broader concept classes [5,6,26,38]. However, while these results are important, it should also be noted that they make assumptions which render them generally less applicable to the complex applications where active learning in most useful, although this is certainly a direction for future study.…”
Section: Related Workmentioning
confidence: 99%