2003
DOI: 10.1007/978-3-540-45167-9_28
|View full text |Cite
|
Sign up to set email alerts
|

Learning Probabilistic Linear-Threshold Classifiers via Selective Sampling

Abstract: We introduce efficient margin-based algorithms for selective sampling and filtering in binary classification tasks. Experiments on real-world textual data reveal that our algorithms perform significantly better than popular and similarly efficient competitors. Using the so-called Mammen-Tsybakov low noise condition to parametrize the instance distribution, and assuming linear label noise, we show bounds on the convergence rate to the Bayes risk of a weaker adaptive variant of our selective sampler. Our analysi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
20
0

Year Published

2004
2004
2011
2011

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 22 publications
(20 citation statements)
references
References 18 publications
0
20
0
Order By: Relevance
“…A standard tf-idf bag-of-words encoding was used to transform each news story into a normalized vector of real attributes (see Cesa-Bianchi et al, 2003, for details on preprocessing).…”
Section: Methodsmentioning
confidence: 99%
“…A standard tf-idf bag-of-words encoding was used to transform each news story into a normalized vector of real attributes (see Cesa-Bianchi et al, 2003, for details on preprocessing).…”
Section: Methodsmentioning
confidence: 99%
“…If there is disagreement between these experts (usually measured using metrics such as Kullback-Leibler divergence [59] or Jensen-Shannon divergence [270]) beyond a specified threshold, the example is queried and used to re-train the classifiers. [273], and demonstrated their practical applicability in optical character recognition. Dredze and Crammer [263] proposed an online active learner for natural language processing, where the distance of a point from the margin in a large-margin classifier is combined with parameter confidence.…”
Section: Online Active Learning: Related Workmentioning
confidence: 99%
“…Many criteria reflecting various measures of the expected error reduction have been proposed [5,14,18,16,8], with sometimes encouraging results in, for instance, pharmaceutical industry [25]. Specific algorithms and methods have been developed for active learning in linear and kernel spaces, either heuristically [20]or with theoretical foundations [3,8,2].…”
Section: State Of the Artmentioning
confidence: 99%