2011 IEEE 11th International Conference on Data Mining Workshops 2011
DOI: 10.1109/icdmw.2011.20
|View full text |Cite
|
Sign up to set email alerts
|

Active Learning from Positive and Unlabeled Data

Abstract: Abstract-During recent years, active learning has evolved into a popular paradigm for utilizing user's feedback to improve accuracy of learning algorithms. Active learning works by selecting the most informative sample among unlabeled data and querying the label of that point from user. Many different methods such as uncertainty sampling and minimum risk sampling have been utilized to select the most informative sample in active learning. Although many active learning algorithms have been proposed so far, most… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(14 citation statements)
references
References 16 publications
0
14
0
Order By: Relevance
“…Positive and unlabeled (PU) data can be regarded as a kind of noisy-label data, in which we mainly consider the probability that positive samples are mislabeled as negative ones. Ghasemi et al [12] proposed an active learning algorithm for PU data, which works by separately estimating probability density of positive and unlabeled points and then computing expected value of informativeness to get rid of a hyperparameter and have a better measure of informativeness. Plessis et al [33] proposed a cost-sensitive classifier, which utilizes a non-convex loss to prevent the superfluous penalty term in the objective function.…”
Section: Noisy-label Robust Learningmentioning
confidence: 99%
“…Positive and unlabeled (PU) data can be regarded as a kind of noisy-label data, in which we mainly consider the probability that positive samples are mislabeled as negative ones. Ghasemi et al [12] proposed an active learning algorithm for PU data, which works by separately estimating probability density of positive and unlabeled points and then computing expected value of informativeness to get rid of a hyperparameter and have a better measure of informativeness. Plessis et al [33] proposed a cost-sensitive classifier, which utilizes a non-convex loss to prevent the superfluous penalty term in the objective function.…”
Section: Noisy-label Robust Learningmentioning
confidence: 99%
“…These actions could exist of labeling part of the unseen data, such that it could be used for retraining a supervised machine learning algorithm (e.g. active learning [2,3]), to quickly generalize to the unseen data. Or by using methods such as data augmentation [4], transfer learning [1,5,6] or representation learning [7,8], which are commonly used to extend the scope of machine learning algorithms.…”
Section: Introductionmentioning
confidence: 99%
“…Besides the three proposed query strategies, our experiments considered the active learning approaches "lh" uncertainty sampling [5], expected margin sampling [7], entropy sampling [7], outlier sampling [6] and random sampling. The two kernel density estimation strategies of [7] were implemented using the publicly available code by Ghasemi [18]. In the case of large datasets, batches consisting of multiple samples were queried after each iterative training step in order to reduce the computational effort.…”
Section: A Setupmentioning
confidence: 99%
“…Ghasemi et al presented a different approach which makes use of the distribution of target and unlabeled samples and does not consider the classification results of an OCC for active learning. Based on kernel density estimation, they proposed the two query strategies expected margin sampling and entropy sampling [7]. There exist further query strategies which are not considered in this paper due to their limitation to special OCCs or high computational cost.…”
Section: Introductionmentioning
confidence: 99%