2016
DOI: 10.1016/j.neucom.2016.01.089
|View full text |Cite
|
Sign up to set email alerts
|

Positive and unlabeled learning in categorical data

Abstract: International audienceIn common binary classification scenarios, the presence of both positive and negative examples in training datais needed to build an efficient classifier. Unfortunately, in many domains, this requirement is not satisfied andonly one class of examples is available. To cope with this setting, classification algorithms have been introducedthat learn from Positive and Unlabeled (PU) data. Originally, these approaches were exploited in the context ofdocument classification. Only few works addr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 30 publications
(24 citation statements)
references
References 29 publications
0
21
0
Order By: Relevance
“…Step 3 S-EM [62] Spy EM NB ∆E Roc-SVM [53] Rocchio Iterative SVM [107,106] 1-DNF Iterative SVM Last A-EM [55] Augmented Negatives EM NB ∆F LGN [54] Single Negative BN / PE PUC [108] PE (EM) NB Unspecified WVC/PSOC [77] 1-DNF * Iterative SVM Vote CR-SVM [56] Rocchio * SVM / MCLS [13] k-means Iterative LS-SVM Last C-CRNE [63] C-CRNE TFIPNDF / Pulce [37] DILCA DILCA-KNN / PGPU [31] PGPU biased SVM /…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Step 3 S-EM [62] Spy EM NB ∆E Roc-SVM [53] Rocchio Iterative SVM [107,106] 1-DNF Iterative SVM Last A-EM [55] Augmented Negatives EM NB ∆F LGN [54] Single Negative BN / PE PUC [108] PE (EM) NB Unspecified WVC/PSOC [77] 1-DNF * Iterative SVM Vote CR-SVM [56] Rocchio * SVM / MCLS [13] k-means Iterative LS-SVM Last C-CRNE [63] C-CRNE TFIPNDF / Pulce [37] DILCA DILCA-KNN / PGPU [31] PGPU biased SVM /…”
Section: Methodsmentioning
confidence: 99%
“…This assumption allows identifying reliable negative examples as those that are far from all the labeled examples. This can be done by using different similarity (or distance) measures such as tf-idf for text [53] or DILCA for categorical attributes [37]. This assumption is important for two-step techniques (Section 5.1).…”
Section: Definition 5 (Smoothness)mentioning
confidence: 99%
See 1 more Smart Citation
“…A novel algorithm named Pulce [23] was proposed by Dino Ienco and Ruggero G. Pensa for the positive and unlabeled (PU) learning in categorical data. An efficient classifier can be built by using both positive and negative examples, but this requirement is not satisfied in many domains.…”
Section: Related Workmentioning
confidence: 99%
“…the values of the same context attributes. DILCA has been successfully used in different scenarios including clustering (Ienco et al 2012), semi-supervised learning (Ienco and Pensa 2016) and anomaly detection (Ienco et al 2017). However, if applied to a secret dataset, it may disclose a lot of private information.…”
mentioning
confidence: 99%