2006
DOI: 10.1007/11731139_10
|View full text |Cite
|
Sign up to set email alerts
|

Using Weighted Nearest Neighbor to Benefit from Unlabeled Data

Abstract: Abstract. The development of data-mining applications such as textclassification and molecular profiling has shown the need for machine learning algorithms that can benefit from both labeled and unlabeled data, where often the unlabeled examples greatly outnumber the labeled examples. In this paper we present a two-stage classifier that improves its predictive accuracy by making use of the available unlabeled data. It uses a weighted nearest neighbor classification algorithm using the combined example-sets as … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
37
0

Year Published

2008
2008
2021
2021

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 41 publications
(37 citation statements)
references
References 7 publications
0
37
0
Order By: Relevance
“…Driessens et al [16] presented a two-stage classifier, YATSI, that improves its predictive accuracy by making use of the available unlabeled data. It used a weighted nearest neighbor classification algorithm using the combined examplesets as a knowledge base.…”
Section: Example 1 In the All-aml Leukaemia Gene Expression Datasetmentioning
confidence: 99%
“…Driessens et al [16] presented a two-stage classifier, YATSI, that improves its predictive accuracy by making use of the available unlabeled data. It used a weighted nearest neighbor classification algorithm using the combined examplesets as a knowledge base.…”
Section: Example 1 In the All-aml Leukaemia Gene Expression Datasetmentioning
confidence: 99%
“…Catal and Diri (2009) proposed the immune-based YATSI (Yet Another Two Stage Idea) (Driessens, 2006) algorithm to predict faulty software, which is a semisupervised meta-algorithm that can be wrapped around any supervised (base) classifier.…”
Section: Related Workmentioning
confidence: 99%
“…And graphs are introduced in semi-supervised learning [14]. Work in [10][13] [15] is most related to our work. [10] proposed a semi-supervised learning based on a Gaussian random field model.…”
Section: Related Workmentioning
confidence: 99%
“…Different from [10] and [13], we present an efficient iterative algorithm for large data sets and for circumstances where matrix computation fails because of the nonexistent inverse matrix. [15] introduced a two-stage approach. In the first stage, a model is built based on training data.…”
Section: Related Workmentioning
confidence: 99%