2015
DOI: 10.1016/j.knosys.2015.06.010
|View full text |Cite
|
Sign up to set email alerts
|

Nearest neighbor regression in the presence of bad hubs

Abstract: Prediction on a numeric scale, i.e., regression, is one of the most prominent machine learning tasks with various applications in finance, medicine, social and natural sciences. Due to its simplicity, theoretical performance guarantees and successful real-world applications, one of the most popular regression techniques is the k nearest neighbor regression. However, k nearest neighbor approaches are affected by the presence of bad hubs, a recently observed phenomenon according to which some of the instances ar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
21
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
4

Relationship

2
6

Authors

Journals

citations
Cited by 40 publications
(21 citation statements)
references
References 35 publications
(47 reference statements)
0
21
0
Order By: Relevance
“…This means, roughly speaking, that bad hubs are expected in complex data, such as drug-target interaction data. For a more detailed discussion, we refer to [7].…”
Section: Ecknn: K-nearest Neighbor Regression With Error Correctionmentioning
confidence: 99%
See 1 more Smart Citation
“…This means, roughly speaking, that bad hubs are expected in complex data, such as drug-target interaction data. For a more detailed discussion, we refer to [7].…”
Section: Ecknn: K-nearest Neighbor Regression With Error Correctionmentioning
confidence: 99%
“…[8] [28] [46], and hubness-aware classifiers have been developed, see [45] for a survey. More recently, hubness-aware regression techniques, including knearest neighbor with error correction (ECkNN), were developed that allow for predictions on a continuous scale [7]. Despite the fact that hubness-aware techniques are among the most promising recent machine learning approaches, their potential to enhance drug-target interaction prediction methods has not been exploited yet: to the best of our knowledge, our initial work [6] is the only one aiming to apply hubness-aware models to the drug-target prediction problem.…”
Section: Introductionmentioning
confidence: 99%
“…In order to keep the example simple, we use k = 1 nearest neighbour to calculate the corrected labels of training instances. In the figure, directed edges point from each instance to its first nearest For more details about ECkNN we refer to [18]. As mentioned previously, the dynamics of typing is captured by time series data.…”
Section: Nearest Neighbour Regression With Error Correctionmentioning
confidence: 99%
“…We set k = 5 for ECkNN which is in accordance with other works on hubness-aware machine learning [18,22].…”
Section: Evaluation Of Pairwise Modelsmentioning
confidence: 99%
“…Hubness has been identified as a detrimental factor in similarity-based machine learning, impairing several classification [44], clustering [47,60], regression [7], graph analysis [22], visualization [17], and outlier detection [18,19,45] methods. Reports on affected tasks include multimedia retrieval [51], recommendation [48], collaborative filtering [25,34], speaker verification [50], speech recognition [62], and image data classification [58].…”
Section: Introductionmentioning
confidence: 99%