2009 Ninth IEEE International Conference on Data Mining 2009
DOI: 10.1109/icdm.2009.27
|View full text |Cite
|
Sign up to set email alerts
|

A Deep Non-linear Feature Mapping for Large-Margin kNN Classification

Abstract: KNN is one of the most popular data mining methods for classification, but it often fails to work well with inappropriate choice of distance metric or due to the presence of numerous class-irrelevant features. Linear feature transformation methods have been widely applied to extract classrelevant information to improve kNN classification, which is very limited in many applications. Kernels have also been used to learn powerful non-linear feature transformations, but these methods fail to scale to large dataset… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
43
0

Year Published

2010
2010
2021
2021

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 60 publications
(44 citation statements)
references
References 9 publications
1
43
0
Order By: Relevance
“….127 (8) .156(3) .123 (10) .156(3) .124 (9) .128(7) .141(5) .140 (6) .619 (2) .699(1) Arrhythmia .160 (7) .214(4) .167 (6) .229(3) .083 (10) .134 (9) .187(5) .156 (8) .346 (2) .385(1) Balance .127(7) .130(5) .145 (2) .149(1) .135 (4) .091 (9) .129 (6) .142(3) .092 (8) .089(10) Cleveland .889 (8) .897 (2) .890 (6) .897 (1) .889 (7) .895(3) .893 (5) .893(4) .806 (10) .846(9) Cmc .346 (9) .383 (2) .357 (7) .384(1) .358 (6) .341 (10) .365 (5) .371(4) .356 (8) .380(3) Credit .888 (7) .895 (2) .887 (8) .894(3) .891 (5) .903(1) .891 (6) .893 (4) .871 (9) .868(10) Ecoli .943(3) .948 (1) .938 (5) .941 (4) .926 (7) .920 (8) .945 (2) .933 (6) .566 (10) .584(9) German .535 (7) .541 (2) .533 …”
Section: (7) 484(1) 434(6) 475(4) 479(2) 343(8) 454(5) 477(mentioning
confidence: 97%
See 2 more Smart Citations
“….127 (8) .156(3) .123 (10) .156(3) .124 (9) .128(7) .141(5) .140 (6) .619 (2) .699(1) Arrhythmia .160 (7) .214(4) .167 (6) .229(3) .083 (10) .134 (9) .187(5) .156 (8) .346 (2) .385(1) Balance .127(7) .130(5) .145 (2) .149(1) .135 (4) .091 (9) .129 (6) .142(3) .092 (8) .089(10) Cleveland .889 (8) .897 (2) .890 (6) .897 (1) .889 (7) .895(3) .893 (5) .893(4) .806 (10) .846(9) Cmc .346 (9) .383 (2) .357 (7) .384(1) .358 (6) .341 (10) .365 (5) .371(4) .356 (8) .380(3) Credit .888 (7) .895 (2) .887 (8) .894(3) .891 (5) .903(1) .891 (6) .893 (4) .871 (9) .868(10) Ecoli .943(3) .948 (1) .938 (5) .941 (4) .926 (7) .920 (8) .945 (2) .933 (6) .566 (10) .584(9) German .535 (7) .541 (2) .533 …”
Section: (7) 484(1) 434(6) 475(4) 479(2) 343(8) 454(5) 477(mentioning
confidence: 97%
“…All experiments are carried out using 5×2 folds cross-validations, and the final results are the average of the repeated runs. (8) .037(4) .036 (6) .043(1) .040(3) .036 (5) .035(7) .042(2) .024 (10) .025(9) Churn .101(7) .113(2) .101 (6) .115(1) .108(4) .100 (8) .107(5) .111(3) .092 (10) .099(9) Upselling .219 (8) .243 (5) .218 (9) .241(6) .288(3) .212 (10) .231(7) .264(4) .443(1) .437(2) Ada.agnostic .641 (9) .654(5) .646 (8) .652 (6) .689(3) .636 (10) .648 (7) .670(4) .723(1) .691(2) Ada.prior .645 (8) .669 (2) .654(7) .668(3) .661 (5) .639 (9) .657 (6) .664(4) .682(1) .605(10) Sylva.agnostic .930 (2) .926 (8) .930(3) .925 (9) .928(6) .922 (10) .928(4) .926 (7) .934(1) .928(5) Sylva.prior .965(4) .965(2) .965 (6) .965(4) .904 (10) .974(1) .965(3) .935 (9) .946 (8) .954(7) BrazilTourism .176 (9) .242(1) .232(5) .241 (2) .233(4) .184 (8) .209 (6) .237(3) .152 (10) .199 (7) Marketing .112 (10) .157 (2) .113 (9) .161(1) .124 (8) .150(3) .134(5) .142(4) .130 (6) .125(7) Backache .311(7) .325(3) .307 (8) .328(2) .317 …”
Section: Experiments and Analysismentioning
confidence: 97%
See 1 more Smart Citation
“…It depends on a metric used as a distance to measure how similar are the data points. Here we consider the Euclidean distance [11]. The data matrix N (test set) which is j x l, where j=100 and l=4.…”
Section: Discussionmentioning
confidence: 99%
“…Such a stage involves a classification step using the well-known K-Nearest Neighbors (K-NN) multilabel classifier that depends of a metric used for the calculus of the Euclidian distance between two points for select the k neighbors nearest in terms of p attributes [11]. As well, due to the high dimension of acquired data (four dimensions) , we perform a DR step based on Principal Component Analysis (PCA) [8].…”
Section: Introductionmentioning
confidence: 99%