2018
DOI: 10.17694/bajece.419551
|View full text |Cite
|
Sign up to set email alerts
|

A Distributed K Nearest Neighbor Classifier for Big Data

Abstract: The K-Nearest Neighbor classifier is a well-known and widely applied method in data mining applications. Nevertheless, its high computation and memory usage cost makes the classical K-NN not feasible for today's Big Data analysis applications. To overcome the cost drawbacks of the known data mining methods, several distributed environment alternatives have emerged. Among these alternatives, Hadoop MapReduce distributed ecosystem attracted significant attention. Recently, several K-NN based classification algor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 15 publications
0
2
0
Order By: Relevance
“…The classical KNN algorithm is based on calculating the distance between the test data instances to be classified and all of the instances in the training data set and finding the closest K number of training instances. After detecting the K number of closest training instances, the KNN algorithm applies majority voting which is the process of detecting the data class with the maximum number of instances among the K selected instances [6].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The classical KNN algorithm is based on calculating the distance between the test data instances to be classified and all of the instances in the training data set and finding the closest K number of training instances. After detecting the K number of closest training instances, the KNN algorithm applies majority voting which is the process of detecting the data class with the maximum number of instances among the K selected instances [6].…”
Section: Introductionmentioning
confidence: 99%
“…Since the classical KNN algorithm is completely based on individual instance proximities, it heavily suffers from high computation cost. In addition, since the algorithm decisionmaking strategy is relying on the individual instance proximities rather than stronger class representations, the algorithm's classification accuracy is also not adequate for modern data big analysis that requires rapid and accurate classification results [6].…”
Section: Introductionmentioning
confidence: 99%