2017 3rd International Conference on Big Data Computing and Communications (BIGCOM) 2017
DOI: 10.1109/bigcom.2017.48
|View full text |Cite
|
Sign up to set email alerts
|

Cluster-Based Best Match Scanning for Large-Scale Missing Data Imputation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 8 publications
0
5
0
Order By: Relevance
“…Yu et al [3] proposed a modification of the K-NN imputation algorithm, known as Cluster-Based Best Match Scanning (CBMS) in terms of improved computational complexity and improved space/memory usage with comparable level of accuracy to K-NN . Simulation was carried upon a large smart meter reading dataset and imputation testing accuracy was measured using the mean absolute deviation method.…”
Section: A Completeness Dqdmentioning
confidence: 99%
See 4 more Smart Citations
“…Yu et al [3] proposed a modification of the K-NN imputation algorithm, known as Cluster-Based Best Match Scanning (CBMS) in terms of improved computational complexity and improved space/memory usage with comparable level of accuracy to K-NN . Simulation was carried upon a large smart meter reading dataset and imputation testing accuracy was measured using the mean absolute deviation method.…”
Section: A Completeness Dqdmentioning
confidence: 99%
“…Wu and Zhu [16] proposed two main methods to deal with the problem of noisy data: 1) applying data cleansing methods to eliminate data quality issues as far as possible, and 2) make data mining applications more robust so that they can tolerate the presence of noisy data. The first method presents some drawbacks, such as: (1) data cleansing algorithms deal with only certain types of errors,(2) data cleansing cannot result into perfect data, (3) data cleansing cannot always be applied to all data sources, (4) eliminating noisy data may lead to crucial data loss for further mining/analytics and (5) the data mining/analytics algorithm cannot consider the original data source context after data cleansing has been applied. However, making data mining applications more tolerant towards the presence of noisy data is based upon a very important assumption, that there is sufficient knowledge of the type of errors that are present as part of a dataset before the actual analytics is applied.…”
Section: B Accuracy Dqdmentioning
confidence: 99%
See 3 more Smart Citations