2008
DOI: 10.1016/j.knosys.2008.03.013
|View full text |Cite
|
Sign up to set email alerts
|

A selective Bayes Classifier for classifying incomplete data based on gain ratio

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
12
0

Year Published

2011
2011
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 30 publications
(12 citation statements)
references
References 5 publications
(6 reference statements)
0
12
0
Order By: Relevance
“…In addition, some other classifiers have a better performance compared with previously published results, for example, best‐first, J48, and REPT classifiers in the analysis of TH dataset; RF classifier in the analysis of BW dataset; RS classifier in the analysis of HC dataset; MLP and RBF classifiers in the analysis of HH dataset; RBF, SMO, and RF classifiers in the analysis of SH dataset; RBF and RAF classifiers in the analysis of HE dataset; SMO classifier in the analysis of LY dataset; J48 in the analysis of BC dataset; NBU classifier in the analysis of HS dataset; SC classifier in the analysis of HC dataset; RF classifier in the analysis of AR dataset; RS, STC, and zeroR classifiers in the analysis of PO dataset; and NBU in the analysis of PT dataset (Table ). Furthermore, some combinations of feature extraction and classification methods, hybrid‐based, and evolutionary learning‐based classification methods in the published literature have a better accuracy than the selected classification methods in the analysis of selected healthcare data, such as AISWNB, CFSWNB, GRWNB, MIWNB, ReFWNB, Tree‐WNB, and RMWNB (Wu et al, ), LWNB and AODE (Jiang, Zhang, et al, ), pedagogical, decompositional, SVMs, a combination of HC and pedagogical (Stoean & Stoean, ), PSO, ABC, and GSA (Bahrololoum, Nezamabadi‐Pour, Bahrololoum, & Saeed, ), and RBF‐MS, RBF‐HS, and RBF‐NNTS (Jaganathan & Kuppuchamy, ) in the analysis of BW dataset; a combination of SVM and feature selection methods (Sun et al, ) in the analysis of DE dataset; a combination of PPPCA and SVM (Shah et al, ) in the analysis of HH dataset; CDW‐NN (Paredes & Vidal, ) and ABC (Schiezaro & Pedrini, ) in the analysis of SH dataset; RF (Azar, Elshazly, Hassanien, & Elkorany, ) in the analysis of LY dataset; NPBC (Soria, Garibaldi, Ambrogi, Biganzoli, & Ellis, ) in the analysis of HS dataset; SRBC and SRBCBG (Chen et al, ) in the analysis of LC dataset; FND (Rodríguez, García‐Osorio, & Maudes, ), a combination of SVM and feature selection methods (Sun et al, ), and SRBC and SRBCBG (Chen et al, ) in the analysis of AR dataset; and PELM (Rong, Ong, Tan, & Zhu, ) in the analysis of LY dataset (Table ). It is also important to note that in the analysis of some datasets, such as TH, BW, and DE, most of the classification methods have a better performance, whereas for some other datasets, such as PO, LI, and PT, a poor performance is observed.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition, some other classifiers have a better performance compared with previously published results, for example, best‐first, J48, and REPT classifiers in the analysis of TH dataset; RF classifier in the analysis of BW dataset; RS classifier in the analysis of HC dataset; MLP and RBF classifiers in the analysis of HH dataset; RBF, SMO, and RF classifiers in the analysis of SH dataset; RBF and RAF classifiers in the analysis of HE dataset; SMO classifier in the analysis of LY dataset; J48 in the analysis of BC dataset; NBU classifier in the analysis of HS dataset; SC classifier in the analysis of HC dataset; RF classifier in the analysis of AR dataset; RS, STC, and zeroR classifiers in the analysis of PO dataset; and NBU in the analysis of PT dataset (Table ). Furthermore, some combinations of feature extraction and classification methods, hybrid‐based, and evolutionary learning‐based classification methods in the published literature have a better accuracy than the selected classification methods in the analysis of selected healthcare data, such as AISWNB, CFSWNB, GRWNB, MIWNB, ReFWNB, Tree‐WNB, and RMWNB (Wu et al, ), LWNB and AODE (Jiang, Zhang, et al, ), pedagogical, decompositional, SVMs, a combination of HC and pedagogical (Stoean & Stoean, ), PSO, ABC, and GSA (Bahrololoum, Nezamabadi‐Pour, Bahrololoum, & Saeed, ), and RBF‐MS, RBF‐HS, and RBF‐NNTS (Jaganathan & Kuppuchamy, ) in the analysis of BW dataset; a combination of SVM and feature selection methods (Sun et al, ) in the analysis of DE dataset; a combination of PPPCA and SVM (Shah et al, ) in the analysis of HH dataset; CDW‐NN (Paredes & Vidal, ) and ABC (Schiezaro & Pedrini, ) in the analysis of SH dataset; RF (Azar, Elshazly, Hassanien, & Elkorany, ) in the analysis of LY dataset; NPBC (Soria, Garibaldi, Ambrogi, Biganzoli, & Ellis, ) in the analysis of HS dataset; SRBC and SRBCBG (Chen et al, ) in the analysis of LC dataset; FND (Rodríguez, García‐Osorio, & Maudes, ), a combination of SVM and feature selection methods (Sun et al, ), and SRBC and SRBCBG (Chen et al, ) in the analysis of AR dataset; and PELM (Rong, Ong, Tan, & Zhu, ) in the analysis of LY dataset (Table ). It is also important to note that in the analysis of some datasets, such as TH, BW, and DE, most of the classification methods have a better performance, whereas for some other datasets, such as PO, LI, and PT, a poor performance is observed.…”
Section: Discussionmentioning
confidence: 99%
“…Combinations of feature extraction and classification methods and different classification methods have been also implemented in healthcare data classification, for instance, the combination of radial basis function (RBF) neural network, mean selection, half selection, and neural network for threshold selection (NNTS)‐based feature selection method in the classification of BW dataset (accuracy of 95.85–97.28%), PI dataset (accuracy of 73.83–76.04%), HS dataset (accuracy of 84.44–85.19%), HE dataset (accuracy of 82.58–85.16%), and HC dataset (accuracy of 81.75–84.46%; Jaganathan & Kuppuchamy, ); combination of random forest (RAF) and J48 in the classification of BW dataset (accuracy of 97.31%) and SH dataset (accuracy of 83.68%; Tan et al, ); combination of principal component analysis (PCA) and SVM, and kernel PCA and SVM in the classification of BW and PT datasets (maximum accuracy of 96.35% for BW dataset using PCA + SVM method; Li, Liu, & Hu, ); combination of probabilistic PCA and SVM in the classification of HC and HH datasets (maximum accuracy of 85.82% for HH dataset; Shah et al, ), and combinations of C4.5 with standard, ad hoc, and pairwise learning in the classification of BW and PI datasets (maximum accuracy of 44.89% for PO dataset; Fernández et al, 2013). Besides, some novel and hybrid classification methods have been developed in healthcare data classification, for example, artificial immune system‐based self‐adaptive attribute weighting method for NB in the classification of TH, AU, BC, BW, HC, HH, HS, HE, LY, and PT datasets (maximum accuracy of 75.49 ± 8% for AU dataset; Wu et al, ); superparent‐one‐dependence estimator method for classification of TH, AU, BC, BW, PI, HC, HH, HS, HE, LY, and PT datasets (maximum accuracy of 48.38% for PT dataset; Wu, Pan, Zhu, Zhang, & Zhang, ); a two‐stage evolutionary algorithm in the classification of DE, BC, LY, and PI datasets (maximum accuracy of 69.34 ± 2.30% in the classification of BC dataset; Gutiérrez, Hervás‐Martínez, Martínez‐Estudillo, & Carbonero, ); and selective robust Bayes classifier (SRBC) and SRBC for incomplete data based on a gain ratio (SRBCBG) in the classification of AR, AU, BC, and LC datasets (maximum accuracy of 76.08 ± 0.74% for AR dataset; Chen, Huang, Tian, & Tian, ).…”
Section: Introductionmentioning
confidence: 99%
“…To improve the naive Bayes performance, many different approaches have been proposed in literature [7,8,11]. Solutions like the Extended Naive Bayes [10] and the NB+ [12] are only two examples of techniques being validated over a number of data sets to show a better classification accuracy than the traditional naive Bayes.…”
Section: Discussionmentioning
confidence: 99%
“…According to Chen and colleagues [11] methods of constructing classifiers for incomplete data deserve more attention. Classifiers such as naive Bayes classifiers and C4.5 often adopt two simple strategies to deal with incomplete data: to ignore the instances with unknown entries or to ascribe these unknown entries to a specified dummy value of the respective attribute variables.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation