2017
DOI: 10.1016/j.compbiomed.2017.09.011
|View full text |Cite
|
Sign up to set email alerts
|

Optimal feature selection using a modified differential evolution algorithm and its effectiveness for prediction of heart disease

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
39
0
1

Year Published

2018
2018
2022
2022

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 142 publications
(50 citation statements)
references
References 24 publications
0
39
0
1
Order By: Relevance
“…In this section, a survey of recent techniques is discussed which is used to predict the CV and diabetes diseases among the patients. Vivekanandan and Iyengar (2017), performed the feature selection and optimized the selected features by using modified Differential Evolution (DE) algorithm for CV disease. A Fuzzy Analytic Hierarchy Process (FAHP) and Feed-Forward Neural Network (FFNN) were used to predict the heart disease with selected critical features.…”
Section: Literature Reviewmentioning
confidence: 99%
“…In this section, a survey of recent techniques is discussed which is used to predict the CV and diabetes diseases among the patients. Vivekanandan and Iyengar (2017), performed the feature selection and optimized the selected features by using modified Differential Evolution (DE) algorithm for CV disease. A Fuzzy Analytic Hierarchy Process (FAHP) and Feed-Forward Neural Network (FFNN) were used to predict the heart disease with selected critical features.…”
Section: Literature Reviewmentioning
confidence: 99%
“…The best feature subset (called the optimal) is measured based on an evaluation condition. However, discovering the optimal feature is generally intractable this is due to the fact that the increase in dimensionality increases the number of features as well [3], [5], [13]. Numerous problems connected to FS are proved to be NP-hard.…”
Section: Feature Selectionmentioning
confidence: 99%
“…The wrapper, on the other hand, used a mining algorithm to determine the goodness of selected features, the subset that provides higher performance are selected [14]. The major drawback of the wrapper model is classifier dependency, computationally expensive and is not suitable for large datasets [3], [13].…”
Section: Feature Selectionmentioning
confidence: 99%
See 2 more Smart Citations