2016
DOI: 10.5120/ijca2016908317
|View full text |Cite
|
Sign up to set email alerts
|

Literature Review on Feature Selection Methods for High-Dimensional Data

Abstract: Feature selection plays a significant role in improving the performance of the machine learning algorithms in terms of reducing the time to build the learning model and increasing the accuracy in the learning process. Therefore, the researchers pay more attention on the feature selection to enhance the performance of the machine learning algorithms. Identifying the suitable feature selection method is very essential for a given machine learning task with highdimensional data. Hence, it is required to conduct t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
32
0
3

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 68 publications
(43 citation statements)
references
References 72 publications
0
32
0
3
Order By: Relevance
“…The random search examines feature space in a random manner. It can begin with a random feature or specified feature and add features randomly to get the best subset found [37][38][39].…”
Section: Feature Selectionmentioning
confidence: 99%
“…The random search examines feature space in a random manner. It can begin with a random feature or specified feature and add features randomly to get the best subset found [37][38][39].…”
Section: Feature Selectionmentioning
confidence: 99%
“…One of the challenges in data mining is high dimensional data analysis [1][2][3][4][5][6][7]. Having a small sample set adds to the difficulty of the problem.…”
Section: Introductionmentioning
confidence: 99%
“…Generally, high-dimensional remotely sensed datasets contain irrelevant information and highly redundant features. Such dimensionality deteriorates quantitative (e.g., leaf area index and biomass) and qualitative (e.g., land-cover) performance of statistical algorithms by overfitting data [10]. High dimensional data are often associated with the Hughes effects or the curse of dimensionality, a phenomenon that occurs when the number of features in a dataset is greater than the number of samples [11,12].…”
Section: Introductionmentioning
confidence: 99%
“…There are two main components of dimension reduction strategies: feature extraction or construction and feature selection or feature ranking. Feature extraction (e.g., Principle Component Analysis (PCA)), constructs a new and low dimensional feature space using linear or non-linear combinations of the original high-dimensional feature space [14] while feature selection (e.g., Fisher Score and Information Gain) extracts subsets from existing features [10]. Although feature extraction methods produce higher classification accuracies, the interpretation of generated results is often challenging [2].…”
Section: Introductionmentioning
confidence: 99%