2021
DOI: 10.48550/arxiv.2111.12140
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Filter Methods for Feature Selection in Supervised Machine Learning Applications -- Review and Benchmark

Abstract: The amount of data for machine learning (ML) applications is constantly growing. Not only the number of observations, especially the number of measured variables (features) increases with ongoing digitization. Selecting the most appropriate features for predictive modeling is an important lever for the success of ML applications in business and research. Feature selection methods (FSM) that are independent of a certain ML algorithm-so-called filter methods-have been numerously suggested, but little guidance fo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 37 publications
0
0
0
Order By: Relevance
“…The methods can be divided into methods related to the data model so-called wrapper methods (e.g. recursive feature elimination, heuristic methods) or embedded methods (Random Forests, LASSO, Ridge Regression) [Lal et al 2006] and model-independent filtering methods [Sánchez-Maroño et al 2007;Hopf 2021] e.g. based on variable correlation or mutual information (MI) measures [Vergara 2014;Gajowniczek et al 2022].…”
Section: Methodsmentioning
confidence: 99%
“…The methods can be divided into methods related to the data model so-called wrapper methods (e.g. recursive feature elimination, heuristic methods) or embedded methods (Random Forests, LASSO, Ridge Regression) [Lal et al 2006] and model-independent filtering methods [Sánchez-Maroño et al 2007;Hopf 2021] e.g. based on variable correlation or mutual information (MI) measures [Vergara 2014;Gajowniczek et al 2022].…”
Section: Methodsmentioning
confidence: 99%
“…2006] oraz niezależne od modelu metody filtrowania [Sánchez-Maroño i in. 2007;Hopf 2021] np. oparte na korelacji zmiennych czy mierze informacji wzajemnej (MI) [Vergara 2014;Gajowniczek i in.…”
Section: Selekcja Zmiennychunclassified
“…Regarding predictor selection techniques in supervised algorithms, filter, and wrapper methods are usually employed to forecast the most relevant predictors for good performance and to understand the model's response [56]. Filter methods assess the importance of the predictor based on the dataset, independently of the applied ML algorithm and its performance [57], for example, using mutual information (MI) [58]. The MI is calculated on the dataset to provide pre-interpretations, i.e., before training, aiming to measure the correlation between predictors and the output.…”
Section: Machine Learning For Path Loss Predictionmentioning
confidence: 99%