2020
DOI: 10.1080/00031305.2020.1790217
|View full text |Cite
|
Sign up to set email alerts
|

Feature Engineering and Selection: A Practical Approach for Predictive Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 38 publications
(21 citation statements)
references
References 6 publications
0
21
0
Order By: Relevance
“…The idea is to include only those features that significantly contribute to the ability of the algorithm to distinguish between different classes, effectively limiting the number of computations necessary to obtain an accurate prediction. Thus, the exact number of selected features may vary from one model to the next, depending on the type of algorithm used, structure and volume of raw data, and the main tasks that the machine learning classifier will be expected to complete [92].…”
Section: ) Feature Selectionmentioning
confidence: 99%
See 1 more Smart Citation
“…The idea is to include only those features that significantly contribute to the ability of the algorithm to distinguish between different classes, effectively limiting the number of computations necessary to obtain an accurate prediction. Thus, the exact number of selected features may vary from one model to the next, depending on the type of algorithm used, structure and volume of raw data, and the main tasks that the machine learning classifier will be expected to complete [92].…”
Section: ) Feature Selectionmentioning
confidence: 99%
“…In terms of the principles used to rank the features, we can talk about Filter methods (such as variance threshold, correlation coefficient, or Chi-square test) which capture some of the native properties of each feature, and Wrapper methods (i.e. forward feature selection or backward feature elimination), which measure how a proposed set of features works with a particular algorithm [92]. There are also Embedded (LASSO regularization or random forest importance) and Hybrid approaches, which combine some of the main strengths of both Filter and Wrapper methods.…”
Section: ) Feature Selectionmentioning
confidence: 99%
“…They are a set of techniques that assign a score to input features based on how useful they are at predicting a target variable. Feature importance scores play an important role in a predictive modeling project, providing an insight into data, an insight into the model, and the basis for dimensionality reduction and feature selection which can improve the efficiency and effectiveness of a predictive model on the problem [36].…”
Section: E Feature Importancementioning
confidence: 99%
“…In the first stage, four statistical measures based on correlation were used. According to Butcher and Smith (2020), inferential statistical approaches provide a better solution to appraising the contribution of a predictor to the underlying model or the dataset. These measures are Analysis of Variance (ANOVA), Pearson's Correlation Coefficient, Mutual Information, and Chi-Squared.…”
Section: Proposed Feature Engineering Approachmentioning
confidence: 99%
“…Feature selection and engineering are essential factors in machine learning as they can increase the predictive power of ML algorithms. Feature engineering involves understanding the domain knowledge of the dataset to create new features or combine existing features that make ML algorithms perform better (Butcher & Smith, 2020). Hence, the contribution of this study is the feature selection and engineering processes on the BoT-IoT dataset in an attempt to improve the performance of ML algorithms.…”
Section: Introductionmentioning
confidence: 99%