2019
DOI: 10.1016/j.is.2019.02.003
|View full text |Cite
|
Sign up to set email alerts
|

Boosting decision stumps for dynamic feature selection on data streams

Abstract: Feature selection targets the identification of which features of a dataset are relevant to the learning task. It is also widely known and used to improve computation times, reduce computation requirements, and to decrease the impact of the curse of dimensionality and enhancing the generalization rates of classifiers. In data streams, classifiers shall benefit from all the items above, but more importantly, from the fact that the relevant subset of features may drift over time. In this paper, we propose a nove… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 30 publications
(18 citation statements)
references
References 47 publications
(67 reference statements)
0
12
0
Order By: Relevance
“…-Feature drift This is a type of change in data streams that happens when a subset of features becomes, or stops to be, relevant to the learning task (Barddal et al 2017). Additionally, new features may emerge (thus extending the feature space), while the old ones may cease to arrive (Barddal et al 2019a). Therefore, classifiers need to adapt to these changes in feature space (Barddal et al 2016) by performing a dynamic feature selection (Yuan et al 2018;Barddal et al 2019b), using randomness in selected features (Abdulsalam et al 2011), or employing a sliding window and feature space transformation (Nguyen et al 2012).…”
Section: Overviewmentioning
confidence: 99%
“…-Feature drift This is a type of change in data streams that happens when a subset of features becomes, or stops to be, relevant to the learning task (Barddal et al 2017). Additionally, new features may emerge (thus extending the feature space), while the old ones may cease to arrive (Barddal et al 2019a). Therefore, classifiers need to adapt to these changes in feature space (Barddal et al 2016) by performing a dynamic feature selection (Yuan et al 2018;Barddal et al 2019b), using randomness in selected features (Abdulsalam et al 2011), or employing a sliding window and feature space transformation (Nguyen et al 2012).…”
Section: Overviewmentioning
confidence: 99%
“…This selection method was shown to improve the performance of two different types of classifiers. Adaptive Boosting for FS (ABFS) was introduced by Barddal et al [6] and uses a combination of boosting [23] and decision stumps (a decision tree whereby the root node is connected to the terminal nodes) to select features. Boosting gives higher weights to training instances which are harder to classify, then decision stumps are used to select features from these difficult-to-classify samples.…”
Section: Related Workmentioning
confidence: 99%
“…if |AA|==m (13) \*Selecting features in the storage window*\ (14) for ii = 1 to |AA| do (15) Calculate the feature repulsion loss for each feature by Equation (12); (16) end for (17) Select the feature f ii with the largest FRL; (18) S=S∪{f ii } (19) AA=AA-{f ii }; (20) end if (21) end if (22) end for (23) end for (24) end while; (25) Until no feature groups are available; (26) Return S;…”
Section: Streaming Feature Selectionmentioning
confidence: 99%
“…Nevertheless, the aforementioned algorithms assume full knowledge of the entire feature space, while features emerge incrementally or dynamically in numerous modern applications [11,21,22]. For example, hot news is constantly updated with different keywords in each piece of news, indicating that features of every data sample cannot be necessarily available in advance.…”
Section: Introductionmentioning
confidence: 99%