2018 24th International Conference on Pattern Recognition (ICPR) 2018
DOI: 10.1109/icpr.2018.8545422
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic Ensemble Active Learning: A Non-Stationary Bandit with Expert Advice

Abstract: Active learning aims to reduce annotation cost by predicting which samples are useful for a human teacher to label. However it has become clear there is no best active learning algorithm. Inspired by various philosophies about what constitutes a good criteria, different algorithms perform well on different datasets. This has motivated research into ensembles of active learners that learn what constitutes a good criteria in a given scenario, typically via multi-armed bandit algorithms. Though algorithm ensemble… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(16 citation statements)
references
References 29 publications
(74 reference statements)
0
16
0
Order By: Relevance
“…Their approach performed in line with the best two individual query strategies and outperformed Hsu and Lin (2015). Pang et al (2018) proposed a modification of multi-armed bandits with experts, to account for non-stationary loss functions (i.e., the best expert might vary over time), in binary and multi-class classification tasks. Their approach outperformed or performed in line with the best individual strategies and outperformed both Baram, El-Yaniv, and Luz (2004) and Hsu and Lin (2015) in non-stationary datasets.…”
Section: Combining Online Learning and Active Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…Their approach performed in line with the best two individual query strategies and outperformed Hsu and Lin (2015). Pang et al (2018) proposed a modification of multi-armed bandits with experts, to account for non-stationary loss functions (i.e., the best expert might vary over time), in binary and multi-class classification tasks. Their approach outperformed or performed in line with the best individual strategies and outperformed both Baram, El-Yaniv, and Luz (2004) and Hsu and Lin (2015) in non-stationary datasets.…”
Section: Combining Online Learning and Active Learningmentioning
confidence: 99%
“…Since our end goal is to improve the MT ensemble, we measure its improvement at each weight update by considering the expected regret R M for not choosing the best MT system's translation at each iteration, up to the current iteration T (Eq. 9), which can be seen as a dynamic regret (Pang et al 2018). Note that this regret formulation deviates from the traditional formulation, in that we compare the forecaster to the best sequence of decisions overall (whose cumulative loss is given by T t=1 min j=1,...,J j,t ), instead of the best expert overall (whose cumulative loss would be given by min j=1,...,J T t=1 j,t ).…”
Section: Figurementioning
confidence: 99%
“…Many other data-driven approaches for pool-based AL processes have been proposed recently. While Bachman et al (2017) and Pang et al (2018b) used RL to build the learning model, Liu et al (2018) formulated learning AL strategies as an imitation learning problem (i.e., the machine is trained to perform a task from demonstrations by learning a mapping between observations and actions), Contardo et al (2017) and Ravi and Larochelle (2018) Pang et al (2018a) extended the LSA approach using non-stationary multi-armed bandit with expert advice.…”
Section: Strategy-free Approachesmentioning
confidence: 99%
“…Conversely, selecting an example in an already sampled region allows to locally refine the predictive model. We do not intend to provide an exhaustive overview of existing AL strategies and refer to [37], [38] for a detailed overview, [39]- [41] for some recent benchmark and a new way to treat uncertainty in [42] Another meta active learning paradigm exists, which combines conventional strategies using bandit algorithms [43]- [48]. These meta-learning algorithms intend to select online the best AL strategy according to the observed improvements of the classifier.…”
Section: B Axis 2: Inexact Supervision -Labels At the Right Proxy Vs ...mentioning
confidence: 99%