2011
DOI: 10.1007/s10472-011-9228-z
|View full text |Cite
|
Sign up to set email alerts
|

Algorithm portfolio selection as a bandit problem with unbounded losses

Abstract: We propose a method that learns to allocate computation time to a given set of algorithms, of unknown performance, with the aim of solving a given sequence of problem instances in a minimum time. Analogous meta-learning techniques are typically based on models of algorithm performance, learned during a separate of f line training sequence, which can be prohibitively expensive. We adopt instead an online approach, named GambleTA, in which algorithm performance models are iteratively updated, and used to guide a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2013
2013
2021
2021

Publication Types

Select...
6
2
2

Relationship

0
10

Authors

Journals

citations
Cited by 24 publications
(14 citation statements)
references
References 36 publications
0
14
0
Order By: Relevance
“…The performance model is usually built by applying machine learning approaches onto a dataset reporting the algorithm performances on a comprehensive set of benchmark problem instances (with the exception of [5], using a multi-armed bandit approach). Such machine learning approaches range from k-nearest neighbors [16] to ridge regression [23], random forests [24], collaborative filtering [20,15], or learning to rank approaches [17].…”
Section: Algorithm Selectorsmentioning
confidence: 99%
“…The performance model is usually built by applying machine learning approaches onto a dataset reporting the algorithm performances on a comprehensive set of benchmark problem instances (with the exception of [5], using a multi-armed bandit approach). Such machine learning approaches range from k-nearest neighbors [16] to ridge regression [23], random forests [24], collaborative filtering [20,15], or learning to rank approaches [17].…”
Section: Algorithm Selectorsmentioning
confidence: 99%
“…Furthermore, as cameras learn, the learning problem facing the other cameras changes in response. We therefore consider that a camera is faced with a non-stationary online algorithm selection problem [Gagliolo and Schmidhuber 2011]. Our approach is to consider this as an variant of the multi-armed bandit problem [Auer et al 2002].…”
Section: Learning Efficient Configurations Using Bandit Solversmentioning
confidence: 99%
“…A prominent example is the distribution of delays in end-to-end network routing [5], a typical application domain for bandits-see, e.g., [2]. Another interesting example is the distribution of running times of heuristics for solving hard combinatorial problems [6], where bandit algorithms have been used to select the heuristics [7].…”
Section: Introductionmentioning
confidence: 99%