2019
DOI: 10.48550/arxiv.1904.02642
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Meta-Learning Acquisition Functions for Transfer Learning in Bayesian Optimization

Abstract: Transferring knowledge across tasks to improve data-efficiency is one of the open key challenges in the field of global black-box optimization. Readily available algorithms are typically designed to be universal optimizers and, therefore, often suboptimal for specific tasks. We propose a novel transfer learning method to obtain customized optimizers within the well-established framework of Bayesian optimization, allowing our algorithm to utilize the proven generalization capabilities of Gaussian processes. Usi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 16 publications
0
6
0
Order By: Relevance
“…We find this especially promising for meta-learning, potentially building on LEO (Rusu et al, 2018). Inspired by DCEM, other more powerful sampling-based optimizers could be made differentiable in the same way, potentially optimizers that leverage gradient-based information in the inner optimization steps (Sekhon & Mebane, 1998;Theodorou et al, 2010;Stulp & Sigaud, 2012;Maheswaranathan et al, 2018) or by also learning the hyper-parameters of structured optimizers (Li & Malik, 2016;Volpp et al, 2019;Chen et al, 2017).…”
Section: Discussionmentioning
confidence: 99%
“…We find this especially promising for meta-learning, potentially building on LEO (Rusu et al, 2018). Inspired by DCEM, other more powerful sampling-based optimizers could be made differentiable in the same way, potentially optimizers that leverage gradient-based information in the inner optimization steps (Sekhon & Mebane, 1998;Theodorou et al, 2010;Stulp & Sigaud, 2012;Maheswaranathan et al, 2018) or by also learning the hyper-parameters of structured optimizers (Li & Malik, 2016;Volpp et al, 2019;Chen et al, 2017).…”
Section: Discussionmentioning
confidence: 99%
“…A possible extension of PROFET would be to consider multi-fidelity benchmarks (Klein et al, 2017a;Kandasamy et al, 2017;Klein et al, 2017c) where cheap, but approximate fidelities of the objective function are available, e. g. learning curves or dataset subsets. Furthermore, since PROFET also provides gradient information it could serve as a training distribution for learning-to-learn approaches (Chen et al, 2017;Volpp et al, 2019).…”
Section: Comparing State-of-the-art Hpo Methodsmentioning
confidence: 99%
“…This may be done by i) updating the predictive mean and variance of the surrogate model learned in the target campaign, 42 or ii) by means of an ensemble of acquisition functions. 43,44 Prior work has argued that the latter option better aggregates information from source campaigns with varying objective value ranges 43 (see ESI † Sec. S.2.A and S.2.B for additional details).…”
Section: Related Workmentioning
confidence: 99%
“…Sets of source measurements are then sampled from these perturbed surfaces. Additional details on the generation (following the procedure proposed by Volpp et al 44 ) of source data are given in ESI † Sec. S. 4.…”
Section: Analytical Benchmarksmentioning
confidence: 99%