Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2017
DOI: 10.1145/3097983.3098043
|View full text |Cite
|
Sign up to set email alerts
|

Google Vizier

Abstract: Any sufficiently complex system acts as a black box when it becomes easier to experiment with than to understand. Hence, black-box optimization has become increasingly important as systems have become more complex. In this paper we describe Google Vizier, a Google-internal service for performing black-box optimization that has become the de facto parameter tuning engine at Google. Google Vizier is used to optimize many of our machine learning models and other systems, and also provides core capabilities to Goo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
101
0
4

Year Published

2018
2018
2020
2020

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 380 publications
(105 citation statements)
references
References 12 publications
0
101
0
4
Order By: Relevance
“…2) using Google Hypertune (Golovin et al, 2017). Hypertune is an automatic function optimizer that tries to find a minimum of a function in a bounded space.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…2) using Google Hypertune (Golovin et al, 2017). Hypertune is an automatic function optimizer that tries to find a minimum of a function in a bounded space.…”
Section: Resultsmentioning
confidence: 99%
“…In designing our network, we used such a system: an early version of Google Hypertune (Golovin et al, 2017). Hypertune has two components: a learning component models the effect on network performance of various hyperparameters, and an optimization component suggests new hyperparameter settings to evaluate in an attempt to find the best setting.…”
Section: Star Methodsmentioning
confidence: 99%
“…We used the Google-Vizier system for black-box optimization (Golovin et al 2017) to automatically tune our hyperparameters, including those for the input representations (e.g., number of bins, bin width), model architecture (e.g., number of fully connected layers, number of convolutional layers, kernel size), and training (e.g., dropout probability). Each Vizier "study" trained several thousand models to find the hyperparameter configuration that maximized the area under the receiver operating characteristic curve (see Section 5.1) over the validation set.…”
Section: Implementation and Training Proceduresmentioning
confidence: 99%
“…To quantify model uncertainty for our mortality prediction task, we explore the use of deep RNN ensembles and various Bayesian RNNs. For the deep ensembles approach, we optimize for the ideal hyperparameter values for our RNN model via black-box Bayesian optimization [19], and then train M replicas of the best model. Only the random seed differs between the replicas.…”
Section: Predictive Uncertainty Distributionsmentioning
confidence: 99%