2019
DOI: 10.48550/arxiv.1909.12830
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Differentiable Cross-Entropy Method

Brandon Amos,
Denis Yarats

Abstract: We study the Cross-Entropy Method (CEM) for the non-convex optimization of a continuous and parameterized objective function and introduce a differentiable variant (DCEM) that enables us to differentiate the output of CEM with respect to the objective function's parameters. In the machine learning setting this brings CEM inside of the end-to-end learning pipeline where this has otherwise been impossible. We show applications in a synthetic energy-based structured prediction task and in non-convex continuous co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(12 citation statements)
references
References 48 publications
0
12
0
Order By: Relevance
“…Recently, [2,6] showed how to efficiently differentiate through convex cone programs by applying the implicit function theorem to a residual map introduced in [27], and [1] showed how to differentiate through convex optimization problems by an automatable reduction to convex cone programs; our method for learning convex optimization models builds on this recent work. Optimization layers have been used in many applications, including control [7,11,15,3], game-playing [46,45], computer graphics [37], combinatorial tasks [58,52,53,21], automatic repair of optimization problems [14], and data fitting more generally [9,17,16,10]. Differentiable optimization for nonconvex problems is often performed numerically by differentiating each individual step of a numerical solver [33,48,32,36], although sometimes it is done implicitly; see, e.g., [7,47,4].…”
Section: Related Workmentioning
confidence: 99%
“…Recently, [2,6] showed how to efficiently differentiate through convex cone programs by applying the implicit function theorem to a residual map introduced in [27], and [1] showed how to differentiate through convex optimization problems by an automatable reduction to convex cone programs; our method for learning convex optimization models builds on this recent work. Optimization layers have been used in many applications, including control [7,11,15,3], game-playing [46,45], computer graphics [37], combinatorial tasks [58,52,53,21], automatic repair of optimization problems [14], and data fitting more generally [9,17,16,10]. Differentiable optimization for nonconvex problems is often performed numerically by differentiating each individual step of a numerical solver [33,48,32,36], although sometimes it is done implicitly; see, e.g., [7,47,4].…”
Section: Related Workmentioning
confidence: 99%
“…Relation to Differentiable Cross-Entropy: Particular importance should be given to a recent paper [9], since, to the best of our knowledge, is the first to suggest sampling-based optimization instead of gradient descent, and features some similarities with our approach. The authors in [9] propose a differentiable approximation of the cross-entropy method (CEM) [21,22], called differentiable cross-entropy (DCEM). To obtain this approximation, they need to approximate CEM's eliteness threshold operation, which is non-differentiable.…”
Section: Further Background and Related Workmentioning
confidence: 99%
“…With respect to the latter, a significant amount of attention has been devoted to incorporating optimization blocks or modules operating at some part of the network. This has been motivated by large number of applications, including meta-learning [1][2][3], differentiable physics simulators [4], classification [5], GANs [6], reinforcement learning with constraints, latent spaces, or safety [7][8][9][10], model predictive control [11,12], as well as tasks relying on the use of energy networks [13,3], among many others. Local 2 optimization modules lead to nested optimization operations, as they interact with the global, end-to-end training of the network that contains them.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations