2019
DOI: 10.48550/arxiv.1903.08689
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Implicit Generation and Generalization in Energy-Based Models

Abstract: Energy based models (EBMs) are appealing due to their generality and simplicity in likelihood modeling, but have been traditionally difficult to train. We present techniques to scale MCMC based EBM training on continuous neural networks, and we show its success on the high-dimensional data domains of ImageNet32x32, ImageNet128x128, CIFAR-10, and robotic hand trajectories, achieving better samples than other likelihood models and nearing the performance of contemporary GAN approaches, while covering all modes o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
157
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 73 publications
(178 citation statements)
references
References 21 publications
1
157
0
Order By: Relevance
“…To put this goal in context, in recent energy models for high-dimensional data (Xie et al, 2016;Nijkamp et al, 2019;Du & Mordatch, 2019;Zhao et al, 2020;Du et al, 2020;Xie et al, 2021), sampling using MCMC quickly breaks or collapses to a mode and chains longer than a few hundred steps were not reported. Thus, evaluation in prior work relies on samples from independent MCMC chains, where in addition there are heuristics like "replay buffer" (Du & Mordatch, 2019), not supported by theory. In this work, we report FID scores obtained by single MCMC chains, the first result of its kind, which we consider as a benchmark for future works on long run MCMC chains.…”
Section: Resultsmentioning
confidence: 99%
“…To put this goal in context, in recent energy models for high-dimensional data (Xie et al, 2016;Nijkamp et al, 2019;Du & Mordatch, 2019;Zhao et al, 2020;Du et al, 2020;Xie et al, 2021), sampling using MCMC quickly breaks or collapses to a mode and chains longer than a few hundred steps were not reported. Thus, evaluation in prior work relies on samples from independent MCMC chains, where in addition there are heuristics like "replay buffer" (Du & Mordatch, 2019), not supported by theory. In this work, we report FID scores obtained by single MCMC chains, the first result of its kind, which we consider as a benchmark for future works on long run MCMC chains.…”
Section: Resultsmentioning
confidence: 99%
“…Energy Based Models. Our work is related to existing work on energy-based models [5,7,9,11,13,23,34,43,46]. Most similar to our work is that of [8], which proposes a framework of utilizing EBMs to compose several object descriptions together.…”
Section: Generated Imagementioning
confidence: 90%
“…We model each relational probability distribution utilizing an Energy-Based Model (EBM) [7,27]. EBMs are a class of unnormalized probability models, which parameterize a probability distribution p θ (x) utilizing a learned energy function E θ (x):…”
Section: Energy-based Modelsmentioning
confidence: 99%
“…(Ho et al, 2020) 6.12 90 2.64 DDIM, K = 30 (Song et al, 2020a) 5.85 30 0.88 Improved DDPM, K = 45 5.96 45 1.37 SNGAN (Miyato et al, 2018) 21.7 1 -BigGAN (cond.) (Brock et al, 2018) 14.73 1 -StyleGAN2 (Karras et al, 2020a) 8.32 1 -StyleGAN2 + ADA (Karras et al, 2020a) 2.92 1 -NVAE 23.5 1 -Glow (Kingma & Dhariwal, 2018) 48.9 1 -EBM (Du & Mordatch, 2019) 38.2 60 -VAEBM (Xiao et al, 2020) 12.2 16 -G.9 SAMPLES…”
Section: G8 Comparison To Other Classes Of Generative Modelsmentioning
confidence: 99%