2020
DOI: 10.48550/arxiv.2007.14660
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Ergodicity of the underdamped mean-field Langevin dynamics

Anna Kazeykina,
Zhenjie Ren,
Xiaolu Tan
et al.

Abstract: We study the long time behavior of an underdamped mean-field Langevin (MFL) equation, and provide a general convergence as well as an exponential convergence rate result under different conditions. The results on the MFL equation can be applied to study the convergence of the Hamiltonian gradient descent algorithm for the overparametrized optimization. We then provide a numerical example of the algorithm to train a generative adversarial networks (GAN).

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(10 citation statements)
references
References 30 publications
0
10
0
Order By: Relevance
“…( 8)), and the convergence rate of which is more difficult to establish. (Monmarché, 2017;Guillin et al, 2019;Kazeykina et al, 2020;Guillin et al, 2021), based on hypocoercivity (Villani, 2009) or coupling techniques (Eberle et al, 2019). In addition, Bou-Rabee and Schuh (2020); Bou-Rabee and Eberle (2021) studied the dis-crete time convergence of Hamiltonian Monte Carlo with interaction potential.…”
Section: Related Literaturementioning
confidence: 99%
See 1 more Smart Citation
“…( 8)), and the convergence rate of which is more difficult to establish. (Monmarché, 2017;Guillin et al, 2019;Kazeykina et al, 2020;Guillin et al, 2021), based on hypocoercivity (Villani, 2009) or coupling techniques (Eberle et al, 2019). In addition, Bou-Rabee and Schuh (2020); Bou-Rabee and Eberle (2021) studied the dis-crete time convergence of Hamiltonian Monte Carlo with interaction potential.…”
Section: Related Literaturementioning
confidence: 99%
“…Recent works have studied the convergence rate of the mean field Langevin dynamics, including its underdamped (kinetic) version. However, most existing analyses either require sufficiently strong regularization (Hu et al, 2019;Jabir et al, 2019), or build upon involved mathematical tools (Kazeykina et al, 2020;Guillin et al, 2021). Our goal is to provide a simpler convergence proof that covers general and more practical machine learning settings, with a focus on neural network optimization in the mean field regime.…”
Section: Introductionmentioning
confidence: 99%
“…The particle interacts with its environment through the non-local term K * ρ whilst being confined by an external potential g, experiencing a frictional force f and a diffusive force Wt. We have just presented a statistical physics interpretation of (5.3), however, more recently there has been a resurgence of interest in such equations among those in the machine learning community who want to rigorously prove trainability of neural networks, see [KRTY20] and references therein.…”
Section: Vlasov-fokker-planck Equation (Vfpe)mentioning
confidence: 99%
“…Algorithms using Langevin dynamics have a better long-time behaviour compared to the overdamped Langevin dynamics [15,16], which forms a degenerated special case of the Langevin dynamics, where the limit for γ to infinity is taken [41,Section 6.5.1]. Therefore, nonlinear Langevin dynamics became recently popular for training networks as the Generative Adversarial Network (GAN) [33].…”
Section: Introductionmentioning
confidence: 99%