2022
DOI: 10.48550/arxiv.2201.10469
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Convex Analysis of the Mean Field Langevin Dynamics

Abstract: As an example of the nonlinear Fokker-Planck equation, the mean field Langevin dynamics recently attracts attention due to its connection to (noisy) gradient descent on infinitely wide neural networks in the mean field regime, and hence the convergence property of the dynamics is of great theoretical interest. In this work, we give a simple and self-contained convergence rate analysis of the mean field Langevin dynamics with respect to the (regularized) objective function in both continuous and discrete time s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(14 citation statements)
references
References 28 publications
0
14
0
Order By: Relevance
“…• Under these assumptions, the convergence of µ s to µ * in relative entropy and in Wasserstein distance, also holds, with the same rate [27,9].…”
Section: Exponential Convergencementioning
confidence: 85%
See 2 more Smart Citations
“…• Under these assumptions, the convergence of µ s to µ * in relative entropy and in Wasserstein distance, also holds, with the same rate [27,9].…”
Section: Exponential Convergencementioning
confidence: 85%
“…Let us recall the statement of Theorem 3.3 and prove it, by an application of a convergence result proved independently in [27] and [9].…”
Section: Proof Of Theorem 33mentioning
confidence: 94%
See 1 more Smart Citation
“…Upon completion of this work, we became aware of the preprint Nitanda et al [2022] which also proves the exponential convergence of the Mean-Field Langevin dynamics with the same proof technique. Our work was conducted independently and simultaneously, and their contribution is not reflected in the present version of our paper (beyond this paragraph).…”
Section: Contributions and Related Workmentioning
confidence: 95%
“…We shall note that two recent independent works (Chizat (2022); Nitanda et al (2022)) also tried to prove the linear convergence result for neural networks trained by noisy SGD in the meanfield regime, but as pointed out by the authors (Chizat (2022)), their assumptions of boundedness and smoothness in both works cannot be applied to the vanilla two-layer neural networks (we will discuss details in later sections). Our work differs from them in both assumptions and proof techniques.…”
Section: Related Workmentioning
confidence: 99%