2019
DOI: 10.48550/arxiv.1909.02102
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Accelerated Information Gradient flow

Abstract: We present a systematic framework for the Nesterov's accelerated gradient flows in the spaces of probabilities embedded with information metrics. Here two metrics are considered, including both the Fisher-Rao metric and the Wasserstein-2 metric. For the Wasserstein-2 metric case, we prove the convergence properties of the accelerated gradient flows, and introduce their formulations in Gaussian families. Furthermore, we propose a practical discrete-time algorithm in particle implementations with an adaptive res… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5

Relationship

3
2

Authors

Journals

citations
Cited by 6 publications
(11 citation statements)
references
References 37 publications
0
11
0
Order By: Relevance
“…demonstrate that underdamped LD accelerates the steepest descent steps taken by the overdamped LD, forming an analog of Nesterov acceleration for MCMC methods. Wang and Li (2019) present a framework for Nesterov's accelerated gradient method in the Wasserstein space, which consists of augmenting the energy functional with the kinetic energy of an additional momentum variable.…”
Section: Stein's Methods and Other Relevant Workmentioning
confidence: 99%
“…demonstrate that underdamped LD accelerates the steepest descent steps taken by the overdamped LD, forming an analog of Nesterov acceleration for MCMC methods. Wang and Li (2019) present a framework for Nesterov's accelerated gradient method in the Wasserstein space, which consists of augmenting the energy functional with the kinetic energy of an additional momentum variable.…”
Section: Stein's Methods and Other Relevant Workmentioning
confidence: 99%
“…where h is computed at the current samples w n l , e.g., as the median of their square distances (Liu and Wang, 2016) or through optimization (Wang and Li, 2019). For the step size α n l in ( 16), we use a line search technique (Chen et al, 2019b;.…”
Section: Projected Wasserstein Gradient Descentmentioning
confidence: 99%
“…The first example is a bi-modal posterior distribution with a Gaussian prior. WGD-MED and WGD-BM denote WGD with kernel bandwidth calculated by the MED method (Liu and Wang, 2016) and the BM method (Wang and Li, 2019) respectively. We compare WGD-MED, WGD-BM with SVGD.…”
Section: 1mentioning
confidence: 99%
“…We next present the following two categories of gradient flows in Hessian density manifold. Firstly, we introduce a class of transport Newton's flows [36].…”
Section: Proof Of Claimmentioning
confidence: 99%
“…In this paper, we extend the area of TIG into the category of Hessian geometry. One direct application is to formulate optimization techniques for Bayesian sampling problems [36,37]. See related developments in information geometry [32,33].…”
Section: Introductionmentioning
confidence: 99%