2021
DOI: 10.48550/arxiv.2107.06330
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Multi-variance replica exchange stochastic gradient MCMC for inverse and forward Bayesian physics-informed neural network

Guang Lin,
Yating Wang,
Zecheng Zhang

Abstract: Physics-informed neural network (PINN) has been successfully applied in solving a variety of nonlinear non-convex forward and inverse problems. However, the training is challenging because of the non-convex loss functions and the multiple optima in the Bayesian inverse problem. In this work, we propose a multi-variance replica exchange stochastic gradient Langevin diffusion method to tackle the challenge of the multiple local optima in the optimization and the challenge of the multiple modal posterior distribu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
4
1

Relationship

4
1

Authors

Journals

citations
Cited by 5 publications
(9 citation statements)
references
References 15 publications
0
9
0
Order By: Relevance
“…However, compared to vanilla SGLD [26], the computational cost of reSGLD doubles. To reduce the computational cost of reSGLD, we adopt the idea proposed in [24] and develop m-reSGLD (Algorithm 2); that is, an algorithm that uses (1) a full training regime for the low temperature particle θ 1 , which exploits the landscape of the energy function U , and (2) a accelerated training regime for the high temperature particle θ 2 , which explores the landscape of U . To this end, we first fully train both particles for a fixed number of burn-in epochs.…”
Section: Accelerated Bayesian Training Of Deeponetsmentioning
confidence: 99%
See 2 more Smart Citations
“…However, compared to vanilla SGLD [26], the computational cost of reSGLD doubles. To reduce the computational cost of reSGLD, we adopt the idea proposed in [24] and develop m-reSGLD (Algorithm 2); that is, an algorithm that uses (1) a full training regime for the low temperature particle θ 1 , which exploits the landscape of the energy function U , and (2) a accelerated training regime for the high temperature particle θ 2 , which explores the landscape of U . To this end, we first fully train both particles for a fixed number of burn-in epochs.…”
Section: Accelerated Bayesian Training Of Deeponetsmentioning
confidence: 99%
“…They demonstrated that the samples generated by their proposed algorithm converge to the target posterior distribution. Replicaexchange MCMC and (its stochastic gradient variants) were proposed [24,5,12] to accelerate the Langevin diffusion. In replica-exchange frameworks, instead of using one particle to sample from the posterior, one uses two particles with different temperature parameters.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…One choice is to estimate each state p k (x) given a sequence of observations up to k in time by the posterior distribution. This task of sequential prediction based on the online observations can then be categorized as the standard Bayesian filtering problems which have been thoroughly studied in the control theory [10,5,24,20]. This motivates us to formulate the inversion problem in the Bayesian framework.…”
Section: Bayesian Formulation and Outlinementioning
confidence: 99%
“…Physics-informed neural network (PINN) is a neural network approach of solving the partial differential equations (PDE) [22,17,16]. The idea of the PINN is to approximate the solution of the PDE by a network.…”
Section: Introductionmentioning
confidence: 99%