2021
DOI: 10.1007/s00498-021-00309-8
|View full text |Cite
|
Sign up to set email alerts
|

Parameter calibration with stochastic gradient descent for interacting particle systems driven by neural networks

Abstract: We propose a neural network approach to model general interaction dynamics and an adjoint-based stochastic gradient descent algorithm to calibrate its parameters. The parameter calibration problem is considered as optimal control problem that is investigated from a theoretical and numerical point of view. We prove the existence of optimal controls, derive the corresponding first-order optimality system and formulate a stochastic gradient descent algorithm to identify parameters for given data sets. To validate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 10 publications
(10 citation statements)
references
References 22 publications
0
9
0
Order By: Relevance
“…solutions of partial differential equations (42,43) or parameter estimation of multiagent models (44,45). Neural networks, and especially deep neural networks, are mathematically little understood, and their theoretical underpinnings are sparse and mainly restricted to shallow networks (networks with only one hidden layer) (46,47).…”
Section: Significancementioning
confidence: 99%
See 1 more Smart Citation
“…solutions of partial differential equations (42,43) or parameter estimation of multiagent models (44,45). Neural networks, and especially deep neural networks, are mathematically little understood, and their theoretical underpinnings are sparse and mainly restricted to shallow networks (networks with only one hidden layer) (46,47).…”
Section: Significancementioning
confidence: 99%
“…More recently, a promising new method has emerged in the form of artificial neural nets. Neural networks have of course prominently been used as powerful pattern-recognition devices and predictive models ( 41 ), but, as they become more and more accessible to the scientific community at large, researchers are beginning to apply their computational capabilities across the mathematical disciplines, including in fields heretofore dominated by more classical methods: Examples include finding solutions of partial differential equations ( 42 , 43 ) or parameter estimation of multiagent models ( 44 , 45 ). Neural networks, and especially deep neural networks, are mathematically little understood, and their theoretical underpinnings are sparse and mainly restricted to shallow networks (networks with only one hidden layer) ( 46 , 47 ).…”
mentioning
confidence: 99%
“…In the context of calibration, the global maximum of the objective function is sought such that the model optimally matches the computational constraints. The CMA-ES algorithm, as an evolution strategy, is more suited to find such a global maximum compared to gradient-based approaches which are more likely to converge to local maxima, such as gradient descent (Galtier and Wainrib, 2013) or stochastic gradient descent (Göttlich and Totzeck, 2021). Figure 4 shows an example of CMA-ES maximization starting from an initial guess at iteration 0 until converging to one of two global maxima at iteration 20.…”
Section: Objective Function Maximization Illustrationmentioning
confidence: 99%
“…In general, this is done by using an training algorithm like the back-propagation algorithm [241]. The aim of this training algorithm is to adjust the network settings in a way, that minimizes the given cost function [242]. Common cost functions are the mean square error or the cross entropy.…”
Section: A Technically Oriented Comparisonmentioning
confidence: 99%