2022
DOI: 10.1007/s10915-022-02044-x
|View full text |Cite
|
Sign up to set email alerts
|

Solving Non-linear Kolmogorov Equations in Large Dimensions by Using Deep Learning: A Numerical Comparison of Discretization Schemes

Abstract: Non-linear partial differential Kolmogorov equations are successfully used to describe a wide range of time dependent phenomena, in natural sciences, engineering or even finance. For example, in physical systems, the Allen-Cahn equation describes pattern formation associated to phase transitions. In finance, instead, the Black-Scholes equation describes the evolution of the price of derivative investment instruments. Such modern applications often require to solve these equations in high-dimensional regimes in… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4

Relationship

3
1

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 49 publications
(145 reference statements)
0
3
0
Order By: Relevance
“…A deep neural network is a type of ML model, and when a deep network is fitted to data, this is referred to as deep learning [31]. Deep learning (DL) has shown very powerful empirical performance for solving very complex real-world problems in areas such as computer vision [32], natural language processing [33,34], speech recognition [35], recommendation systems [36], drug discovery [37], differential equations [38,39], and much more [40][41][42]. In simple words, DL can be seen a neural network [43], composed by many layers, that takes some data set D, input and targets, and learns the rules for forecasting new input data.…”
Section: Introductionmentioning
confidence: 99%
“…A deep neural network is a type of ML model, and when a deep network is fitted to data, this is referred to as deep learning [31]. Deep learning (DL) has shown very powerful empirical performance for solving very complex real-world problems in areas such as computer vision [32], natural language processing [33,34], speech recognition [35], recommendation systems [36], drug discovery [37], differential equations [38,39], and much more [40][41][42]. In simple words, DL can be seen a neural network [43], composed by many layers, that takes some data set D, input and targets, and learns the rules for forecasting new input data.…”
Section: Introductionmentioning
confidence: 99%
“…The MIS is important for applications in computer science, operations research, and engineering via such uses as graph coloring, assigning channels to the radio stations, register allocation in a compiler, artificial intelligence etc. [5][6][7][8].…”
Section: Introductionmentioning
confidence: 99%
“…If one wants to use the MC method for computing the global minimum, one can either run the algorithm at T = 0 or change slowly the temperature from an initial value to T = 0 : this is the so-called Simulated Annealing algorithm 14 . A key property that allows for a solid theory of MC is the so-called detailed balance condition that ensures the algorithm admits a limiting distribution at large times 15,16 .SGD [17][18][19] is a popular optimization algorithm used in the development of state-of-the-art machine learning 20 and deep learning models 21,22 , which have shown tremendous success in numerous fields, becoming indispensable tools for many advanced applications [23][24][25][26][27][28][29][30][31] . It is an extension of the gradient descent algorithm 32 that uses a subset of the training data to compute the gradient of the objective function at each iteration.…”
mentioning
confidence: 99%
“…SGD [17][18][19] is a popular optimization algorithm used in the development of state-of-the-art machine learning 20 and deep learning models 21,22 , which have shown tremendous success in numerous fields, becoming indispensable tools for many advanced applications [23][24][25][26][27][28][29][30][31] . It is an extension of the gradient descent algorithm 32 that uses a subset of the training data to compute the gradient of the objective function at each iteration.…”
mentioning
confidence: 99%