Nowadays many financial derivatives, such as American or Bermudan options, are of early exercise type. Often the pricing of early exercise options gives rise to high-dimensional optimal stopping problems, since the dimension corresponds to the number of underlying assets. High-dimensional optimal stopping problems are, however, notoriously difficult to solve due to the well-known curse of dimensionality. In this work, we propose an algorithm for solving such problems, which is based on deep learning and computes, in the context of early exercise option pricing, both approximations of an optimal exercise strategy and the price of the considered option. The proposed algorithm can also be applied to optimal stopping problems that arise in other areas where the underlying stochastic process can be efficiently simulated. We present numerical results for a large number of example problems, which include the pricing of many high-dimensional American and Bermudan options, such as Bermudan max-call options in up to 5000 dimensions. Most of the obtained results are compared to reference values computed by exploiting the specific problem design or, where available, to reference values from the literature. These numerical results suggest that the proposed algorithm is highly effective in the case of many underlyings, in terms of both accuracy and speed.
In recent years deep artificial neural networks (DNNs) have been successfully employed in numerical simulations for a multitude of computational problems including, for example, object and face recognition, natural language processing, fraud detection, computational advertisement, and numerical approximations of partial differential equations (PDEs). These numerical simulations indicate that DNNs seem to have the fundamental flexibility to overcome the curse of dimensionality in the sense that the number of real parameters used to describe the DNN grows at most polynomially in both the reciprocal of the prescribed approximation accuracy ε > 0 and the dimension d ∈ N of the function which the DNN aims to approximate in such computational problems. There is also a large number of rigorous mathematical approximation results for artificial neural networks in the scientific literature but there are only a few special situations where results in the literature can rigorously justify the success of DNNs to approximate high-dimensional functions. The key contribution of this article is to reveal that DNNs do overcome the curse of dimensionality in the numerical approximation of Kolmogorov PDEs with constant diffusion and nonlinear drift coefficients. We prove that the number of parameters used to describe the employed DNN grows at most polynomially in both the reciprocal of the prescribed approximation accuracy ε > 0 and the PDE dimension d ∈ N. A crucial ingredient in our proof is the fact that the artificial neural network used to approximate the solution of the PDE is indeed a deep artificial neural network with a large number of hidden layers.
We show that if a sequence of piecewise affine linear processes converges in the strong sense with a positive rate to a stochastic process that is strongly Hölder continuous in time, then this sequence converges in the strong sense even with respect to much stronger Hölder norms and the convergence rate is essentially reduced by the Hölder exponent. Our first application hereof establishes pathwise convergence rates for spectral Galerkin approximations of stochastic partial differential equations. Our second application derives strong convergence rates of multilevel Monte Carlo approximations of expectations of Banach-space-valued stochastic processes.
In spite of the accomplishments of deep learning based algorithms in numerous applications and very broad corresponding research interest, at the moment there is still no rigorous understanding of the reasons why such algorithms produce useful results in certain situations. A thorough mathematical analysis of deep learning based algorithms seems to be crucial in order to improve our understanding and to make their implementation more effective and efficient. In this article we provide a mathematically rigorous full error analysis of deep learning based empirical risk minimisation with quadratic loss function in the probabilistically strong sense, where the underlying deep neural networks are trained using stochastic gradient descent with random initialisation. The convergence speed we obtain is presumably far from optimal and suffers under the curse of dimensionality. To the best of our knowledge, we establish, however, the first full error analysis in the scientific literature for a deep learning based algorithm in the probabilistically strong sense and, moreover, the first full error analysis in the scientific literature for a deep learning based algorithm where stochastic gradient descent with random initialisation is the employed optimisation method.
In this paper we propose and analyze explicit space-time discrete numerical approximations for additive space-time white noise driven stochastic partial differential equations (SPDEs) with non-globally monotone nonlinearities such as the stochastic Burgers equation with space-time white noise. The main result of this paper proves that the proposed explicit space-time discrete approximation method converges strongly to the solution process of the stochastic Burgers equation with space-time white noise. To the best of our knowledge, the main result of this work is the first result in the literature which establishes strong convergence for a space-time discrete approximation method in the case of the stochastic Burgers equations with space-time white noise.
In this article we study the differentiability of solutions of parabolic semilinear stochastic evolution equations (SEEs) with respect to their initial values. We prove that if the nonlinear drift coefficients and the nonlinear diffusion coefficients of the considered SEEs are n-times continuously Fréchet differentiable, then the solutions of the considered SEEs are also ntimes continuously Fréchet differentiable with respect to their initial values. Moreover, a key contribution of this work is to establish suitable enhanced regularity properties of the derivative processes of the considered SEE in the sense that the dominating linear operator appearing in the SEE smoothes the higher order derivative processes.
Stochastic wave equations appear in several models for evolutionary processes subject to random forces, such as the motion of a strand of DNA in a liquid or heat flow around a ring. Semilinear stochastic wave equations can typically not be solved explicitly, but the literature contains a number of results which show that numerical approximation processes converge with suitable rates of convergence to solutions of such equations. In the case of approximation results for strong convergence rates, semilinear stochastic wave equations with both additive or multiplicative noise have been considered in the literature. In contrast, the existing approximation results for weak convergence rates assume that the diffusion coefficient of the considered semilinear stochastic wave equation is constant, that is, it is assumed that the considered wave equation is driven by additive noise, and no approximation results for multiplicative noise are known. The purpose of this work is to close this gap and to establish essentially sharp weak convergence rates for spatial spectral Galerkin approximations of semilinear stochastic wave equations with multiplicative noise. In particular, our weak convergence result establishes as a special case essentially sharp weak convergence rates for the continuous version of the hyperbolic Anderson model. Our method of proof makes use of the Kolmogorov equation and the Hölder-inequality for Schatten norms.
Although for a number of semilinear stochastic wave equations existence and uniqueness results for corresponding solution processes are known from the literature, these solution processes are typically not explicitly known and numerical approximation methods are needed in order for mathematical modelling with stochastic wave equations to become relevant for real world applications. This, in turn, requires the numerical analysis of convergence rates for such numerical approximation processes. A recent article by the authors proves upper bounds for weak errors for spatial spectral Galerkin approximations of a class of semilinear stochastic wave equations. The findings there are complemented by the main result of this work, that provides lower bounds for weak errors which show that in the general framework considered the established upper bounds can essentially not be improved.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.