Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2018 Aug 21;115(34):8505-8510.
doi: 10.1073/pnas.1718942115. Epub 2018 Aug 6.

Solving high-dimensional partial differential equations using deep learning

Affiliations

Solving high-dimensional partial differential equations using deep learning

Jiequn Han et al. Proc Natl Acad Sci U S A. .

Abstract

Developing algorithms for solving high-dimensional partial differential equations (PDEs) has been an exceedingly difficult task for a long time, due to the notoriously difficult problem known as the "curse of dimensionality." This paper introduces a deep learning-based approach that can handle general high-dimensional parabolic PDEs. To this end, the PDEs are reformulated using backward stochastic differential equations and the gradient of the unknown solution is approximated by neural networks, very much in the spirit of deep reinforcement learning with the gradient acting as the policy function. Numerical results on examples including the nonlinear Black-Scholes equation, the Hamilton-Jacobi-Bellman equation, and the Allen-Cahn equation suggest that the proposed algorithm is quite effective in high dimensions, in terms of both accuracy and cost. This opens up possibilities in economics, finance, operational research, and physics, by considering all participating agents, assets, resources, or particles together at the same time, instead of making ad hoc assumptions on their interrelationships.

Keywords: Feynman–Kac; backward stochastic differential equations; deep learning; high dimension; partial differential equations.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Fig. 1.
Fig. 1.
Plot of θu0 as an approximation of u(t=0,x=(100,,100)) against the number of iteration steps in the case of the 100-dimensional nonlinear Black–Scholes equation with 40 equidistant time steps (N=40) and learning rate 0.008. The shaded area depicts the mean ± the SD of θu0 as an approximation of u(t=0,x=(100,,100)) for five independent runs. The deep BSDE method achieves a relative error of size 0.46% in a runtime of 1,607 s.
Fig. 2.
Fig. 2.
(Top) Relative error of the deep BSDE method for u(t=0,x=(0,,0)) when λ=1 against the number of iteration steps in the case of the 100-dimensional HJB Eq. 13 with 20 equidistant time steps (N=20) and learning rate 0.01. The shaded area depicts the mean ± the SD of the relative error for five different runs. The deep BSDE method achieves a relative error of size 0.17% in a runtime of 330 s. (Bottom) Optimal cost u(t=0,x=(0,,0)) against different values of λ in the case of the 100-dimensional HJB Eq. 13, obtained by the deep BSDE method and classical Monte Carlo simulations of Eq. 14.
Fig. 3.
Fig. 3.
(Top) Relative error of the deep BSDE method for u(t=0.3,x=(0,,0)) against the number of iteration steps in the case of the 100-dimensional Allen–Cahn Eq. 15 with 20 equidistant time steps (N=20) and learning rate 0.0005. The shaded area depicts the mean ± the SD of the relative error for five different runs. The deep BSDE method achieves a relative error of size 0.30% in a runtime of 647 s. (Bottom) Time evolution of u(t,x=(0,,0)) for t[0,0.3] in the case of the 100-dimensional Allen–Cahn Eq. 15 computed by means of the deep BSDE method.
Fig. 4.
Fig. 4.
Illustration of the network architecture for solving semilinear parabolic PDEs with H hidden layers for each subnetwork and N time intervals. The whole network has (H+1)(N1) layers in total that involve free parameters to be optimized simultaneously. Each column for t=t1,t2,,tN1 corresponds to a subnetwork at time t. hn1,,hnH are the intermediate neurons in the subnetwork at time t=tn for n=1,2,,N1.

References

    1. Bellman RE. Dynamic Programming. Princeton Univ Press; Princeton: 1957.
    1. Goodfellow I, Bengio Y, Courville A. Deep Learning. MIT Press; Cambridge, MA: 2016.
    1. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521:436–444. - PubMed
    1. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. In: Bartlett P, Pereira F, Burges CJC, Bottou L, Weinberger KQ, editors. Advances in Neural Information Processing Systems. Vol 25. Curran Associates, Inc.; Red Hook, NY: 2012. pp. 1097–1105.
    1. Hinton G, et al. Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups. IEEE Signal Process Mag. 2012;29:82–97.

Publication types