A new efficient approximation scheme for solving high-dimensional semilinear PDEs: control variate method for Deep BSDE solver
Abstract
This paper introduces a new approximation scheme for solving high-dimensional semilinear partial differential equations (PDEs) and backward stochastic differential equations (BSDEs). First, we decompose a target semilinear PDE (BSDE) into two parts, namely “dominant” linear and “small” nonlinear PDEs.
Then, we employ a Deep BSDE solver with a new control variate method to solve those PDEs, where approximations based on an asymptotic expansion technique are effectively applied to the linear part and also used as control variates for the nonlinear part.
Moreover, our theoretical result indicates that errors of the proposed method become much smaller than those of the original Deep BSDE solver. Finally, we show numerical experiments
to demonstrate the validity of our method, which is consistent with the theoretical result in this paper.
Keyword. Deep learning, Semilinear partial differential equations, Backward stochastic differential equations, Deep BSDE solver, Asymptotic expansion, Control variate method
1 Introduction
High-dimensional semilinear partial differential equations (PDEs) are often used to describe various complex, large-scale phenomena appearing in physics, applied mathematics, economics and finance. Such PDEs typically have the form:
(1.1) | |||
where is a nonlinear function, is a second order differential operator of the type:
(1.2) |
and the dimension is assumed to be high. To solve the nonlinear PDE, we have to rely on some numerical schemes since they have no closed-form solutions especially in high-dimensional cases. Classical methods such as finite differences and finite elements fail in high-dimensional cases due to their exponential growth of complexity. In the last two decades, probabilistic approaches have been studied with Monte Carlo methods for backward stochastic differential equations (BSDEs) since solutions of semilinear PDEs can be represented by the ones of corresponding BSDEs through the nonlinear Feynman-Kac formula (see Zhang (2017) [42] for instance).
In Weinan E et al. (2017) [3], a novel computational scheme called the Deep BSDE method is proposed. In the Deep BSDE method, a stochastic target problem is considered with a forward-discretization scheme of the related BSDE. Then, the control problem is solved with a deep learning algorithm. The Deep BSDE method has opened the door to tractability of higher dimensional problems, which enables us to solve high-dimensional semilinear PDEs within realistic computation time. Recently, notable related works, mostly with neural networks have developed new methods for solving various types of high dimensional PDEs. See [2][4][5][10][11][12][14][15][17][18][16][23][30][41] for example.
While high-dimensional semilinear PDEs can be feasibly solved by the Deep BSDE method, the deviation of its estimated value from the true one is not small with reasonable computational time. Then, constructing an acceleration scheme for the Deep BSDE method is desirable.
Fujii et al. (2019) [9] proposed an improved scheme for the Deep BSDE method. They used a prior knowledge with an asymptotic expansion method for a target BSDE and obtained its fast approximation. Then, they found that numerical errors become small in accordance with the fast decrease in values of the corresponding loss function. The scheme enables us to reduce processing load of the original Deep BSDE solver. For details of the asymptotic expansion method, a key technique applied in their article, see Takahashi (1999, 2015) [31][32], Kunitomo and Takahashi (2001, 2003) [21][22] and references therein. Moreover, Naito and Yamada (2020) [27] presented an extended scheme of Fujii et al. (2019) [8] by applying the backward Euler scheme for a BSDE with a good initial detection of the solution to a target PDE so that the Deep BSDE method works more efficiently.
In the current work, we develop a new deep learning-based approximation for solving high-dimensional semilinear PDEs by extending the schemes in Weinan E et al. (2017) [3], Fujii et al. (2019) [9] and Naito and Yamada (2020) [27]. In particular, we propose an efficient control variate method for the Deep BSDE solver in order to obtain more accurate and stable approximations. Let us briefly explain the strategy considered in this paper. We first decompose the semilinear PDE (1.1) into two parts,
as follows:
(1.3) | |||
and
(1.4) | |||
Here, we remark that the solution of the semilinear PDE (1.1) is given by the sum of the solutions and of PDEs (1.3) and (1.4), respectively. Also, we note that is the solution to the “dominant” linear PDE and is the solution to the “small” residual nonlinear PDE with null terminal condition whose magnitude is governed by the driver , which is generally expected to have small nonlinear effects on the solution of . Consequently, the decomposition of the target is represented as follows:
(1.5) |
We next approximate
-
1.
by an asymptotic expansion method denoted by ;
-
2.
by the Deep BSDE method, denoted by .
We expect that in the approximation
(1.6) |
becomes a control variate. Furthermore, and in the approximate driver of will be doubly the control variates. The current work shows how the proposed method works well as a new deep learning-based approximation in both theoretical and numerical aspects.
The organization of this paper is as follows: The next section briefly introduces the deep BSDE solver and acceleration schemes with asymptotic expansions. Section 3 explains our proposed method with the main theoretical result and Section 4 presents our numerical scheme with its experiment.
2 Deep BSDE solver and acceleration scheme with asymptotic expansion
Let and be a filtered probability space equipped with a -dimensional Brownian motion and a square-integrable -valued random variable , which is independent of . The filtration is generated by . Under this setting we consider the following FBSDE:
(2.1) | ||||
(2.2) |
where is a -valued function on , is a -valued function on , , are some functions so that the FBSDE has the unique solution, and are some small parameters. Here, we assume that and are bounded and smooth in and have bounded derivatives with any orders. Also, is uniformly Lipschitz continuous function with the Lipschitz constant and at most linear growth in the variables . The function is assumed to be -class. The functions are uniformly Hölder- continuous with respect to . Furthermore, we put the condition that there is such that for all and . We sometimes omit the subscripts or if no confusion arises.
The corresponding semilinear PDE is given by
(2.3) | |||
where , , and is the generator:
(2.4) |
The purpose of this paper is to estimate
(2.5) |
against high dimensional FBSDEs/semilinear PDEs. In particular, we introduce an approximation with a deep BSDE solver to propose an efficient control variate method for solving semilinear PDEs. To explain how our method works well as a new scheme, we briefly review the deep BSDE method proposed in Weinan E et al. (2017) [3] and an approximation method developed by Fujii et al. (2019) [9].
2.1 Deep BSDE method by Weinan E et al. (2017)
In Weinan E et al. (2017) [3], the authors considered the minimization problem of the loss function:
(2.6) |
where , subject to
(2.7) |
where is the continuous Euler-Maruyama scheme with number of discretization time steps :
(2.8) |
with . They solved the problem by using a deep learning algorithm and checked the effectiveness of the method for nonlinear BSDEs/PDEs even for the high dimension . The method is known as Deep BSDE solver.
Then, we have
(2.9) |
where is obtained by solving (2.6), which is justified by the following estimate shown in Han and Long (2020) [14].
Theorem 1 (Han and Long (2020)).
There exists such that
(2.10) |
for .
2.2 An approximation method by Fujii et al. (2019)
In Fujii et al. (2019) [9], the authors considered the problem
(2.11) |
subject to
(2.12) | ||||
(2.13) |
where is a prior knowledge of which is easily computed by an asymptotic expansion method, and they solve the minimization problem with respect to and by Deep BSDE solver. The authors showed that the scheme gives better accuracy than the original Deep BSDE solver. Furthermore, Naito and Yamada (2020) [27] proposed an acceleration scheme by extending the method of Fujii et al. (2019) [9] with a good initial detection of and the backward Euler scheme of . They confirmed that the numerical error of the method becomes smaller even if the number of iteration steps is few, in other words, the scheme gives faster computation for nonlinear BSDEs/PDEs than the original deep BSDE method ([3]) and Fujii et al. (2019) [9].
3 New method
We propose a new method as an extension of Fujii et al. (2019) [9] and Naito and Yamada (2020) [27]. The new scheme is regarded as a control variate method for solving high-dimensional nonlinear BSDEs/PDEs which is motivated by the perturbation scheme in Takahashi and Yamada (2015) [34]. In the following, let us explain the proposed method. We first decompose as and by introducing
(3.1) | ||||
(3.2) |
Here, we note that is the solution of a linear BSDE and that can be interpreted as the solution of a “residual (nonlinear) BSDE”.
Let be the solution of the linear PDE corresponding to :
(3.3) | |||
3.1 Deep BSDE solver for explicitly solvable
We start with a case that is explicitly solvable as a closed-form in order to explain our motivation of the paper. Even in this case, can not be obtained in closed-form due to the nonlinearity of the driver . Hence, we apply the deep BSDE method to the residual nonlinear BSDE . Then, the following will be an approximation for the target :
(3.4) |
where is obtained as a solution of the following problem based on the deep BSDE method with closed-form functions for and :
(3.5) |
subject to
(3.6) |
where
(3.7) |
with the continuous Euler-Maruyama scheme and closed-form functions and .
In this case, we have the following error estimate with a small -effect in the residual nonlinear BSDE. The proof will be shown as a part of the one for Theorem 3 in the next subsection. Particularly, see the sentences after (3.24) and (3.33).
Theorem 2.
There exists such that
(3.8) |
for all and .
3.2 General case: Deep BSDE solver for unsolvable
In most cases, is unsolvable as a closed-form, particularly it is the case when the dimension is high. In such cases, we need to approximate . However, constructing tractable approximations of , , and especially , , is not an easy task because it includes the gradient of . A possible solution is to use an asymptotic expansion approach with stochastic calculus. We prepare some notations of Malliavin calculus. Let be the space of smooth Wiener functionals in the sense of Malliavin. For a nondegenerate and , for a multi-index , there exists such that for all . See Chapter V.8-10 in Ikeda and Watanabe (1989) [20] and Chapter 1-2 in Nualart (2006) [28] for the details.
First, we give approximations of and . For , we approximate and with asymptotic expansions up to the -th order and Malliavin calculus, by applying or extending the methods in [25][32][33][34][40]. Let us consider be the solution of
(3.9) |
Then the -dimensional forward process can be expanded as follows: for ,
(3.10) |
for some , , which are independent of (see Watanabe (1987) [38] for example). Here, is the solution of , and , .
Let us define for . The functions and are approximated by the asymptotic expansion.
Proposition 1.
Let and . There is satisfying that there exist and such that
(3.11) |
for all , and , where the is given by
(3.12) |
which satisfy , . Also, there is satisfying that there exist and such that
(3.13) |
for all , and , where the is given by
(3.14) |
which satisfy , .
Proof. See Appendix A.1.
For example, the stochastic weight has the representation in general:
(3.15) |
See Section 2.2 in Takahashi and Yamada (2012) [33] and Section 6.1 in Takahashi (2015) [32] for more details. The functions and have more explicit representation. Actually when , and have the following forms which are easily computed by taking advantage of the fact that (and ) is a Gaussian random variable. In particular, the representation , the multidimensional expansion of is new, which is an extension of [25][40].
Proposition 2.
For , ,
(3.16) | |||
(3.17) | |||
and
with
Here, is an entry in -th row and -th column of a matrix, is an -th element of and () is used.
Proof. See Appendix A.2.
Using and , we define
(3.18) |
Furthermore, we compute and numerically by the deep BSDE method by solving
(3.19) |
subject to
(3.20) |
where
(3.21) |
with the continuous Euler-Maruyama scheme .
We have the main theoretical result in this paper as follows.
Theorem 3.
There exists such that
(3.22) |
for all and .
Proof. In the proof, we use a generic constant which varies from line to line. Let be the solution of the following BSDE:
(3.23) |
Then we have
(3.24) |
First, we estimate the term . We note that this term becomes null in (3.8), i.e. the error estimate of Theorem 2 for the case that is explicitly solvable as a closed-form.
Since we have
(3.25) |
and
(3.26) | ||||
(3.27) |
it holds that
(3.28) |
Here, the estimates
(3.29) | |||
(3.30) |
are obtained by (3.11) and (3.13). Also, by Theorem 4.2.3 in Zhang (2017) [42], we have
(3.31) |
where the estimates (3.29) and (3.30) are applied in the last inequality. Therefore, we get
(3.32) |
Next, we estimate
(3.33) |
in (3.24). We note that only this term appears in (3.8), i.e. the error estimate of Theorem 2 for the case that is explicitly solvable as a closed-form.
Since we have
(3.34) |
the upper bound of can be decomposed as
(3.35) |
Then, the following holds:
(3.36) | |||
(3.37) |
since for all , and are in and , respectively. Thus, we have
(3.38) |
By Theorem 1 of Han and Long (2020) [14], it holds that
(3.39) |
Therefore, we get
(3.40) |
and the assertion is obtained as:
(3.41) | |||
By the theorem above, it holds that
(3.42) |
where is obtained by solving (3.19) with Deep BSDE method. The process and work as control variates for the nonlinear BSDE.
Here, let us briefly make comments on comparison of the theoretical error estimates of our proposed method, namely (3.8) in Theorem 2 for the explicitly solvable case and (3.22) in Theorem 3 for the unsolvable case with the one provided by Han and Long (2020) for the method of Weinan E et al (2017), i.e. (2.10) in Theorem 1. Given the number of discretized time steps for Euler-Maruyama scheme, those are relisted below:
-
•
Proposed method (for the solvable case):
-
•
Proposed method (for the unsolvable case):
-
•
Method of Weinan E et al. (2017) (error estimate by Han and Long (2020)):
Thanks to the following advantages of our proposed method, we can see that it works better as a new Deep BSDE solver, more precisely, its errors are expected to be smaller:
-
•
(i) Decomposition into a “dominant” linear PDE with original terminal and a “small” nonlinear PDE with zero terminal, i.e.
and an application of Deep BSDE solver only to the “small” nonlinear PDE.
(ii) Closed form solutions/approximations for the linear PDE, which also work as control variates for the driver of the nonlinear PDE.
Thanks to (i) and (ii), we can obtain the term in the error bound, rather than .
Moreover, we note that our method enjoys the effects of a small parameter in the nonlinear driver for this term, as well as for the discretization error term caused by Euler-Maruyama scheme, which is given as rather than .
-
•
Regarding the unsolvable case, our asymptotic expansions with respect to a small parameter in the diffusion coefficient enable us to obtain closed form approximations and : Particularly, in (3.22) the coefficients and are associated with errors of the approximations for terminal and driver , respectively.
We will check the effectiveness of the new method by numerical experiments in the next section.
Remark 1.
We give an important remark on the new method. While the proposed scheme provides a fine result, we can further improve it by replacing our approximation for the linear part in the decomposition (3.42) with
the methods of [36][35][39][26]
[29][20].
For example, based on Takahashi and Yamada (2016) [35], the following result will be an improvement of the proposed scheme. Let , , with a parameter , and , . Define
(3.43) |
and consider the quantity
(3.44) |
where is the same as in (3.42). Then, (3.44) will be the improved approximation, as
(3.45) |
in the following sense.
Corollary 1.
There exist and such that
(3.46) |
for all and .
Remark 2.
If the driver of a BSDE contains a linear part, we can transform the BSDE to the one considered in the current paper, namely the equation (2.2). For instance, let us solve FSDE (2.1) and the following BSDE:
where and are -valued and -valued bounded functions, respectively.
Let , , and , , which is a Brownian motion under a probability measure obtained by the change of measure with process . Then, we have
4 Numerical results
In the numerical examples, we demonstrate that the deep BSDE method with the first order asymptotic expansion obtained in Proposition 2 provides enough accuracy in solving semilinear PDEs. The dimension in (2.2) is assumed to be or .
We investigate the accuracy of the new method by comparing to the standard Deep BSDE method in Weinan E et al. (2017) [3] and the Deep BSDE method with a prior knowledge in Fujii et al. (2019) [9] for the model (2.2), where the target BSDEs with FSDEs are specified later.
4.1 Numerical schemes used in experiments
In this subsection, we explain the details of schemes used in numerical experiments. To construct the deep neural networks for each method, we follow Weinan E et al. (2017) [3] and employ the adaptive moment estimation (Adam) with mini-batches. The parameters for the networks are set as follows: there are of hidden layers except batch normalization layers. For all learning steps, sample paths are generated and the learning rate is taken as .
(Numerical scheme) Now, let us briefly explain the schemes used in the numerical experiment in the following subsections.
-
1.
(Deep BSDE method based on Weinan E et al. (2017)) In forward discretization of , the Euler-Maruyama scheme is applied with time step . The initial guess of is generated by uniform random number around , which is a prior knowledge for the Deep BSDE method.
In the study of Weinan E et al. (2017), it is known that the estimated value by the Deep BSDE method converges to the true value of if we take a sufficient number of iteration steps.
In Section 4.2 below, the estimate values based on this scheme are shown by the green lines labeled with “Deep BSDE” in the figures.
-
2.
(Deep BSDE method with an enhanced version of Fujii et al. (2019)) In forward discretization of in the Deep BSDE solver, as an approximation of we apply , , with the function defined by (3.17) and the Euler-Maruyama scheme with time step to obtain an estimate of by optimization in the Deep BSDE solver. As the initial value of , we use with the function defined by (3.16), an approximation of , which appears in the linear part of our decomposition of the BSDE with and . Thus, the scheme is an improved version of Fujii et al. (2019) [9], since it applies the higher order term than the leading order term that Fujii et al. (2019) [9] uses.
Through the study of Fujii et al. (2019) [9], it is also known that the estimated value by the enhanced Deep BSDE method converges to the true value of with a much smaller number of iteration steps than by the original Deep BSDE method in Weinan E et al. (2017).
In Section 4.2 below, the estimate values based on this scheme are shown by the red lines labeled with “Deep BSDE[]+AE[ and ]” in the figures.
-
3.
(New scheme) Following the main result introduced in Section 3, particularly Theorem 2, we employ our approximation (3.42) for the decomposition with and , namely,
where we compute the nonlinear part with (3.19)–(3.21) by the Deep BSDE solver, while by with the function defined by (3.16).
Specifically, in computation of by the Deep BSDE solver with the equation:
we use and as approximations for and in the driver , respectively.
In Section 4.2 below, the estimated values based on this new scheme are shown by the blue lines labeled with “New method [, ]” in the figures.
The initial value of or is generated by uniform random number with the range for each method. Numerical experiments presented in the following subsections are implemented by Python with TensorFlow on Google Colaboratory.
4.2 Numerical experiments
We show some numerical examples which show that our proposed method substantially outperforms other methods in terms of terminal errors (numerical values of loss functions), variations and convergence speed.
4.2.1 The case of
This subsection presents the numerical results for the case of .
We first check the performance of our method in the model, where the explicit value of the solution is obtained by the Picard iteration. We consider an option pricing model in finance that takes CVA(credit value adjustment) into account as follows:
(4.1) | ||||
(4.2) |
with and . in (2.1) and (2.2). We note that (loss rate in default)(default intensity) in a finance model of CVA.
In computation we set , , , , , with (ATM case) and (OTM case).
In this case an explicit value of is computed as . More precisely, by the -Picard iteration of the backward equation:
(4.3) | ||||
with | (4.4) |
it is easy to see that , and thus one has
Then, the true values are given as in the ATM case and in the OTM case by the -Picard iteration, which provides enough convergence and hence accuracy.
Figure 1 and 2 show the numerical values of loss functions and the approximate values of , respectively against the number of iteration steps for the ATM case, and Figure 3 and 4 for the OTM case.
By Figure 2 and 4, the numerical values of “New method [,” converge to the true values substantially faster with smaller variations comparing to other schemes. Also, we can see that the errors of “New method” are much smaller according to the behavior of their loss functions against the number of iteration steps in Figure 1 and 3.




Next, we present numerical examples for the model, where explicit values of can not be obtained without numerical schemes such as Monte Carlo simulations. Let us consider
(4.5) | ||||
(4.6) |
with , , , , , and
(4.7) |
with , , and
(4.8) |
As we mentioned in Section 4.1, the estimated value by the methods “Deep BSDE” and “Deep BSDE[]+AE[ and ]” converges to the true value of . Then, in the experiments, we check whether the estimated value by “New method [,” converges faster than the ones computed by the methods “Deep BSDE” and “Deep BSDE[]+AE[ and ]”.
Figure 5 shows the numerical values of loss functions against the number of iteration steps. While “Deep BSDE[]+AE[ and ]” is superior to the original “Deep BSDE”, we see that “New method [,” gives much more stable and accurate convergence than other schemes.
Figure 6 shows the approximate values of against the number of iteration steps. It is observed that “New method [,” provides the fastest convergence with the smallest standard deviation, while “Deep BSDE[]+AE[ and ]” gives better approximation than “Deep BSDE”.


4.2.2 The case of
We show the main numerical result for . The same experiment as in the case of is performed. Let us consider
(4.9) | ||||
(4.10) |
with , (), , , , and
(4.11) |
with , where is defined by , and
(4.12) |
The result is given in Figure 7, 8 and 9. It seems that the convergence speed of the original deep BSDE method is too slow to obtain the precise result. On the contrary, “Deep BSDE[]+AE[ and ]” and “New method [,” work well even in this high dimensional case. Particularly, “New method [,” provides a remarkable performance in terms of convergence speed, accuracy (numerical values of loss functions) and variations. Moreover, comparing the results of our new method and “Deep BSDE[]+AE[ and ]” closely, Figure 9 shows that the variation of by our method is much smaller, which is consistent with much smaller values of loss functions for the new method appearing in Figure 7.



5 Conclusion and future works
This paper has introduced a new control variate method for Deep BSDE solver to improve the methods such as in Weinan E et al. (2017) [3] and Fujii et al. (2019) [9]. First, we decompose a target semilinear PDE (BSDE) to two parts, namely dominant linear and residual nonlinear PDEs (BSDEs). When the dominant part is obtained as a closed-form or approximated based on an asymptotic expansion scheme, the small nonlinear PDE part is efficiently computed by Deep BSDE solver, where the asymptotic expansion crucially works as a control variate. The main theorem provides the validity of our proposed method. Moreover, numerical examples for one and 100 dimensional BSDEs corresponding to target nonlinear PDEs show the effectiveness of our scheme, which is consistent with our initial conjecture and theoretical result.
As mentioned in Remark 1, even if the accuracy of the standard asymptotic expansion scheme becomes worse, the linear PDE part can be more efficiently approximated by the existing methods such as [36][35][39][26][29][20]. We should check those performances in such cases against various nonlinear models. Also, it will be a challenging task to examine whether the high order automatic differentiation schemes proposed in [40][37] work as efficient approximations of in nonlinear BSDEs or in nonlinear PDEs. These are left for future studies.
Appendix
Appendix A Proof of Propositions
A.1 Proof of Proposition 1
Also, note that for and a multi-index , and hold for . Then, for , i.e. , .
For , we have the representation
for a matrix-valued Wiener functional such that , , satisfying for and a multi-index , for . Then, for , i.e. , .
A.2 Proof of Proposition 2
For the derivations, we use Malliavin calculus. Let be a tempered distribution and be a nondegenerate Wiener functional in the sense of Malliavin.
Then, is well-defined as an element of the space of Watanabe distributions , that is the dual space of . Also, for , a (generalized)
expectation is understood as
a coupling of
and , namely
.
Note that and in
(A.1) |
and
(A.2) | |||
(A.3) |
have expansions in whose expansion coefficients are given by iterated stochastic integrals: and . In particular,
(A.4) | ||||
(A.5) | ||||
(A.6) |
and
(A.7) | |||
(A.8) |
where the followings are used: for a smooth function ,
(A.9) |
(A.10) |
We expand in (A.1) and in (A.3) to obtain explicit expressions of and . Next, let us recall the following formulas.
Lemma 1.
Let be a tempered distribution.
-
1.
For an adapted process ,
(A.11) where represents the -th element of the Malliavin derivative
for . -
2.
For ,
(A.12)
Proof of Lemma 1. Use the duality formula (see Theorem 1.26 of [24] or Proposition 1.3.11 of [28]), with for (see Proof of Proposition 2.1.9 of [28] or Proof of Theorem 2.6 of [33]) to obtain the first assertion. Also, the second assertion is immediately obtained by the integration by parts.
In the expansions of (A.1) and (A.3), iterated integrals such as
(A.13) |
appear, for which the following calculation holds with use of (A.11):
(A.14) | ||||
Note that is deterministic, and one has
(A.15) |
Thus, we get
(A.16) | ||||
(A.17) | ||||
(A.18) |
Using the above calculation with (A.12), we have
(A.19) | ||||
(A.20) | ||||
(A.21) | ||||
(A.22) | ||||
(A.23) | ||||
(A.24) | ||||
(A.25) |
Therefore, we get (3.16) as:
(A.26) | ||||
(A.27) | ||||
(A.28) | ||||
(A.29) |
Acknowledgements
This work is supported by JSPS KAKENHI (Grant Number 19K13736) and JST PRESTO (Grant Number JPMJPR2029), Japan.
References
- [1] C. Beck, F. Hornung, M. Hutzenthaler and A. Jentzen and T. Kruse, Overcoming the curse of dimensionality in the numerical approximation of Allen-Cahn partial differential equations via truncated full-history recursive multilevel Picard approximations, Journal of Numerical Mathematics (2020)
- [2] J. Berner, P. Grohs and A. Jentzen, Analysis of the generalization error: Empirical risk minimization over deep artificial neural networks overcomes the curse of dimensionality in the numerical approximation of Black–Scholes partial differential equations, SIAM Journal on Mathematics of Data Science, 2(3), 631-657 (2020)
- [3] Weinan E, J. Han and A. Jentzen, Deep learning-based numerical methods for high-dimensional parabolic partial differential equations and backward stochastic differential equations, Communications in Mathematics and Statistics, 5(4) 349-380 (2017)
- [4] Weinan E, J. Han and A. Jentzen, Algorithms for Solving High Dimensional PDEs: From Nonlinear Monte Carlo to Machine Learning, arXiv (2020)
- [5] D. Elbrächter, P. Grohs, A. Jentzen and C. Schwab, DNN expression rate analysis of high-dimensional PDEs: Application to option pricing, arXiv (2018)
- [6] El Karoui, S. Peng and M.C. Quenez, Backward stochastic differential equations in finance, Mathematical Finance, 7(1), 1-71 (1997)
- [7] M. Fujii and A. Takahashi, Analytical approximation for non-linear FBSDEs with perturbation scheme, International Journal of Theoretical and Applied Finance, (2011)
- [8] M. Fujii and A. Takahashi, Solving backward stochastic differential equations with quadratic-growth drivers by connecting the short-term expansions, Stochastic Processes and their Applications, 129(5) (2019)
- [9] M. Fujii, A. Takahashi and M. Takahashi, Asymptotic expansion as prior knowledge in deep learning method for high dimensional BSDEs, Asia-Pacific Financial Markets, (2019)
- [10] A. Gnoatto, C. Reisinger and A. Picarelli, Deep xVA solver-A neural network based counterparty credit risk management framework, SSRN (2020)
- [11] P. Grohs, A. Jentzen and D. Salimova, Deep neural network approximations for Monte Carlo algorithms, arXiv (2019)
- [12] P. Grohs, F. Hornung, A. Jentzen and P. Zimmermann, Space-time error estimates for deep neural network approximations for differential equations, arXiv (2019)
- [13] M.B. Giles, A. Jentzen and T. Welti, Generalised multilevel Picard approximations, arXiv (2020)
- [14] J. Han and J. Long, Convergence of the Deep BSDE method for coupled FBSDEs, Probability, Uncertainty and Quantitative Risk, 5(5) (2020)
- [15] J. Han, J. Lu and M.Zhou, Solving high-dimensional eigenvalue problems using deep neural networks: A diffusion Monte Carlo like approach, Journal of Computational Physics, Volume 423, 15, December 2020, 109792 (2020)
- [16] J. Han, L. Zhang and Weinan E, Solving many-electron Schrödinger equation using deep neural networks, Journal of Computational Physics, Volume 399, 15, December 2019, 108929 (2019)
- [17] F. Hornung, A. Jentzen and D. Salimova, Space-time deep neural network approximations for high-dimensional partial differential equations, arXiv (2020)
- [18] M. Hutzenthaler, A. Jentzen, T. Kruse, T.A. Nguyen and P.V. Wurstemberger, Overcoming the curse of dimensionality in the numerical approximation of semilinear parabolic partial differential equations, arXiv (2018)
- [19] N. Ikeda and S. Watanabe, Stochastic Differential Equations and Diffusion Processes, 2nd ed., North-Holland, Amsterdam, Kodansha, Tokyo (1989)
- [20] Y. Iguchi and T. Yamada, Operator splitting around Euler-Maruyama scheme and high order discretization of heat kernels, ESAIM: Mathematical Modelling and Numerical Analysis, to appear (2020)
- [21] N. Kunitomo and A. Takahashi, The asymptotic expansion approach to the valuation of interest rate contingent claims, Mathematical Finance, 11, 117-151 (2001)
- [22] N. Kunitomo and A. Takahashi, On validity of the asymptotic expansion approach in contingent claim analysis, Annals of Applied Probability, 13(3), 914-952 (2003)
- [23] Y. Li, J. Lu and A. Mao, Variational training of neural network approximations of solution maps for physical models, Journal of Computational Physics, Volume 409, 15 May 2020, 109338 (2020)
- [24] P. Malliavin and A. Thalmaier, Stochastic Calculus of Variations in Mathematical Finance, Springer (2006)
- [25] R. Matsuoka, A. Takahashi and Y. Uchida, A new computational scheme for computing Greeks by the asymptotic expansion approach, Asia Pacific Financial Markets, 11, 393-430 (2006)
- [26] R. Naito and T. Yamada, A third-order weak approximation of multidimensional Itô stochastic differential equations, Monte Carlo Methods and Applications, vol 25 (2), 97-120 (2019)
- [27] R. Naito and T. Yamada, An acceleration scheme for deep learning-based BSDE solver using weak expansions, International Journal of Financial Engineering, (2020)
- [28] D. Nualart, The Malliavin Calculus and Related Topics, Springer (2006)
- [29] Y. Okano and T. Yamada, A control variate method for weak approximation of SDEs via discretization of numerical error of asymptotic expansion, Monte Carlo Methods and Applications, 25(3) (2019)
- [30] J. Sirignano and K. Spiliopoulos, DGM: A deep learning algorithm for solving partial differential equations, Journal of Computational Physics, Vol 375, 1339-1364 (2018)
- [31] A. Takahashi, An asymptotic expansion approach to pricing financial contingent claims, Asia-Pacific Financial Markets, 6(2), 115-151 (1999)
- [32] A. Takahashi, Asymptotic expansion approach in finance, Large Deviations and Asymptotic Methods in Finance (P. Friz, J. Gatheral, A. Gulisashvili, A. Jacquier and J. Teichmann ed.), Springer Proceedings in Mathematics & Statistics (2015)
- [33] A. Takahashi and T. Yamada, An asymptotic expansion with push-down of Malliavin weights, SIAM Journal on Financial Mathematics, 3, 95-136 (2012)
- [34] A. Takahashi and T. Yamada, An asymptotic expansion of forward-backward SDEs with a perturbed driver, International Journal of Financial Engineering, 2(2) (2015)
- [35] A. Takahashi and T. Yamada, A weak approximation with asymptotic expansion and multidimensional Malliavin weights, Annals of Applied Probability, 26(2), 818-856 (2016)
- [36] A. Takahashi and N. Yoshida, Monte Carlo simulation with asymptotic method, Journal of the Japan Statistical Society 35(2), 171-203 (2005)
- [37] K. Tokutome and T. Yamada, Acceleration of automatic differentiation of solutions to parabolic partial differential equations: a higher order discretization, Numerical Algorithms, Vol.86, 593-635 (2021)
- [38] S. Watanabe, Analysis of Wiener functionals (Malliavin calculus) and its applications to heat kernels, Annals of Probability, 15, 1-39 (1987)
- [39] T. Yamada, An arbitrary high order weak approximation of SDE and Malliavin Monte Carlo: application to probability distribution functions, SIAM Journal on Numerical Analysis, 57(2), 563-591 (2019)
- [40] T. Yamada and K. Yamamoto, Second order discretization of Bismut-Elworthy-Li formula: application to sensitivity analysis, SIAM/ASA Journal on Uncertainty Quantification, 7(1), 143-173 (2019)
- [41] Y. Zang, G. Bao, X. Ye and H. Zhou, Weak adversarial networks for high-dimensional partial differential equations, Journal of Computational Physics, 411 (2020)
- [42] J. Zhang, Backward Stochastic Differential Equations, Springer (2017)