This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A framework for adaptive Monte-Carlo procedures

Bernard Lapeyre111 Université Paris-Est, CERMICS, Projet MathFi ENPC-INRIA-UMLV, 6 et 8 avenue Blaise Pascal, 77455 Marne La Vallée, Cedex 2, France , e-mail : bernard.lapeyre@enpc.fr.    Jérôme Lelong222 Laboratoire Jean Kuntzmann, Université de Grenoble et CNRS, BP 53, 38041 Grenoble Cédex 9, France, e-mail : jerome.lelong@imag.fr
(July 25, 2025)
Abstract

Adaptive Monte Carlo methods are recent variance reduction techniques. In this work, we propose a mathematical setting which greatly relaxes the assumptions needed by for the adaptive importance sampling techniques presented in [24, 23, 1, 2]. We establish the convergence and asymptotic normality of the adaptive Monte Carlo estimator under local assumptions which are easily verifiable in practice. We present one way of approximating the optimal importance sampling parameter using a randomly truncated stochastic algorithm. Finally, we apply this technique to some examples of valuation of financial derivatives.

1 Introduction

Monte-Carlo methods aim at computing the expectation 𝔼(Z){\mathbb{E}}(Z) of a real-valued random variable ZZ using samples along the law of ZZ. In this work, we focus on cases where there exists a parametric representation of the expectation

𝔼(Z)=𝔼(H(θ,X))for all θd,{\mathbb{E}}(Z)={\mathbb{E}}\left(H(\theta,X)\right)\quad\mbox{for all }\theta\in{\mathbb{R}}^{d}, (1)

where XX is a random vector with values in m{\mathbb{R}}^{m} and H:d×mH:{\mathbb{R}}^{d}\times{\mathbb{R}}^{m}\longmapsto{\mathbb{R}} is a measurable function satisfying 𝔼|H(θ,X)|<{\mathbb{E}}|H(\theta,X)|<\infty for all θd\theta\in{\mathbb{R}}^{d}. We also impose that

θv(θ)=Var(H(θ,X)) is finite for all θd,\theta\longmapsto v(\theta)=\mathop{\rm Var}\nolimits(H(\theta,X))\mbox{ is finite for all }\theta\in{\mathbb{R}}^{d}, (2)

We want to make the most of this free parameter θ\theta to settle an automatic variance reduction method, see [8] for a recent survey on adaptive variance reduction. It consists in first finding a minimiser θ\theta^{\star} of the variance vv and then in plugging it into a Monte Carlo method with a narrower confidence interval. This technique heavily relies on the ability to find a parametric representation and to effectively minimise the function vv. Many papers have been written on how to construct parametric representations H(θ,X)H(\theta,X) for several kinds of random variables ZZ. We mainly have in mind examples based on control variates (see [4, 13, 12]) or importance sampling (see [24, 23, 1, 2]). We refer the reader to section 4 for a presentation of a few examples.

Assume we have a parametric representation of the form H(θ,X)H(\theta,X) satisfying Equations (1) and (2). Let (Xn)n(X_{n})_{n} be an independent and identically distributed sequence of random vectors following the law of XX. Assume we know how to use the sequence (Xn)n(X_{n})_{n} to build an estimator θn\theta_{n} of θ\theta^{\star} adapted to the filtration n=σ(X1,,Xn){\mathcal{F}}_{n}=\sigma(X_{1},\dots,X_{n}). Once such an approximation is available, there are at least two ways of using it to devise a variance reduction method.

The non-adaptive algorithm
Algorithm 1.1 (Non adaptive importance sampling (NADIS)).

Let nn be the number of samples used for the Monte Carlo computation. Draw a second set of nn samples (X1,,Xn)(X^{\prime}_{1},\dots,X^{\prime}_{n}) independent of (X1,,Xn)(X_{1},\dots,X_{n}) and compute

ξ¯n=1ni=1nH(θn,Xi).\bar{\xi}_{n}=\frac{1}{n}\sum_{i=1}^{n}H(\theta_{n},X^{\prime}_{i}).

Since the sequence (θn)n(\theta_{n})_{n} converges to θ\theta^{*}, the convergence of (ξ¯n)n(\bar{\xi}_{n})_{n} to 𝔼(Z){\mathbb{E}}(Z) ensues from the strong law of large numbers and the sequence (ξ¯n)n(\bar{\xi}_{n})_{n} satisfies a central limit theorem

n(ξ¯n𝔼(Z))nlaw𝒩(0,v(θ)).\sqrt{n}(\bar{\xi}_{n}-{\mathbb{E}}(Z))\xrightarrow[n\rightarrow\infty]{law}{\mathcal{N}}\left(0,v(\theta^{\star})\right).

This algorithm has been studied in [23, 2] and required 2n2n samples. It may use less than 2n2n samples if the estimation of θ\theta^{\star} is performed on a smaller number of samples but then it raises the question of how many samples to use.

The adaptive algorithm

The adaptive approach is to use the same samples (X1,,Xn)(X_{1},\dots,X_{n}) to compute θn\theta_{n} and the Monte Carlo estimator. Compared to the sequential algorithm, the adaptive one uses half of the samples.

Algorithm 1.2 (Adaptive Importance Sampling (ADIS)).

Let nn be the number of samples used for the Monte Carlo computation.
For θ0\theta_{0} fixed in d{\mathbb{R}}^{d}, compute

ξn=1ni=1nH(θi1,Xi).\xi_{n}=\frac{1}{n}\sum_{i=1}^{n}H(\theta_{i-1},X_{i}). (3)

Note that the sequence (ξi)i(\xi_{i})_{i} can be written in a recursive manner so that it can be updated online each time a new iterate θi\theta_{i} is drawn

ξi+1=ii+1ξi+1i+1H(θi,Xi+1),with ξ0=0.\xi_{i+1}=\frac{i}{i+1}\xi_{i}+\frac{1}{i+1}H(\theta_{i},X_{i+1}),\quad\mbox{with $\xi_{0}=0$}.

Being able to update the sequence (ξi)i(\xi_{i})_{i} online has the advantage that there is no need to store the whole sequence (X1,,Xn)(X_{1},\dots,X_{n}) for computing ξn\xi_{n}. This adaptive algorithm was first studied in [1] in which the author studied the convergence of the sequence (ξn)n(\xi_{n})_{n} under assumptions to be verified along the path (θn)n(\theta_{n})_{n} which makes them hard to check in practise. In this article, we prove a new convergence result under local integrability conditions on the function HH, namely we impose that for any compact subset KK of n{\mathbb{R}}^{n}, supθK𝔼(|H(θ,X)|2)<\sup_{\theta\in K}{\mathbb{E}}(|H(\theta,X)|^{2})<\infty. We refer the reader to section 2.1 for a precise statement and proof of these results. We want to emphasize that such assumptions only involving properties of the function HH and not of the sequence (θn)n(\theta_{n})_{n} are far easier to check in practice.

Sofar, we have assumed that we knew how to devise a convergent estimator of θ\theta^{\star}, but this may not be so simple as when no closed form expression is available for 𝔼(H(θ,X)){\mathbb{E}}\left(H(\theta,X)\right), there is hardly no chance that the function vv can be computed explicitly. Henceforth, it is needed to approximate θ\theta^{\star} without being able to compute the variance itself. In this work, we recall the methodology based on stochastic approximation developed in [23, 24, 2] to estimate θ\theta^{\star} using some stochastic gradient style algorithms. We aim at applying this methodology to the evaluation of financial derivatives and the main difficulty in approximating θ\theta^{\star} comes from the non-boundedness of the payoff functions usually considered and consequently the non-boundedness of the HH functions. To encompass this problem, several authors as in [24, 23] restrict the parameter θ\theta to lie in a compact set, which is obviously unknown in practice; therefore, this compact set will have to be quite large. Although, it permits to prove the theoretical convergence of the Robbins-Monro algorithm it does not help to build a numerically convergent estimator of θ\theta^{\star}. We all know that the true convergence of stochastic algorithms highly relies on the fine tuning of the gain sequence which reveals to be very difficult when dealing with an artificially bounded parameter set.

In this work following [2], we would rather use a randomly truncated algorithm which is known to converge for a much wider class of functions. We give a unified framework with easily verifiable assumptions under which Algorithm 1.2 converges and satisfies a central limit theorem. Then, we combine this convergence result with the new results on randomly truncated stochastic algorithm from [16] to revisit the adaptive algorithm in the Gaussian framework studied in [1].

The paper is organised as follows. In Section 2, we focus on the mathematical foundation of the method and give both a strong law of large numbers and a central limit theorem for the adaptive estimator under weak assumptions. In Section 3, we present one way of constructing a convergent estimator of θ\theta^{\star} and recall some recent results on stochastic approximation. Then, we give in Section 4 some examples of how to construct a parametric estimator using importance sampling or other more elaborate transformations. Finally, we illustrate the convergence results obtained in Section 2 on numerical examples coming from financial problems.

2 Mathematical foundations of the method

Notations:

  • We encode any elements of m{\mathbb{R}}^{m} as column vectors.

  • If xmx\in{\mathbb{R}}^{m}, xx^{*} is a row vector. We use the “” notation to denote the transpose operator for vectors and matrices.

  • If x,ymx,y\in{\mathbb{R}}^{m}, xyx\cdot y denotes the Euclidean scalar product of xx and yy and the associated norm is denoted by |||\cdot|.

In this section, (Xn)n1(X_{n})_{n\geq 1} is an i.i.d. sequence following the law of XX and we introduce the σ\sigma-algebra n{\mathcal{F}}_{n} it generates n=σ(X1,,Xn){\mathcal{F}}_{n}=\sigma(X_{1},\dots,X_{n}). For technical reasons, we assume that the variance vv does not vanish, i.e. infθnv(θ)>0\inf_{\theta\in{\mathbb{R}}^{n}}v(\theta)>0. If such is not the case, it means that we are actually in a better situation as far as variance reduction is concerned but it does not fit in our framework.

2.1 An adaptive strong law of large numbers

Theorem 2.1 (Adaptive strong law of large numbers).

Assume Equation (1) and (2) hold. Let (θn)n0(\theta_{n})_{n\geq 0} be a (n)({\mathcal{F}}_{n})-adapted sequence with values in d{\mathbb{R}}^{d} such that for all n0n\geq 0, θn<\theta_{n}<\infty a.s and for any compact subset KdK\subset{\mathbb{R}}^{d}, supθK𝔼(|H(θ,X)|2)<\sup_{\theta\in K}{\mathbb{E}}(|H(\theta,X)|^{2})<\infty. If

infθdv(θ)>0and1nk=0nv(θk)<a.s.,\inf_{\theta\in{\mathbb{R}}^{d}}v(\theta)>0\qquad\mbox{and}\qquad\frac{1}{n}\sum_{k=0}^{n}v(\theta_{k})<\infty\quad\mbox{a.s.}, (4)

then ξn\xi_{n} converges a.s. to 𝔼(Z){\mathbb{E}}(Z).

Proof.

For any p0p\geq 0, we define τp=inf{k0;|θk|p}\tau_{p}=\inf\{k\geq 0;|\theta_{k}|\geq p\}. The sequence (τp)p(\tau_{p})_{p} is an increasing sequence of (n)({\mathcal{F}}_{n})-stopping times such that limpτp\lim_{p\rightarrow\infty}\tau_{p}\uparrow\infty a.s.. Let Mn=i=0n1H(θi,Xi+1)𝔼(Z)M_{n}=\sum_{i=0}^{n-1}H(\theta_{i},X_{i+1})-{\mathbb{E}}(Z). We introduce Mnτp=MτpnM^{\tau_{p}}_{n}=M_{\tau_{p}\wedge n} defined by

Mnτp=i=0n1τpH(θi,Xi+1)𝔼(Z)=i=0n1(H(θi,Xi+1)𝔼(Z))𝟏{iτp}.M^{\tau_{p}}_{n}=\sum_{i=0}^{n-1\wedge\tau_{p}}H(\theta_{i},X_{i+1})-{\mathbb{E}}(Z)=\sum_{i=0}^{n-1}(H(\theta_{i},X_{i+1})-{\mathbb{E}}(Z)){\bf 1}_{\{i\leq\tau_{p}\}}.

𝔼(|H(θi,Xi+1)𝔼(Z)|2𝟏{iτp})𝔼(𝟏{iτp}𝔼(|H(θ,X)𝔼(Z)|2)θ=θi){\mathbb{E}}(|H(\theta_{i},X_{i+1})-{\mathbb{E}}(Z)|^{2}{\bf 1}_{\{i\leq\tau_{p}\}})\leq{\mathbb{E}}({\bf 1}_{\{i\leq\tau_{p}\}}{\mathbb{E}}(|H(\theta,X)-{\mathbb{E}}(Z)|^{2})_{\theta=\theta_{i}}). On the set {iτp}\{i\leq\tau_{p}\}, the conditional expectation is bounded from above by sup|θ|pv(θ)\sup_{|\theta|\leq p}v(\theta). Hence, the sequence (Mnτp)n(M^{\tau_{p}}_{n})_{n} is square integrable and it is obvious that (Mnτp)n(M^{\tau_{p}}_{n})_{n} is a martingale, which means that the sequence (Mn)n(M_{n})_{n} is a locally square integrable martingale (i.e. a local martingale which is locally square integrable).

Mn=i=0n1𝔼((H(θi,Xi+1)𝔼(Z))2|i)=i=0n1v(θi).\langle M\rangle_{n}=\sum_{i=0}^{n-1}{\mathbb{E}}((H(\theta_{i},X_{i+1})-{\mathbb{E}}(Z))^{2}|{\mathcal{F}}_{i})=\sum_{i=0}^{n-1}v(\theta_{i}).

By Condition (4), we have a.s. limsupn1nMn<\lim\sup_{n}\frac{1}{n}\langle M\rangle_{n}<\infty and liminfn1nMn>0\lim\inf_{n}\frac{1}{n}\langle M\rangle_{n}>0. Applying the strong law of large numbers for locally 𝕃2{\mathbb{L}}^{2} martingales (see [18]) yields the result. ∎

The sequence (θn)n(\theta_{n})_{n} can be any sequence adapted to (Xn)n1(X_{n})_{n\geq 1} convergent or not. For instance, (θn)n(\theta_{n})_{n} can be an ergodic Markov chain distributed around the minimizer θ\theta^{\star} such as Monte Carlo Markov Chain algorithms.

Remark 2.2.

When the sequence (θn)n0(\theta_{n})_{n\geq 0} converges a.s. to a deterministic constant θ\theta_{\infty}, it is sufficient to assume that vv is continuous at θ\theta_{\infty} and v(θ)>0v(\theta_{\infty})>0 to ensure that Condition (4) is satisfied. Note that there is no need to impose that θ=θ\theta_{\infty}=\theta^{\star} although it is undoubtedly wished in practice. For instance, θ\theta_{\infty} can be an approximation of θ\theta^{\star} obtained either by heuristic arguments such as large deviations.

2.2 A Central limit theorem for the adaptive strong law of large numbers

To derive a central limit theorem for the adaptive estimator ξn\xi_{n}, we need a central limit theorem for locally square integrable martingales, whose convergence rate has been extensively studied. We refer to the works of Rebolledo [21], Jacod and Shiryaev [7], Hall and Heyde [6] and Whitt [25] to find different statements of central limit theorems for locally square integrable càdlàg martingales in continuous time, from which theorems can easily be deduced for discrete time locally square integrable martingales.

Theorem 2.3.

Assume Equation (1) and (2) hold. Let (θn)n0(\theta_{n})_{n\geq 0} be a n{\mathcal{F}}_{n}-adapted sequence with values in d{\mathbb{R}}^{d} such that for all n0n\geq 0, θn<\theta_{n}<\infty a.s and converging to some deterministic value θ\theta_{\infty}. Assume there exists η>0\eta>0 such that the function s2+η:θd𝔼(|H(θ,X)|2+η)s_{2+\eta}:\theta\in{\mathbb{R}}^{d}\longmapsto{\mathbb{E}}\left(|H(\theta,X)|^{2+\eta}\right) is finite for all θd\theta\in{\mathbb{R}}^{d} and continuous at θ\theta_{\infty}. Moreover, if vv is continuous at θ\theta_{\infty} and v(θ)>0v(\theta_{\infty})>0, then, n(ξn𝔼(Z))law𝒩(0,v(θ))\sqrt{n}(\xi_{n}-{\mathbb{E}}(Z))\xrightarrow[]{law}{\mathcal{N}}(0,v(\theta_{\infty})).

Proof.

We know from the proof of Theorem 2.1 that Mn=i=0n1H(θi,Xi+1)𝔼(Z)M_{n}=\sum_{i=0}^{n-1}H(\theta_{i},X_{i+1})-{\mathbb{E}}(Z) is a locally square integrable martingale and that 1nMn\frac{1}{n}\langle M\rangle_{n} converges a.s. to v(θ)v(\theta_{\infty}).

1ni=0n1𝔼(|H(θi,Xi+1)𝔼(Z)|2+η|i)c(1ni=0n1s2+η(θi)+𝔼(Z)2+η).\frac{1}{n}\sum_{i=0}^{n-1}{\mathbb{E}}(|H(\theta_{i},X_{i+1})-{\mathbb{E}}(Z)|^{2+\eta}|{\mathcal{F}}_{i})\leq c\left(\frac{1}{n}\sum_{i=0}^{n-1}s_{2+\eta}(\theta_{i})+{\mathbb{E}}(Z)^{2+\eta}\right).

The term on the r.h.s is bounded thanks to the continuity of s2+ηs_{2+\eta} at θ\theta_{\infty}. Hence, the local martingale (Mn)n(M_{n})_{n} satisfies Lindeberg’s condition. The result ensues from the central limit theorem for locally 𝕃2{\mathbb{L}}^{2} martingales. ∎

Corollary 2.4 (Effective central limit theorem with confidence interval).

Assume Equation (1) and (2) hold. Let (θn)n0(\theta_{n})_{n\geq 0} be a n{\mathcal{F}}_{n}-adapted sequence with values in d{\mathbb{R}}^{d} such that for all n0n\geq 0, θn<\theta_{n}<\infty a.s and converging to some deterministic value θ\theta_{\infty}. Assume there exists η>0\eta>0 such that the function s4+η:θd𝔼(|H(θ,X)|4+η)s_{4+\eta}:\theta\in{\mathbb{R}}^{d}\longmapsto{\mathbb{E}}\left(|H(\theta,X)|^{4+\eta}\right) is finite for all θd\theta\in{\mathbb{R}}^{d} and continuous at θ\theta_{\infty}. Then, σn2=1ni=0n1H(θi,Xi+1)2ξn2a.s.v(θ)\sigma_{n}^{2}=\frac{1}{n}\sum_{i=0}^{n-1}H(\theta_{i},X_{i+1})^{2}-\xi_{n}^{2}\xrightarrow[]{a.s.}v(\theta_{\infty}). If moreover v(θ)>0v(\theta_{\infty})>0, then nσn(ξn𝔼(Z))n+law𝒩(0,1)\frac{\sqrt{n}}{\sigma_{n}}(\xi_{n}-{\mathbb{E}}(Z))\xrightarrow[n\rightarrow+\infty]{law}{\mathcal{N}}(0,1).

Remark 2.5.

Even if v(θ)>0v(\theta_{\infty})>0, σn\sigma_{n} may take negative values for nn small. This corollary is really essential from a practical point of view because it proves that confidence intervals can be built as in the case of a crude Monte Carlo procedure. The only difference lies in the way of approximating the asymptotic variance.

The assumptions of Theorem 2.3 are fairly easy to check in practice since they are formulated independently of the sequence (θn)n(\theta_{n})_{n}. When θ=θ\theta_{\infty}=\theta^{\star}, which is nonetheless not required, the limiting variance is optimal in the sense that a crude Monte Carlo computation with the optimal parameter θ\theta^{\star} would have lead to the same limiting variance. These assumptions are satisfied in the frameworks introduced in Section 4.

3 Estimation of the optimal variance parameter

From Theorem 2.1 and Theorem 2.3, we know that if we can construct a convergent estimator (θn)n(\theta_{n})_{n} of θ\theta^{\star}, the adaptive estimator ξn\xi_{n} is a convergent and asymptotically normal estimator of the expectation 𝔼(Z){\mathbb{E}}(Z). The challenging issue is now to propose an automatic way of approximating the minimiser θ\theta^{\star} of v(θ)=𝔼(H(θ,X)2)𝔼(Z)2v(\theta)={\mathbb{E}}(H(\theta,X)^{2})-{\mathbb{E}}(Z)^{2}. In the following, we will assume that vv is strictly convex, goes to infinity at infinity and is continuously differentiable. Moreover, we assume that v\nabla v admits a representation as an expectation

v(θ)=𝔼(U(θ,X)),\nabla v(\theta)={\mathbb{E}}(U(\theta,X)),

where U:d×mdU:{\mathbb{R}}^{d}\times{\mathbb{R}}^{m}\longmapsto{\mathbb{R}}^{d} is a measurable and integrable function. We could see in the examples developed in Section 4 that these conditions are very easily satisfied. Stochastic algorithms such as the Robbins Monro algorithm (see [22]) are perfectly well suited to estimate quantities defined as the root of an expectation. Because for the applications we are targeting we cannot impose that 𝔼(|U(θ,X)|2)<c(1+|θ|2){\mathbb{E}}(|U(\theta,X)|^{2})<c(1+|\theta|^{2}), the Robbins-Monro algorithm will fail to converge and we need a more robust algorithm. This will naturally lead us to consider randomly truncated stochastic algorithms as introduced by Chen et al. [3]. When dealing with stochastic approximations, the idea of averaging the iterates comes out quite naturally to smooth the trajectories, see Section 3.2.

3.1 Randomly truncated stochastic algorithms

Let (Xn)n1{(X_{n})}_{n\geq 1} be an i.i.d sequence of random variables following the law of XX and (γn)n1{(\gamma_{n})}_{n\geq 1} be a decreasing sequence of a positive real numbers satisfying

nγn=andnγn2<.\sum_{n}\gamma_{n}=\infty\quad\mbox{and}\quad\sum_{n}\gamma_{n}^{2}<\infty. (5)

The sequence (γn)n(\gamma_{n})_{n} is often called the gain sequence or the step sequence. We define the σ\sigma-field n=σ(Xk,kn){\mathcal{F}}_{n}=\sigma(X_{k},\;k\leq n). We introduce an increasing sequence of compact sets (Kj)j(K_{j})_{j} of d{\mathbb{R}}^{d}

n=0Kn=dandKnK̊n+1\bigcup_{n=0}^{\infty}K_{n}\>=\>{\mathbb{R}}^{d}\quad\mbox{and}\quad K_{n}\subsetneq\mathring{K}_{n+1} (6)

Now, we can present the randomly truncated stochastic algorithm introduced in [3], which essentially consists in a truncation of the Robbins Monro algorithm on an increasing sequence of compact sets. For θ0K0\theta_{0}\in K_{0} and α0=0\alpha_{0}=0, we define the sequences of random variables (θn)n{(\theta_{n})}_{n} and (αn)n{(\alpha_{n})}_{n} by

{θn+12=θnγn+1U(θn,Xn+1),if θn+12𝒦αnθn+1=θn+12 and αn+1=αn,if θn+12𝒦αnθn+1=θ0 and αn+1=αn+1.\begin{cases}&\theta_{n+\frac{1}{2}}=\theta_{n}-\gamma_{n+1}U(\theta_{n},X_{n+1}),\\ \text{if $\theta_{n+\frac{1}{2}}\in\mathcal{K}_{\alpha_{n}}$}&\theta_{n+1}=\theta_{n+\frac{1}{2}}\quad\mbox{ and }\quad\alpha_{n+1}=\alpha_{n},\\ \text{if $\theta_{n+\frac{1}{2}}\notin\mathcal{K}_{\alpha_{n}}$}&\theta_{n+1}=\theta_{0}\quad\mbox{ and }\quad\alpha_{n+1}=\alpha_{n}+1.\end{cases} (7)

θn+12\theta_{n+\frac{1}{2}} is the new sample we draw, either we accept it and set θn+1=θn+12\theta_{n+1}=\theta_{n+\frac{1}{2}} or we reject it and reset the algorithm to θ0\theta_{0} when it tries to jump too far ahead in a single step. Note that θn+12\theta_{n+\frac{1}{2}} is actually drawn along the dynamics of the Robbins Monro algorithm and either we accept it as the new iterate or we reject it when the algorithm tries to jump to far ahead and in this case we reset the new iterate to θ0\theta_{0}. In the following, we write Equation (7) in a more condensed form

θn+1=𝒯Kαn(θnγn+1U(θn,Xn+1)),\theta_{n+1}={\cal T}_{K_{\alpha_{n}}}\left(\theta_{n}-\gamma_{n+1}U(\theta_{n},X_{n+1})\right), (8)

where 𝒯Kαn{\cal T}_{K_{\alpha_{n}}} denotes the truncation on the compact sets KαnK_{\alpha_{n}}.

The use of truncations enables to relax the hypotheses required to ensure the convergence. From the recent results of [16], we can state the following convergence result

Theorem 3.1.

Assume that

  • (A1)

    v\nabla v is continuous and there exists a unique θ\theta^{\star} s.t. v(θ)=0\nabla v(\theta^{\star})=0 and θθ\forall\;\theta\neq\theta^{\star}, (v(θ)|θθ)>0(\nabla v(\theta)\,|\,\theta-\theta^{\star})>0.

  • (A2)

    q>0\forall q>0, sup|θ|q𝔼(|U(θ,Z)|2)<\sup_{|\theta|\leq q}{\mathbb{E}}(|U(\theta,Z)|^{2})<\infty.

Then, the sequence (θn)n{(\theta_{n})}_{n} defined by (7) converges a.s. to θ\theta^{\star} for any sequence of compact sets satisfying (6) and moreover the sequence (αn)n{(\alpha_{n})}_{n} is a.s. finite.

Note that the assumptions required to ensure the convergence are very weak and are formulated independently of the algorithm trajectories, which makes them easy to check. Since the variance reduction technique we settle here aims at being automatic in the sense that it does not require any fiddling with the gain sequence depending on the function UU, it is quite natural to average the procedure defined by Equation (7).

3.2 Averaging a stochastic algorithm

This section is based on the remark that Cesaro type averages tend to smooth the behaviour of convergent estimators at least from a theoretical point of view. Such averaging techniques have already been studied and proved to provide asymptotically efficient estimators (see for instance [20], [14] or [19]).

At the same time, it is well known that true Cesaro averages are not so efficient from a practical point of view because the rate at which the impact of the first iterates vanishes in the average is too slow and it induces some kind of a numerical bias which in turn dramatically slows down the convergence. Combining these two facts has led us to consider a moving window average of Algorithm (7).
In this section, we restrict to gain sequences of the form γn=γ(n+1)a\gamma_{n}=\frac{\gamma}{(n+1)^{a}} with 12<a<1\frac{1}{2}<a<1. Let τ>0\tau>0 be the length of the window used for averaging. For n1n\geq 1, we introduce

θ^n(τ)=γpτi=pp+τ/γpθi with p=sup{k1:k+τ/γkn}n.\hat{\theta}_{n}(\tau)=\frac{\gamma_{p}}{\tau}\sum_{i=p}^{p+\lfloor\tau/\gamma_{p}\rfloor}\theta_{i}\quad\text{ with }p=\sup\{k\geq 1:k+\tau/\gamma_{k}\leq n\}\wedge n. (9)

We use the convention sup=+\sup\emptyset=+\infty. The almost sure convergence of (θ^n(τ))n(\hat{\theta}_{n}(\tau))_{n} can easily be deduced from Theorem 3.1. The asymptotic normality of the sequence (θ^n(τ))n(\hat{\theta}_{n}(\tau))_{n} has been studied in [15]. The definition of θ^n\hat{\theta}_{n} is a little different from the one used in [15] because we want to ensure that the sequence (θ^n)n(\hat{\theta}_{n})_{n} is adapted to the filtration n{\mathcal{F}}_{n} in view of the use of θ^n\hat{\theta}_{n} as an estimator of θ\theta^{\star} in Algorithm 1.2.

4 Examples of parametric Monte-Carlo settings

In this section, we give various examples of cases in which a parametric representation of the expectation of interest is available

𝔼(Z)=𝔼(H(θ,X)).{\mathbb{E}}(Z)={\mathbb{E}}(H(\theta,X)).

In each example, we highlight the strong convexity and the regularity of the function θ𝔼(H2(θ,X))\theta\longmapsto{\mathbb{E}}(H^{2}(\theta,X)) such that the minimiser θ\theta^{\star} is uniquely defined as the one root of θθ𝔼(H2(θ,X))\theta\longmapsto\nabla_{\theta}{\mathbb{E}}(H^{2}(\theta,X)).

4.1 Importance sampling for normal random variables

Let G=(G1,,Gd)G=(G_{1},\ldots,G_{d}) be a dd-dimensional standard normal random vector. For any measurable function h:dh:{\mathbb{R}}^{d}\longrightarrow{\mathbb{R}} such that 𝔼(|h(G)|)<{\mathbb{E}}(|h(G)|)<\infty, one has for all θd\theta\in{\mathbb{R}}^{d}

𝔼(h(G))=𝔼(eθG|θ|22h(G+θ)).{\mathbb{E}}\left(h(G)\right)={\mathbb{E}}\left(e^{-\theta\cdot G-\frac{|\theta|^{2}}{2}}h(G+\theta)\right). (10)

Assume we want to compute 𝔼(f(G)){\mathbb{E}}(f(G)) for a measurable function f:df:{\mathbb{R}}^{d}\longrightarrow{\mathbb{R}} such that f(G)f(G) is integrable. By applying equality (10) to h=fh=f and h(x)=f2(x)eθx+|θ|22h(x)=f^{2}(x)\mathop{\mathrm{e}^{-\theta\cdot x+\frac{|\theta|^{2}}{2}}}, one obtains that the expectation and the variance of the random variable f(G+θ)eθG|θ|22f(G+\theta)\mathop{\mathrm{e}^{-\theta\cdot G\nobreakspace-\frac{|\theta|^{2}}{2}}} are respectively equal to 𝔼(f(G)){\mathbb{E}}(f(G)) and v(θ)𝔼2(f(G))v(\theta)-{\mathbb{E}}^{2}(f(G)) where

v(θ)=𝔼(f2(G)eθG+|θ|22).v(\theta)={\mathbb{E}}\left(f^{2}(G)\mathop{\mathrm{e}^{-\theta\cdot G+\frac{|\theta|^{2}}{2}}}\right).

The strict convexity of the function vv is already known from [23] for instance. For the sake of completeness, we prove here a slightly improved version of this result.

Proposition 4.1.

Assume that

(f(G)0)>0,\displaystyle{\mathbb{P}}(f(G)\neq 0)>0, (11)
ε>0,𝔼(|f(G)|2+ε)<\displaystyle\exists\varepsilon>0,\;{\mathbb{E}}(|f(G)|^{2+\varepsilon})<\infty (12)

Then, vv is infinitely continuously differentiable and strongly convex.

Proof.

The function θf2(G)eθG+|θ|22\theta\mapsto f^{2}(G)e^{-\theta\cdot G+\frac{|\theta|^{2}}{2}} is infinitely continuously differentiable. Since,

sup|θ|M|θjf2(G)eθG+|θ|22|eM22f2(G)(M+(eGj+eGj))k=1d(eMGk+eMGk)\displaystyle\sup_{|\theta|\leq M}|\partial_{\theta^{j}}f^{2}(G)e^{-\theta\cdot G+\frac{|\theta|^{2}}{2}}|\leq e^{\frac{M^{2}}{2}}f^{2}(G)\left(M+(e^{G^{j}}+e^{-G^{j}})\right)\prod_{k=1}^{d}(e^{MG^{k}}+e^{-MG^{k}})

where the right hand side is integrable because by Hölder’s inequality and Equation (12), we have θd,𝔼(f2(G)eθG)<.\forall\theta\in{\mathbb{R}}^{d},{\mathbb{E}}\left(f^{2}(G)\mathop{\mathrm{e}^{-\theta\cdot G}}\right)<\infty. Lebesgue’s theorem ensures that vv is continuously differentiable with θjv(θ)=𝔼(f2(G)(θjGj)eθG+|θ|22)\frac{\partial}{\partial_{\theta^{j}}}v(\theta)={\mathbb{E}}\left(f^{2}(G)(\theta^{j}-G^{j})e^{-\theta\cdot G+\frac{|\theta|^{2}}{2}}\right). Higher order differentiability properties are obtained by similar arguments and in particular the Hessian matrix writes

2v(θ)=𝔼(f2(G)eθG+|θ|22))I+𝔼((θG)(θG)f2(G)eθG+|θ|22)\displaystyle\nabla^{2}v(\theta)={\mathbb{E}}\left(f^{2}(G)e^{-\theta\cdot G+\frac{|\theta|^{2}}{2}})\right)I+{\mathbb{E}}\left((\theta-G)(\theta-G)^{*}f^{2}(G)e^{-\theta\cdot G+\frac{|\theta|^{2}}{2}}\right)

The second term in the above equation is a positive semi-definite matrix, hence

2v(θ)𝔼(f2(G)eθG+|θ|22)=𝔼(f2(G)eθG)𝔼(eθG)𝔼(|f(G)|)2.\displaystyle\nabla^{2}v(\theta)\geq{\mathbb{E}}(f^{2}(G)e^{-\theta\cdot G+\frac{|\theta|^{2}}{2}})={\mathbb{E}}(f^{2}(G)e^{-\theta\cdot G}){\mathbb{E}}(e^{\theta\cdot G})\geq{\mathbb{E}}(|f(G)|)^{2}.

Assumption (11) ensures that 𝔼(|f(G)|)>0{\mathbb{E}}(|f(G)|)>0. Then, the Hessian matrix is uniformly bounded from below by the positive definite matrix 𝔼(|f(G)|)2I{\mathbb{E}}(|f(G)|)^{2}I. This yields the strong convexity of the function vv. ∎

Proposition 4.1 implies that vv has a unique minimiser θ\theta^{\star} characterised by v(θ)=0\nabla v(\theta^{\star})=0, i.e. 𝔼((θG)eθG+|θ|22f2(G))=0{\mathbb{E}}\left((\theta^{\star}-G)e^{-\theta^{\star}\cdot G+\frac{|\theta^{\star}|^{2}}{2}}f^{2}(G)\right)=0.

4.2 Importance sampling for processes

Equality (10) can actually be extended to the Brownian motion framework using Girsanov’s theorem. Let (Wt,0tT)(W_{t},0\leq t\leq T) be a dd-dimensional Brownian motion and {\mathcal{F}} its natural filtration. For any measurable and {\mathcal{F}}-predictable process (θt,0tT)(\theta_{t},0\leq t\leq T) such that 𝔼(e120T|θt|2𝑑t)<{\mathbb{E}}\left(\mathop{\mathrm{e}^{\frac{1}{2}\int_{0}^{T}|\theta_{t}|^{2}dt}}\right)<\infty, one has

𝔼(f(Wt,0tT))=𝔼(e0Tθt𝑑Wt120T|θt|2𝑑tf(Wt+0tθs𝑑s,0tT)).{\mathbb{E}}\left(f(W_{t},0\leq t\leq T)\right)={\mathbb{E}}\left(e^{-\int_{0}^{T}\theta_{t}\cdot dW_{t}-\frac{1}{2}\int_{0}^{T}|\theta_{t}|^{2}dt}f\left(W_{t}+\int_{0}^{t}\theta_{s}ds,0\leq t\leq T\right)\right).

Assume θt=θd\theta_{t}=\theta\in{\mathbb{R}}^{d}, for all t[0,T]t\in[0,T]. The variance of eθWTθ2T2f(Wt,0tT)\mathop{\mathrm{e}^{-\theta\cdot W_{T}-\frac{\theta^{2}T}{2}}}f(W_{t},0\leq t\leq T) writes down v(θ)𝔼(f2(Wt,0tT))v(\theta)-{\mathbb{E}}\left(f^{2}(W_{t},0\leq t\leq T)\right) with

v(θ)=𝔼(eθWT+|θ|22Tf2(Wt+θt, 0tT)).v(\theta)={\mathbb{E}}\left(e^{-\theta\cdot W_{T}+\frac{|\theta|^{2}}{2}T}f^{2}\left(W_{t}+\theta t,\,0\leq t\leq T\right)\right).

A similar result to Proposition 4.1 holds; in particular vv is infinitely continuously differentiable, strictly convex and goes to infinity at infinity.

For more general processes (θt,0tT)(\theta_{t},0\leq t\leq T), we refer the reader to [17].

4.3 The exponential change of measure

The idea of tilting some probability measure to find the ones that minimises the variance is a very common idea which can be also be applied to a wide range of distribution, see for instance the recent results of Kawai [11, 10] in which he applied an exponential change of measure to Lévy processes, also known as the Esscher transform.

Consider a random variable XX with values in d{\mathbb{R}}^{d} and cumulative generating function ψ(θ)=log𝔼(eθX)\psi(\theta)=\log{\mathbb{E}}\big{(}\mathop{\mathrm{e}^{\theta\cdot X}}\big{)}. We assume that ψ(θ)<\psi(\theta)<\infty for all θd\theta\in{\mathbb{R}}^{d}. Let pp denote the density of XX. We define the density pθp_{\theta} by

pθ(x)=p(x)eθxψ(θ),xd.p_{\theta}(x)=p(x)\mathop{\mathrm{e}^{\theta\cdot x-\psi(\theta)}},\quad x\in{\mathbb{R}}^{d}.

Let X(θ)X^{(\theta)} have pθp_{\theta} as a density, then

𝔼(f(X))=𝔼[f(X(θ))p(X(θ))pθ(X(θ))].{\mathbb{E}}(f(X))={\mathbb{E}}\left[f(X^{(\theta)})\frac{p(X^{(\theta)})}{p_{\theta}(X^{(\theta)})}\right].

The variance of f(X(θ))p(X(θ))pθ(X(θ))f(X^{(\theta)})\frac{p(X^{(\theta)})}{p_{\theta}(X^{(\theta)})} writes v(θ)𝔼(f(X(θ))2p(X(θ))2pθ(X(θ))2)v(\theta)-{\mathbb{E}}\Big{(}f(X^{(\theta)})^{2}\frac{p(X^{(\theta)})^{2}}{p_{\theta}(X^{(\theta)})^{2}}\Big{)} with

v(θ)=𝔼(|f(X(θ))p(X(θ))pθ(X(θ))|2)=𝔼(f(X)2eθX+ψ(θ)).v(\theta)={\mathbb{E}}\left(\left|f(X^{(\theta)})\frac{p(X^{(\theta)})}{p_{\theta}(X^{(\theta)})}\right|^{2}\right)={\mathbb{E}}\left(f(X)^{2}\mathop{\mathrm{e}^{-\theta\cdot X+\psi(\theta)}}\right).

Obviously, this change of measure is only valuable as a variance reduction technique if X(θ)X^{(\theta)} can be simulated at approximately the same cost as XX.

Proposition 4.2.

Assume that

ε>0,𝔼(|f(G)|2+ε)<\displaystyle\exists\varepsilon>0,\;{\mathbb{E}}(|f(G)|^{2+\varepsilon})<\infty (13)
lim|θ|pθ(x)=0 for all x in d\displaystyle\lim_{|\theta|\longrightarrow\infty}p_{\theta}(x)=0\text{ for all $x$ in ${\mathbb{R}}^{d}$} (14)

Then, vv is infinitely continuously differentiable, convex and lim|θ|v(θ)=\lim_{|\theta|\longrightarrow\infty}v(\theta)=\infty.

Proof.

To prove the differentiability of vv, it suffices to reproduce the first part of the proof of Proposition 4.1. The convexity of vv comes from the log-convexity of ψ\psi. Moreover,

v(θ)=𝔼(f(X)2eθX+ψ(θ))=𝔼(f(X)2p(X)pθ(X))v(\theta)={\mathbb{E}}\left(f(X)^{2}\mathop{\mathrm{e}^{-\theta\cdot X+\psi(\theta)}}\right)={\mathbb{E}}\left(f(X)^{2}\frac{p(X)}{p_{\theta}(X)}\right)

Combining Equation (14) with Fatou’s Lemma yields that lim|θ|v(θ)=\lim_{|\theta|\longrightarrow\infty}v(\theta)=\infty. ∎

Remark 4.3.

If XX is a random standard normal vector, pθ(x)=p(xθ)p_{\theta}(x)=p(x-\theta) and X(θ)X^{(\theta)} is a random normal vector with mean θ\theta and identity covariance matrix. Hence, we recover Equality (10).

5 Application to the Gaussian random vector framework

5.1 Presentation of the problem

We consider a DD-multidimensional local volatility model in which each asset is supposed to be driven by the following dynamics under the risk neutral measure.

dSti=Sti(rdt+σ(t,Sti)dWti),S0i=si.dS^{i}_{t}=S^{i}_{t}(rdt+\sigma(t,S^{i}_{t})\cdot dW^{i}_{t}),\ S^{i}_{0}=s^{i}.

W=(W1,,WD)W=(W^{1},\dots,W^{D})^{*} is a vector of correlated standard Brownian motions. The covariance structure of the Brownian motions is given by W,Wt=Γt\langle W,W\rangle_{t}=\Gamma t where Γ\Gamma is a definite positive matrix with a diagonal filled with ones. In our numerical examples, we take Γij=𝟏{i=j}+ρ𝟏{ij}\Gamma_{ij}={\bf 1}_{\{i=j\}}+\rho{\bf 1}_{\{i\neq j\}} with ρ(1D1,1)\rho\in(\frac{-1}{D-1},1) to ensure that the matrix Γ\Gamma is positive definite. The function σ\sigma is the local volatility function, rr is the instantaneous interest rate and the vector (s1,,sD)(s^{1},\dots,s^{D}) is the vector of the spot values. In this model, we want to price path-dependent options whose payoffs can be written as a function of (St,tT)(S_{t},t\leq T). Hence, the price is given by the discounted expectation erT𝔼(ψ(St,tT))\mathop{\mathrm{e}^{-rT}}{\mathbb{E}}(\psi(S_{t},t\leq T)). Most of the time, this expectation must be computed by Monte Carlo methods and one has to consider an approximation of ψ(St,tT)\psi(S_{t},t\leq T) on a time grid 0=t0<t1<<tN=T0=t_{0}<t_{1}<\dots<t_{N}=T. Then, the quantity of interest becomes

erT𝔼(ψ^(St0,St1,,StN)).\mathop{\mathrm{e}^{-rT}}{\mathbb{E}}(\hat{\psi}(S_{t_{0}},S_{t_{1}},\ldots,S_{t_{N}})).

The discretisation of the asset SS can for instance be obtained using an Euler scheme, which means that the function ψ^\hat{\psi} can be expressed in terms of the Brownian increments or equivalently using a random normal vector. These remarks finally turn the original pricing problem into the computation of an expectation of the form 𝔼(ϕ(G)){\mathbb{E}}(\phi(G)) where GG is a standard normal random vector in ND{\mathbb{R}}^{ND} and ϕ:N×D\phi:{\mathbb{R}}^{N\times D}\longmapsto{\mathbb{R}} is a measurable and integrable function. Using Equation (10), we have for all θd\theta\in{\mathbb{R}}^{d},

𝔼(ϕ(G))=𝔼(ϕ(G+Aθ)eAθG|Aθ|22),{\mathbb{E}}(\phi(G))={\mathbb{E}}\left(\phi(G+A\theta)e^{-A\theta\cdot G-\frac{|A\theta|^{2}}{2}}\right), (15)

where AA is d×NDd\times ND matrix. The particular choice d=NDd=ND and A=IdA=I_{d} corresponds to Equation (10). When D=1D=1, the choice d=1d=1 and A=(t1,t2t1,,tNtN1)A=(\sqrt{t_{1}},\sqrt{t_{2}-t_{1}},\dots,\sqrt{t_{N}-t_{N-1}})^{*} corresponds to adding a linear drift to the one dimensional standard Brownian motion WW and we recover the Cameron-Martin formula.

Transformation (15) actually relies on an importance sampling change of measure. Other strategies may be applicable such as stratification for instance as it is explained by Glasserman et al. in [5].

5.2 Bespoke estimators for the optimal variance parameter

It ensues from Proposition 4.1, that the second moment

v(θ)=𝔼(ψ(G+Aθ)2e2AθG|Aθ|2)=𝔼(ϕ(G)2eAθG+|Aθ|22)v(\theta)={\mathbb{E}}\left(\psi(G+A\theta)^{2}e^{-2A\theta\cdot G-|A\theta|^{2}}\right)={\mathbb{E}}\left(\phi(G)^{2}e^{-A\theta\cdot G+\frac{|A\theta|^{2}}{2}}\right) (16)

is strongly convex, infinitely differentiable and

v(θ)=𝔼(A(AθG)ϕ(G)2eAθG+|Aθ|22).\nabla v(\theta)={\mathbb{E}}\left(A^{*}(A\theta-G)\phi(G)^{2}e^{-A\theta\cdot G+\frac{|A\theta|^{2}}{2}}\right). (17)

If we apply Equation (10) again, we obtain an other expression for

v(θ)=𝔼(AGϕ(G+Aθ)2e2AθG+|Aθ|2).\nabla v(\theta)={\mathbb{E}}\left(-A^{*}G\phi(G+A\theta)^{2}e^{-2A\theta\cdot G+|A\theta|^{2}}\right). (18)

Let us introduce the following two functions

U1(θ,G)\displaystyle U^{1}(\theta,G) =A(AθG)ϕ(G)2eAθG+|Aθ|22,\displaystyle=A^{*}(A\theta-G)\phi(G)^{2}e^{-A\theta\cdot G+\frac{|A\theta|^{2}}{2}}, (19)
U2(θ,G)\displaystyle U^{2}(\theta,G) =AGϕ(G+Aθ)2e2AθG+|Aθ|2.\displaystyle=-A^{*}G\phi(G+A\theta)^{2}e^{-2A\theta\cdot G+|A\theta|^{2}}. (20)

Using either Equation (17) or (18), we can write v(θ)=𝔼(U2(θ,G))=𝔼(U1(θ,G))\nabla v(\theta)={\mathbb{E}}(U^{2}(\theta,G))={\mathbb{E}}(U^{1}(\theta,G)) and these two functions U1U^{1} and U2U^{2} fit in the framework of Section 3 and enable to construct two estimators of θ\theta^{\star} (θn1)n(\theta^{1}_{n})_{n} and (θn2)n(\theta^{2}_{n})_{n} following Equation (7)

θn+11\displaystyle\theta^{1}_{n+1} =𝒯Kαn(θn1γn+1U1(θn1,Gn+1)),\displaystyle={\cal T}_{K_{\alpha_{n}}}\left(\theta^{1}_{n}-\gamma_{n+1}U^{1}(\theta^{1}_{n},G_{n+1})\right), (21)
θn+12\displaystyle\theta^{2}_{n+1} =𝒯Kαn(θn1γn+1U2(θn2,Gn+1)),\displaystyle={\cal T}_{K_{\alpha_{n}}}\left(\theta^{1}_{n}-\gamma_{n+1}U^{2}(\theta^{2}_{n},G_{n+1})\right), (22)

where GnG_{n} is an i.i.d sequence of random variables following the law of GG. We also introduce their corresponding averaging versions (θ1^n)n(\widehat{\theta^{1}}_{n})_{n} and (θ2^n)n(\widehat{\theta^{2}}_{n})_{n} following Equation (9). Based on Equation (15), we define

H(θ,G)=ϕ(G+Aθ)eAθG|Aθ|22.H(\theta,G)=\phi(G+A\theta)e^{-A\theta\cdot G-\frac{|A\theta|^{2}}{2}}.

Corresponding to the different estimators of θ\theta^{\star} listed above, we can define as many approximations of 𝔼(ϕ(G)){\mathbb{E}}(\phi(G)) following Equation (3)

ξn1=1ni=1nH(θi11,Gi),\displaystyle\xi^{1}_{n}=\frac{1}{n}\sum_{i=1}^{n}H(\theta^{1}_{i-1},G_{i}),\quad ξn2=1ni=1nH(θi12,Gi),\displaystyle\xi^{2}_{n}=\frac{1}{n}\sum_{i=1}^{n}H(\theta^{2}_{i-1},G_{i}),
ξ^n1=1ni=1nH(θ^i11,Gi),\displaystyle\hat{\xi}^{1}_{n}=\frac{1}{n}\sum_{i=1}^{n}H(\hat{\theta}^{1}_{i-1},G_{i}),\quad ξ^n2=1ni=1nH(θ^i12,Gi),\displaystyle\hat{\xi}^{2}_{n}=\frac{1}{n}\sum_{i=1}^{n}H(\hat{\theta}^{2}_{i-1},G_{i}),

where the sequence GiG_{i} has already been used to build the (θn)n(\theta_{n})_{n} estimators. From Proposition 4.1 and Theorems 3.1, 2.1 and 2.3, we can deduce the following result.

Theorem 5.1.

If there exists ε>0\varepsilon>0 such that 𝔼(ϕ(G)4+ε)<{\mathbb{E}}(\phi(G)^{4+\varepsilon})<\infty then, the sequences (θn1)n(\theta^{1}_{n})_{n}, (θn2)n(\theta^{2}_{n})_{n}, (θ1^n)n(\widehat{\theta^{1}}_{n})_{n} and (θ2^n)n(\widehat{\theta^{2}}_{n})_{n} defined by Equations (7) or (9) converge a.s. to θ\theta^{\star} for any increasing sequence of compact sets (Kj)j(K_{j})_{j} satisfying (6) and the adaptive estimator (ξn1)n,(ξn2)n,(ξ^n1)n,(ξ^n2)n(\xi^{1}_{n})_{n},(\xi^{2}_{n})_{n},(\hat{\xi}^{1}_{n})_{n},(\hat{\xi}^{2}_{n})_{n} converge to 𝔼(ϕ(G)){\mathbb{E}}(\phi(G)) and are asymptotically normal with optimal limiting variance v(θ)v(\theta^{\star}).

Proof.

We only do the proof for (θn1)n(\theta^{1}_{n})_{n} and (θ1^n)n(\widehat{\theta^{1}}_{n})_{n} as the same ideas can be applied to (θn2)n(\theta^{2}_{n})_{n} and (θ2^n)n(\widehat{\theta^{2}}_{n})_{n}. We know from Proposition 4.1, that the function vv defined by Equation (16) is strongly convex and infinitely differentiable, hence v\nabla v satisfies Assumption (A(A1)). Let q>0q>0. For any θ\theta satisfying |θ|q|\theta|\leq q, we have

𝔼|U1(θ,G)|2\displaystyle{\mathbb{E}}\left|U^{1}(\theta,G)\right|^{2} 𝔼((A(Aq+|G|)2ϕ(G)4e2AθGeA2q2).\displaystyle\leq{\mathbb{E}}\left((\|A\|(\|A\|q+|G|)^{2}\phi(G)^{4}e^{-2A\theta\cdot G}e^{\|A\|^{2}q^{2}}\right).

Using Hölder’s inequality, it can easily be proved that the expectation on the right hand side is uniformly bounded for |θ|q|\theta|\leq q. Hence Assumption (A(A2)) is satisfied. Therefore, θn1\theta^{1}_{n} and θ^n1\hat{\theta}^{1}_{n} both converge to θ\theta^{\star}. Let η>0\eta>0,

𝔼|H(θ,G)|2+η\displaystyle{\mathbb{E}}\left|H(\theta,G)\right|^{2+\eta} =𝔼(|ϕ(G)|2+ηe(1+η)AθGe|Aθ|2(1+η)2).\displaystyle={\mathbb{E}}\left(|\phi(G)|^{2+\eta}e^{-(1+\eta)A\theta\cdot G}e^{\frac{|A\theta|^{2}(1+\eta)}{2}}\right).

Using the integrability of ϕ(G)\phi(G) and Hölder’s inequality, one can prove that the expectation on the .r.h.s is bounded for θ\theta in a ball. Moreover, combining this with Lebesgue’s theorem, we obtain that the functions θ𝔼|H(θ,G)|2\theta\longmapsto{\mathbb{E}}|H(\theta,G)|^{2} and θ𝔼|H(θ,G)|2+η\theta\longmapsto{\mathbb{E}}|H(\theta,G)|^{2+\eta} are continuous. Therefore, the convergence and asymptotic normality of ξn\xi_{n} issues from Theorems (2.1) and (2.3). ∎

Remark 5.2.

Theorem 5.1 extends the result of [2, Theorem 4]. Our result is valid for any increasing sequences of compact sets (Kj)j(K_{j})_{j} satisfying (6) whereas Arouna needed a condition on the compact sets to ensure the convergence of the (θn)n(\theta_{n})_{n} estimators. The only condition required is some integrability on the payoff function and nothing has to be checked along the algorithm paths, which is a great improvement from a practical point of view.

For the vast majority of payoff functions commonly used, the assumptions of Theorem 5.1 are always satisfied.

5.3 Numerical results

5.3.1 Complexity of the different approximations

In the introduction, we have presented two different strategies for implementing a variance reduction method based on the approximation of the optimal variance parameter. We know from Theorem 5.1, that the adaptive and non-adaptive algorithms both converges at the same rate and the same limiting variance. Therefore, to decide which one is the better, we have to compare their computational costs. In this section, we assume that the computational cost of the different algorithms is determined by the number of evaluations of the function ϕ\phi. We will see in the examples later that this assumption is realistic and therefore it becomes obvious that the averaging and non-averaging estimators of θ\theta^{\star} all have the same computational costs when implemented with expertise.

The non-adaptive algorithm

We know from [2, 23] that the sequential algorithm converges with a rate of v(θ)/n\sqrt{v(\theta^{\star})/n} if we have 2n2n samples at hand and want to implement the sequential algorithm by using the first nn samples for approximating θ\theta^{\star} and the last nn samples for actually computing the Monte Carlo estimator with the previously computed approximation of θ\theta^{\star}. Whatever approximation of θ\theta^{\star} is used, be it (θn1)n(\theta^{1}_{n})_{n}, (θn2)n(\theta^{2}_{n})_{n}, (θ1^n)n(\widehat{\theta^{1}}_{n})_{n} or (θ2^n)n(\widehat{\theta^{2}}_{n})_{n}, this algorithm requires 2n2n evaluations of the function ϕ\phi whereas a crude Monte Carlo method only evaluates the function ϕ\phi nn times and achieves a convergence rate of v(0)/n\sqrt{v(0)/n}, hence this method only becomes efficient when v(θ)v(0)/2v(\theta^{\star})\leq v(0)/2.

The adaptive algorithm

From Theorem 5.1, we know that the adaptive estimators (ξn1)n,(ξn2)n,(ξ^n1)n,(ξ^n2)n(\xi^{1}_{n})_{n},(\xi^{2}_{n})_{n},(\hat{\xi}^{1}_{n})_{n},(\hat{\xi}^{2}_{n})_{n} all converge with the same rate v(θ)/n\sqrt{v(\theta^{\star})/n} but as we will see it they do not have the same computational cost. First, let us concentrate on (ξn1)n(\xi^{1}_{n})_{n} and (ξ^n1)n(\hat{\xi}^{1}_{n})_{n}, at each iteration ii, the function ϕ\phi has to be computed twice : once at the point Gi+1+θi1G_{i+1}+\theta^{1}_{i} (or Gi+1+θ^i1G_{i+1}+\hat{\theta}^{1}_{i}) to update the Monte Carlo estimator and once at the point Gi+1G_{i+1} to update θi+11\theta^{1}_{i+1} or θ^i+11\hat{\theta}^{1}_{i+1}. Hence, the computation of ξn1\xi^{1}_{n} or ξ^n1\hat{\xi}^{1}_{n} requires 2n2n evaluations of the function ϕ\phi. Similarly, the computation of ξ^n2\hat{\xi}^{2}_{n} requires at each step 22 evaluations of the function ϕ\phi : one at the point Gi+1+θ^i2G_{i+1}+\hat{\theta}^{2}_{i} to update the Monte Carlo estimator and one at the point Gi+1+θi2G_{i+1}+\theta^{2}_{i} to update the stochastic algorithm. So the overall cost is still 2n2n evaluations of the function ϕ\phi. But looking closely at the computation of (ξn2)n(\xi^{2}_{n})_{n} immediately highlights the benefit of having put the parameter θ\theta back into the function ϕ\phi in the expression of v\nabla v : the updates of ξi+12\xi^{2}_{i+1} and θi+12\theta^{2}_{i+1} both use the evaluation of the function ϕ\phi at the same point Gi+1+θi2G_{i+1}+\theta^{2}_{i}. Hence, the computation of ξn2\xi^{2}_{n} only needs nn evaluations of the function ϕ\phi instead of 2n2n for all the others algorithms. Obviously, the computational costs of the different estimators cannot really be reduced to the number of times the function ϕ\phi is evaluated so one should not expect that computing ξn2\xi^{2}_{n} is twice less costly than the other estimators but we will see in the examples below that the estimator ξn2\xi^{2}_{n} is indeed faster than the others.

To shortly conclude on the complexity of the different algorithms, be they sequential or adaptive, one should bear in mind that all the estimators except (ξn2)n(\xi^{2}_{n})_{n} roughly require twice the computational time of the crude Monte-Carlo method.

5.3.2 Practical implementation

The choice of using Equations (7) or (9) to build an estimator of θ\theta^{\star} becomes really important when one has to implement the variance reduction procedure either by using Algorithms (1.1) or (1.2). Both the averaging and non-averaging strategies have pros and cons. The averaging algorithm theoretically converges a little slower but has a much smoother behaviour with respect to the proper adjustment of the gain sequence (γn)n(\gamma_{n})_{n}. Then, to have a robust estimator — in the sense that the numerical convergence of the estimator does not depend too much on the choice of of the gain sequence — the averaging procedure proves to be better in practice. The non averaging algorithm should converge a little faster even though we do not notice it in practise as the convergence oscillates too much and is far more sensitive to the proper choice of the sequence (γn)n(\gamma_{n})_{n}. Eventually, both algorithms produce very similar results regarding variance reduction; the averaging one is easier to tune but requires more computational time.

In the numerical experiments of this section, we compare the different algorithms on multi-asset options. The quantity “Var MC” denotes the variance of the crude Monte Carlo estimator computed on-line on a single run of the algorithm. The variance denoted “Var ξ2\xi^{2}” (resp. “Var ξ^2\hat{\xi}^{2}”) is the variance of the ADIS algorithm (see Algorithm 1.2) which uses (θn2)n(\theta^{2}_{n})_{n} (resp. (θ^n2)n(\hat{\theta}^{2}_{n})_{n}) to estimate θ\theta^{\star}. These variances are computed using the on-line estimator given by Corollary 2.4. These adaptive algorithms are also compared to the sequential strategy described by Algorithm 1.1 denoted by “θ2\theta^{2}+MC” or “θ^2\hat{\theta}^{2}+MC” depending on how θ\theta^{\star} is approximated. In all these algorithms, the matrix AA introduced in Equation (15) is chosen as the identity matrix. When AA is not the identity matrix, its purpose is to reduce the dimension of the space in which the optimal θ\theta^{\star} is searched and in such cases the algorithms will be call “reduced”. Note that for the comparison to be fair between the different strategies, we have used nn samples for the adaptive algorithms but 2n2n for the sequential algorithms so that they all satisfy a central limit theorem with the rate n\sqrt{n}.

Basket options

We consider options with payoffs of the form (i=1dωiSTiK)+(\sum_{i=1}^{d}\omega^{i}S_{T}^{i}-K)_{+} where (ω1,,ωd)(\omega^{1},\dots,\omega^{d}) is a vector of algebraic weights (enabling us to consider exchange options).

ρ\rho KK γ\gamma Price Var MC Var ξ2\xi^{2} Var ξ^2\hat{\xi}^{2}
0.1 45 1 7.21 12.24 1.59 1.10
55 10 0.56 1.83 0.19 0.14
0.2 50 0.1 3.29 13.53 1.82 1.76
0.5 45 0.1 7.65 43.25 6.25 4.97
55 0.1 1.90 14.74 1.91 1.4
0.9 45 0.1 8.24 69.47 10.20 7.78
55 0.1 2.82 30.87 2.7 2.6
Table 1: Basket option in dimension d=40d=40 with r=0.05r=0.05, T=1T=1, S0i=50S_{0}^{i}=50, σi=0.2\sigma^{i}=0.2, ωi=1d\omega^{i}=\frac{1}{d} for all i=1,,di=1,\dots,d and n=100 000n=100\,000.
Estimators MC ξ2\xi^{2} ξ^2\hat{\xi}^{2}
CPU time 0.85 0.9 1.64
Table 2: CPU times for the option of Table 1.

The results of Table 1 indicate that the adaptive algorithm using an averaging stochastic approximation outperforms not only the crude Monte Carlo approach but also the adapted algorithms using non-averaging stochastic approximation. The better performance of the algorithms using averaging estimators of θ\theta^{\star} comes from the better smoothness of the averaging algorithm (see Equation (9)). Nonetheless, these good results in terms of variance reduction must be considered together with their computation costs reported in Table 2. As explained in Section 5.3.1, we notice that the computational cost of the estimator ξ2\xi^{2} is very close to the one of the crude Monte Carlo estimator because the implementation made the most of the fact that the updates of ξi+12\xi^{2}_{i+1} and θi+12\theta^{2}_{i+1} both need to evaluate the function ϕ\phi at the same point. Note that, because this implementation trick cannot be applied to ξ^2\hat{\xi}^{2}, the adaptive algorithm using an averaging stochastic approximation is twice slower. For a given precision, the adaptive algorithm is between 55 and 1010 times faster.

Barrier Basket Options

We consider basket options in dimension DD with a discrete barrier on each asset. For instance, if we consider a Down and Out Call option, the payoff writes down (i=1DωiSTiK)+𝟏{iD,jN,StjiLi}(\sum_{i=1}^{D}\omega^{i}S_{T}^{i}-K)_{+}{\bf 1}_{\{\forall i\leq D,\;\forall j\leq N,\;S_{t_{j}}^{i}\geq L^{i}\}} where ω=(ω1,,ωD)\omega=(\omega^{1},\dots,\omega^{D}) is a vector of positive weights, L=(L1,,LD)L=(L^{1},\dots,L^{D}) is the vector of barriers, K>0K>0 the strike value and tN=Tt_{N}=T. We consider one time step per month, which means that for an option with maturity time T=2T=2, the number of time steps is N=24N=24. From now on, we fix D=5D=5. Hence if we use the identity matrix AA, the parameter θ\theta is of size N×D=120N\times D=120. Here, we propose to reduce the dimension of θ\theta and we will in Table 3 that it achieves almost the same variance reduction. Of course the matrix AA cannot be chosen independently of the structure of the problem. Remember that the vector GG actually corresponds to the increments of the Brownian motion BB with values in d{\mathbb{R}}^{d} on the grid (tk=kT/N,k=0,,N)(t_{k}=kT/N,k=0,\dots,N). We recall that we can simulate the Brownian motion BB on the time grid (tk)k(t_{k})_{k} by using the following equality in law

(Bt1Bt2BtN1BtN)=(t1ID000t1IDt2t1ID00tN1tN2ID0t1t2t1IDtN1tN2IDtNtN1ID)G\begin{pmatrix}B_{t_{1}}\\ B_{t_{2}}\\ \vdots\\ B_{t_{N-1}}\\ B_{t_{N}}\end{pmatrix}=\begin{pmatrix}\sqrt{t_{1}}I_{D}&0&0&\ldots&0\\ \sqrt{t_{1}}I_{D}&\sqrt{t_{2}-t_{1}}I_{D}&0&\ldots&0\\ \vdots&\ddots&\ddots&\ddots&\vdots\\ \vdots&\ddots&\ddots&\sqrt{t_{N-1}-t_{N-2}}I_{D}&0\\ \sqrt{t_{1}}&\sqrt{t_{2}-t_{1}}I_{D}&\ldots&\sqrt{t_{N-1}-t_{N-2}}I_{D}&\sqrt{t_{N}-t_{N-1}}I_{D}\end{pmatrix}G

where IDI_{D} is the identity matrix in dimension DD. If we choose

A=(t1IDt2t1IDtNtN1ID)A=\begin{pmatrix}\sqrt{t_{1}}I_{D}\\ \sqrt{t_{2}-t_{1}}I_{D}\\ \vdots\\ \vdots\\ \sqrt{t_{N}-t_{N-1}}I_{D}\end{pmatrix}

then the transformation G+AθG+A\theta corresponds to the transformation (Bt1+θt1,Bt2+θt2,,BtN+θtN)(B_{t_{1}}+\theta t_{1},B_{t_{2}}+\theta t_{2},\dots,B_{t_{N}}+\theta t_{N})^{*} and it reduces the effective dimension of the importance sampling parameter to D=5D=5 rather than DN=120DN=120.

KK γ\gamma Price Var MC Var ξ2\xi^{2} Var ξ^2\hat{\xi}^{2} Var Var ξ2\xi^{2} Var ξ^2\hat{\xi}^{2} Var
θ^2\hat{\theta}^{2}+MC θ^2\hat{\theta}^{2}+MC
reduced reduced reduced
45 0.5 2.37 22.46 4.92 3.52 2.59 2.64 2.62 2.60
50 1 1.18 10.97 1.51 1.30 0.79 0.80 0.80 0.79
55 1 0.52 4.85 0.39 0.38 0.19 0.24 0.23 0.19
Table 3: Down and Out Call option in dimension I=5I=5 with σ=0.2\sigma=0.2, S0=(50,40,60,30,20)S_{0}=(50,40,60,30,20), L=(40,30,45,20,10)L=(40,30,45,20,10), ρ=0.3\rho=0.3, r=0.05r=0.05, T=2T=2, ω=(0.2,0.2,0.2,0.2,0.2)\omega=(0.2,0.2,0.2,0.2,0.2) and n=100 000n=100\,000.
Estimators MC ξ2\xi^{2} ξ^2\hat{\xi}^{2} θ2\theta^{2} + MC ξ2\xi^{2} ξ^2\hat{\xi}^{2} θ2\theta^{2} + MC
reduced reduced reduced
CPU time 1.86 1.93 3.34 4.06 1.89 2.89 3.90
Table 4: CPU times for the option of Table 3.

First, we note from Table 3 that the reduced and non-reduced ADIS algorithm achieve almost the same variance reduction. Actually, it is even advisable to reduce the size of the importance sampling parameter to reduce the noise in the stochastic approximation and therefore in the adaptive Monte Carlo estimator. Comparing the columns “Var MC”, “Var ξ2\xi^{2}” and “Var ξ^2\hat{\xi}^{2}” points out that when the convergence of the estimator of θ\theta^{\star} is too slow the first iterates of the adaptive Monte Carlo estimators use wrong values of θ\theta and therefore cannot reach v(θ)v(\theta^{\star}) whereas if a sequential algorithm is used all the iterates of the Monte Carlo estimator use the same and better approximation of θ\theta. We can see in Table 3 that the variance of “ξ^2\hat{\xi}^{2}+MC” is half the one of ξ2\xi^{2} or ξ^2\hat{\xi}^{2} but its CPU time is twice the one of ξ2\xi^{2} as noted in Table 4.

The reduced algorithms are a little faster than the non-reduced ones but their real advantage is to converge much more stably and to achieve the same variance as “ξ^2\hat{\xi}^{2}+MC” but in far less computational time. As in the previous examples, we still observe that the estimator “ξ2\xi^{2} reduced” is faster than the others and has a variance very close to the best method which is “θ^2\hat{\theta}^{2}+MC”.

6 Conclusion

In this work, we have explained how one could devise an adaptive variance reduction method for computing an expectation with a free parameter. Different algorithms have been studied both from a theoretical point of view and in practice. Although all the adaptive algorithms satisfy the same central limit theorem, they may behave very differently in practice, in particular adaptive algorithms using a non-averaging stochastic approximation of the optimal variance parameter can be implemented in a clever way which makes them as fast as a crude Monte Carlo approach. Nevertheless, the numerical convergence of these stochastic is very sensitive to the tuning of their gain sequence and one way to smooth this behaviour is to plug an averaging procedure on top of the stochastic approximation but then the computational time significantly increases and yet the dependency with respect to the gain sequence is still a serious drawback. To encounter the fine tuning of the algorithm, Jourdain and Lelong [9] have recently suggested to use deterministic optimisation techniques coupled with sample approximation, but their technique can not be implemented in an adaptive manner which increases its computational cost.

References

  • [1] B. Arouna. Adaptative Monte Carlo method, a variance reduction technique. Monte Carlo Methods Appl., 10(1):1–24, 2004.
  • [2] B. Arouna. Robbins-Monro algorithms and variance reduction in finance. The Journal of Computational Finance, 7(2), Winter 2003/2004.
  • [3] H.F. Chen and Y.M. Zhu. Stochastic Approximation Procedure with randomly varying truncations. Scientia Sinica Series, 1986.
  • [4] Paul Glasserman. Monte Carlo methods in financial engineering, volume 53 of Applications of Mathematics (New York). Springer-Verlag, New York, 2004. , Stochastic Modelling and Applied Probability.
  • [5] Paul Glasserman, Philip Heidelberger, and Perwez Shahabuddin. Asymptotically optimal importance sampling and stratification for pricing path-dependent options. Math. Finance, 9(2):117–152, 1999.
  • [6] P. Hall and C. C. Heyde. Martingale limit theory and its application. Academic Press Inc. [Harcourt Brace Jovanovich Publishers], New York, 1980. Probability and Mathematical Statistics.
  • [7] J. Jacod and A.N. Shiryaev. Limit Theorems for Stochastic Processes. Springer-Verlag Berlin, 1987.
  • [8] B. Jourdain. Advanced Financial Modelling, chapter Adaptive variance reduction techniques in finance, pages 205–222. Radon Series Comp. Appl. Math 8. Walter de Gruyter, 2009.
  • [9] B. Jourdain and J. Lelong. Robust adaptive importance sampling for normal random vectors. Annals of Applied Probability, 19(5):1687–1718, 2009.
  • [10] Reiichiro Kawai. Adaptive monte carlo variance reducion for Lévy processes with two-time-scale stochastic approximation. Methodology and Computing in Applied Probability, 10(2):199–223, 2008.
  • [11] Reiichiro Kawai. Optimal importance sampling parameter search for Lévy processes via stochastic approximation. SIAM Journal on Numerical Analysis, 47(1):293–307, 2010.
  • [12] S. Kim and S. G. Henderson. Adaptive control variates. In Proceedings of the 2004 Winter Simulation Conference, 2004.
  • [13] Sujin Kim and Shane G. Henderson. Adaptive control variates for finite-horizon simulation. Math. Oper. Res., 32(3):508–527, 2007.
  • [14] Harold J. Kushner and Jichuan Yang. Stochastic approximation with averaging of the iterates: Optimal asymptotic rate of convergence for general processes. SIAM Journal on Control and Optimization, 31(4):1045–1062, 1993.
  • [15] J. Lelong. Etude asymptotique des algorithmes stochastiques et calcul du prix des options Parisiennes. PhD thesis, Ecole Nationale des Ponts et Chasusées, http://tel.archives-ouvertes.fr/tel-00201373/fr/, 2007.
  • [16] J. Lelong. Almost sure convergence of randomly truncated stochastic algorithms under verifiable conditions. Statistics & Probability Letters, 78(16), 2008.
  • [17] V. Lemaire and G. Pagès. Unconstrained recursive importance sampling. Annals of Applied Probability (to appear), 2009.
  • [18] D. Lépingle. Sur le comportement asymptotique des martingales locales. In Séminaire de Probabilités, XII (Univ. Strasbourg, Strasbourg, 1976/1977), volume 649 of Lecture Notes in Math., pages 148–161. Springer, Berlin, 1978.
  • [19] Mariane Pelletier. Asymptotic almost sure efficiency of averaged stochastic algorithms. SIAM J. Control Optim., 39(1):49–72 (electronic), 2000.
  • [20] B. T. Polyak and A. B. Juditsky. Acceleration of stochastic approximation by averaging. SIAM J. Control Optim., 30(4):838–855, 1992.
  • [21] Rolando Rebolledo. Central limit theorems for local martingales. Z. Wahrsch. Verw. Gebiete, 51(3):269–286, 1980.
  • [22] Herbert Robbins and Sutton Monro. A stochastic approximation method. Ann. Math. Statistics, 22:400–407, 1951.
  • [23] Yi Su and Michael C. Fu. Optimal importance sampling in securities pricing. Journal of Computational Finance, 5(4):27–50, 2002.
  • [24] Felisa J. Vázquez-Abad and Daniel Dufresne. Accelerated simulation for pricing asian options. In WSC ’98: Proceedings of the 30th conference on Winter simulation, pages 1493–1500, Los Alamitos, CA, USA, 1998. IEEE Computer Society Press.
  • [25] Ward Whitt. Proofs of the martingale FCLT. Probab. Surv., 4:268–302, 2007.