This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Hedging of time discrete auto-regressive stochastic volatility options

Joan del Castillo1 and Juan-Pablo Ortega2
Abstract

Numerous empirical proofs indicate the adequacy of the time discrete auto-regressive stochastic volatility models introduced by Taylor [Tay86, Tay05] in the description of the log-returns of financial assets. The pricing and hedging of contingent products that use these models for their underlying assets is a non-trivial exercise due to the incomplete nature of the corresponding market. In this paper we apply two volatility estimation techniques available in the literature for these models, namely Kalman filtering and the hierarchical-likelihood approach, in order to implement various pricing and dynamical hedging strategies. Our study shows that the local risk minimization scheme developed by Föllmer, Schweizer, and Sondermann is particularly appropriate in this setup, especially for at and in the money options or for low hedging frequencies.

Keywords: Stochastic volatility models, ARSV models, hedging techniques, incomplete markets, local risk minimization, Kalman filter, hierarchical-likelihood.

22footnotetext: Universitat Autònoma de Barcelona. E-08192 Barcelona, Spain. castillo@mat.uab.cat 44footnotetext: Corresponding author. Centre National de la Recherche Scientifique, Département de Mathématiques de Besançon, Université de Franche-Comté, UFR des Sciences et Techniques. 16, route de Gray. F-25030 Besançon cedex. France. Juan-Pablo.Ortega@univ-fcomte.fr

1 Introduction

Ever since Black, Merton, and Scholes introduced their celebrated option valuation formula [BS72, Mer76] a great deal of effort has been dedicated to reproduce that result using more realistic stochastic models for the underlying asset. In the discrete time modeling setup, the GARCH parametric family [Eng82, Bol86] has been a very popular and successful choice for which the option pricing and hedging problem has been profusely studied; see for example [Dua95, HN00, BEK+11, GCI10, Ort10], and references therein. Even though the GARCH specification can accommodate most stylized features of financial returns like leptokurticity, volatility clustering, and autocorrelation of squared returns, there are mathematical relations between some of their moments that impose undesirable constraints on some of the parameter values. For example, a well-known phenomenon [CPnR04] has to do with the relation that links the kurtosis of the process with the parameter that accounts for the persistence of the volatility shocks (the sum of the ARCH and the GARCH parameters, bound to be smaller than one in order to ensure stationarity). This relation implies that for highly volatility persistent time series like the ones observed in practice, the ARCH coefficient is automatically forced to be very small, which is in turn incompatible with having sizeable autocorrelation (ACF) for the squares (the ACF of the squares at lag 11 is linear in this coefficient). This situation aggravates when innovations with fat tails are used in order to better reproduce leptokurticity.

The structural rigidities associated to the finiteness of the fourth moment are of particular importance when using quadratic hedging methods for there is an imperative need to remain in the category of square summable objects and finite kurtosis is a convenient way to ensure that (see, for example [Ort10, Theorem 3.1 (iv)]).

The auto-regressive stochastic volatility (ARSV) models [Tay86, Tay05] that we will carefully describe later on in Section 2 are a parametric family designed in part to overcome the difficulties that we just explained. The defining feature of these models is the fact that the volatility (or a function of it) is determined by an autoregressive process that, unlike the GARCH situation, is exclusively driven by past volatility values and by designated innovations that are in principle independent from those driving the returns. The use of additional innovations introduces in the model a huge structural leeway because, this time around, the same constraint that ensures stationarity guarantees the existence of finite moments of arbitrary order which is of much value in the context of option pricing and hedging.

The modifications in the ARSV prescription that allow an enriching of the dynamics come with a price that has to do with the fact that conditional volatilities are not determined, as it was the case in the GARCH situation, by the price process. This causes added difficulties at the time of volatility and parameter estimation; in particular, a conditional likelihood cannot be written down in this context (see later on in Section 2.1). This also has a serious impact when these models are used in derivatives pricing and hedging because the existence of additional noise sources driving the dynamics makes more acute the incomplete character of the corresponding market; this makes more complex the pricing and hedging of contingent products.

The main goal of this paper is showing that appropriate implementations of the local risk minimization scheme developed by Föllmer, Schweizer, and Sondermann [FS86, FS91, Sch01] tailored to the ARSV setup provide an appropriate tool for pricing and hedging contingent products that use these models to characterize their underlying assets. The realization of this hedging scheme in the ARSV context requires two main tools:

  • Use of a volatility estimation technique: as we already said, volatility for these models is not determined by the price process and hence becomes a genuinely hidden variable that needs to be estimated separately. In our work we will use a Kalman filtering approach advocated by [HRS94] and the so called hierarchical likelihood (hh-likelihood) strategy [LN96, LN06, dCL08, WLLdC11] which, roughly speaking, consists of carrying out a likelihood estimation while considering the volatilities as unobserved parameters that are part of the optimization problem. Even though both methods are adequate volatility estimation techniques, the hh-likelihood technique has a much wider range of applicability for it is not subjected to the rigidity of the state space representation necessary for Kalman and hence can be used for stochastic volatility models with complex link functions or when innovations are non-Gaussian.

  • Use of a pricing kernel: this term refers to a probability measure equivalent to the physical one with respect to which the discounted prices are martingales. The expectation of the discounted payoff of a derivative product with respect to any of these equivalent measures yields an arbitrage free price for it. In our work we will also use these martingale measures in order to devise hedging strategies via local risk minimization. This is admittedly just an approximation of the optimal local risk minimization scheme that should be carried out with respect to the physical measure that actually quantifies the local risk. However there are good reasons that advise the use of these equivalent martingale measures mainly having to do with numerical computability and interpretation of the resulting value processes as arbitrage free prices. These arguments are analyzed in Section 3.2.

Regarding the last point, there are two equivalent martingale measures that we will be using. The first one is inspired by the Extended Girsanov Principle introduced in [EM98]. This principle allows the construction of a martingale measure in a general setup under which the process behaves as its “martingale component” used to do under the physical probability; this is the reason why this measure is sometimes referred to as the mean-correcting martingale measure, denomination that we will adopt. This construction has been widely used in the GARCH context (see [BEK+11, Ort10] and references therein) where it admits a particularly simple and numerically efficient expression in terms of the probability density function of the model innovations. As we will see, in the ARSV case, this feature is not anymore available and we will hence work with the expression coming from the predictable situation that even though in the ARSV case produces a measure that does not satisfy the Extended Girsanov Principle, it is still a martingale measure. Secondly, in Theorem 2.3 we construct the so called minimal martingale measure QminQ_{{\rm min}} in the ARSV setup; the importance in our context of this measure is given by the fact that the value process of the local risk-minimizing strategy with respect to the physical measure for a derivative product coincides with its arbitrage free price when using QminQ_{{\rm min}} as a pricing kernel. A concern with QminQ_{{\rm min}} in the ARSV setup is that this measure is in general signed; fortunately, the occurrence of negative Radon-Nikodym derivatives is extremely rare for the usual parameter values that one encounters in financial time series. Consequently, the bias introduced by censoring paths that yield negative Radon-Nikodym derivatives and using QminQ_{{\rm min}} as a well-defined positive measure is hardly noticeable. A point that is worth emphasizing is that even though the value processes obtained when carrying out local risk minimization with respect to the physical and the minimal martingale measures are identical, the hedges are in general not the same and consequently so are the hedging errors; this difference is studied in Proposition 3.6.

A general remark about local risk minimization that supports its choice in applications is its adaptability to different hedging frequencies. Most sensitivity based (delta) hedging methods for time series type underlyings are constructed by discretizing a continuous time hedging argument; consequently, when the hedging frequency diminishes, the use of such techniques loses pertinence. As we explain in Section 3.3, the local risk minimization hedging scheme can be adapted to prescribed changes in the hedging frequency, which regularly happens in real life applications.

The paper is organized in four sections. Section 2 contains a brief introduction to ARSV models and their dynamical features of interest to our study; this section contains two subsections that explain the volatility estimation techniques that we will be using and the martingale measures that we mentioned above. The details about the implementation of the local risk minimization strategy are contained in Section 3. Section 4 contains a numerical study where we carry out a comparison between the hedging performances obtained by implementing the local risk minimization scheme using the different volatility estimation techniques and the martingale measures introduced in Section 2. In order to enrich the discussion, we have added to the study two standard sensitivity based hedging methods that provide good results in other contexts, namely Black-Scholes [BS72] and an adaptation of Duan’s static hedge [Dua95] to the ARSV context. As a summary of the results obtained in that section it can be said that local risk minimization outperforms sensitivity based hedging methods, especially when maturities are long and the hedging frequencies are low; this could be expected due to the adaptability feature of local risk minimization with respect to modifications in this variable. As to the different measures proposed, it is the minimal martingale measure that yields the best hedging performances when implemented using a Kalman based estimation of volatilities, in the case of short and medium maturities and hh-likelihood for longer maturities and low hedging frequencies.

Conventions and notations: The proofs of all the results in the paper are contained in the appendix in Section 6. Given a filtered probability space (Ω,P,,{t}t)(\Omega,P,\mathcal{F},\{\mathcal{F}_{t}\}_{t\in\mathbb{N}}) and X,YX,Y two random variables, we will denote by Et[X]:=E[X|t]E_{t}[X]:=E[X|\mathcal{F}_{t}] the conditional expectation, covt(X,Y):=cov(X,Y|t):=Et[XY]Et[X]Et[Y]{\rm cov}_{t}(X,Y):={\rm cov}(X,Y|\mathcal{F}_{t}):=E_{t}[XY]-E_{t}[X]E_{t}[Y] the conditional covariance, and by vart(X):=Et[X2]Et[X]2{\rm var}_{t}(X):=E_{t}[X^{2}]-E_{t}[X]^{2} the conditional variance. A discrete-time stochastic process {Xt}t\{X_{t}\}_{t\in\mathbb{N}} is predictable when XtX_{t} is t1\mathcal{F}_{t-1}-measurable, for any tt\in\mathbb{N}.

Acknowledgments: we thank Alexandru Badescu for insightful comments about our work that have significantly improved its overall quality.

2 Auto-regressive stochastic volatility (ARSV) models

The auto-regressive stochastic volatility (ARSV) model has been introduced in [Tay86] with the objective of capturing some of the most common stylized features observed in the excess returns of financial time series: volatility clustering, excess kurtosis, and autodecorrelation in the presence of dependence; this last feature can be visualized by noticing that financial log-returns time series exhibit autocorrelation at, say lag 11, close to zero while the autocorrelation of the squared returns is significantly not null. Let StS_{t} be the price at time tt of the asset under consideration, rr the risk-free interest rate, and yt:=log(St/St1)y_{t}:=\log\left(S_{t}/S_{t-1}\right) the associated log-return. In this paper we will be considering the so-called standard ARSV model which is given by the prescription

{yt=r+σtϵt,{ϵt}IIDN(0,1)bt=γ+ϕbt1+wt,{wt}IIDN(0,σw2)\left\{\begin{array}[]{rcl}y_{t}&=&r+\sigma_{t}\epsilon_{t},\qquad\{\epsilon_{t}\}\sim{\rm IIDN}(0,1)\\ b_{t}&=&\gamma+\phi b_{t-1}+w_{t},\qquad\{w_{t}\}\sim{\rm IIDN}(0,\sigma_{w}^{2})\end{array}\right. (2.1)

where bt:=log(σt2)b_{t}:=\log(\sigma_{t}^{2}), γ\gamma is a real parameter, and ϕ(1,1)\phi\in(-1,1). Notice that in this model, the volatility process {σt}\{\sigma_{t}\} is a non-traded stochastic latent variable that, unlike the situation in GARCH like models [Eng82, Bol86] is not a predictable process that can be written as a function of previous returns and volatilities.

It is easy to prove that the unique stationary returns process induced by (2.1) available in the presence of the constraint ϕ(1,1)\phi\in(-1,1) is a white noise (the returns have no autocorrelation) with finite moments of arbitrary order. In particular, the marginal variance and kurtosis are given by

var(yt)=E[σt2]=exp[γ1ϕ+12σb2],kurtosis(yt)=3exp(σb2){\rm var}(y_{t})={\rm E}[\sigma_{t}^{2}]=\exp\left[\frac{\gamma}{1-\phi}+\frac{1}{2}\sigma_{b}^{2}\right],\qquad{\rm kurtosis}\,(y_{t})=3\exp\left(\sigma_{b}^{2}\right) (2.2)

Moreover, let σb2\sigma_{b}^{2} be the marginal variance of the stationary process {bt}\{b_{t}\}, that is,

σb2=σw21ϕ2.\sigma_{b}^{2}=\frac{\sigma_{w}^{2}}{1-\phi^{2}}.

It can be shown [Tay86] that whenever σb2\sigma_{b}^{2} is small and/or ϕ\phi is close to one then the autocorrelation γ(h)\gamma(h) of the squared returns at lag hh can be approximated by

γ(h)exp(σb2)13exp(σb2)1ϕh.\gamma(h)\simeq\frac{\exp(\sigma_{b}^{2})-1}{3\exp(\sigma_{b}^{2})-1}\phi^{h}.

The existence of finite moments is particularly convenient in the context of quadratic hedging methods. For example, in Theorem 3.1 of [Ort10] it is shown that the finiteness of kurtosis is a sufficient condition for the availability of adequate integrability conditions necessary to carry out pricing and hedging via local risk minimization.

Let (Ω,P)(\Omega,P) be the probability space where the model (2.1) has been formulated and let t\mathcal{F}_{t} be the information set generated by the observables {S0,S1,,St}\{S_{0},S_{1},\ldots,S_{t}\}. This statement can be mathematically coded by setting t=σ(S0,S1,,St)\mathcal{F}_{t}=\sigma\left(S_{0},S_{1},\ldots,S_{t}\right), where σ(S0,S1,,St)\sigma\left(S_{0},S_{1},\ldots,S_{t}\right) is the sigma algebra generated by the prices {S0,S1,,St}\{S_{0},S_{1},\ldots,S_{t}\}. As several equivalent probability measures will appear in our discussion, we will refer to PP as the physical or historical probability measure.

Unless we indicate otherwise, in the rest of the discussion we will not assume that the innovations {ϵt}\{\epsilon_{t}\} are Gaussian. We will also need the following marginal and conditional cumulant functions of ϵt\epsilon_{t} with respect to the filtration ={t}t{0,,T}\mathcal{F}=\{\mathcal{F}_{t}\}_{t\in\{0,\ldots,T\}}.

LϵtP(z)\displaystyle L_{\epsilon_{t}}^{P}(z) =\displaystyle= logEP[ezϵt],z,\displaystyle\log E^{P}\left[e^{z\epsilon_{t}}\right],\qquad z\in\mathbb{R},
KϵtP(u)\displaystyle K_{\epsilon_{t}}^{P}(u) =\displaystyle= logEP[euϵt|t1],with u a random variable.\displaystyle\log E^{P}\left[e^{u\epsilon_{t}}|{\cal F}_{t-1}\right],\qquad\text{with }u\mbox{ a random variable}.\quad

If the innovations {ϵt}\{\epsilon_{t}\} are Gaussian, we obviously have LϵtP(z)=z2/2L_{\epsilon_{t}}^{P}(z)=z^{2}/2. When the random variables uu and ϵt\epsilon_{t} are PP-independent, the following relation between LϵtPL_{\epsilon_{t}}^{P} and KϵtPK_{\epsilon_{t}}^{P} holds.

Lemma 2.1

Let u:(Ω,P)u:(\Omega,P)\rightarrow\mathbb{R} be a random variable independent of ϵt\epsilon_{t} and ={t}t{0,,T}\mathcal{F}=\{\mathcal{F}_{t}\}_{t\in\{0,\ldots,T\}} the filtration introduced above. Then

KϵtP(u)=logEP[eLϵtP(u)t1].K_{\epsilon_{t}}^{P}(u)=\log E^{P}\left[e^{L^{P}_{\epsilon_{t}}(u)}\mid{\cal F}_{t-1}\right]. (2.3)

The cumulant functions that we just introduced are very useful at the time of writing down the conditional means and variances of price changes with respect to the physical probability; these quantities will show up frequently in our developments later on. Before we provide these expressions we introduce the following notation for discounted prices:

St~:=Stert,\widetilde{S_{t}}:=S_{t}e^{-rt},

where rr denotes the risk-free interest rate. The tilde will be used in general for discounted processes. Using this notation and Lemma 2.1, a straightforward computation using the independence at any given time step tt of the innovations {ϵt}\{\epsilon_{t}\} and the volatility process {σt}\{\sigma_{t}\} defined by  (2.1), shows that

EP[St~S~t1t1]\displaystyle E^{P}\left[\widetilde{S_{t}}-\widetilde{S}_{t-1}\mid\mathcal{F}_{t-1}\right] =\displaystyle= S~t1(eKϵtP(σt)1)=S~t1EP[eLϵtP(σt)t1],\displaystyle\widetilde{S}_{t-1}\left(e^{K_{\epsilon_{t}}^{P}(\sigma_{t})}-1\right)=\widetilde{S}_{t-1}E^{P}\left[e^{L^{P}_{\epsilon_{t}}(\sigma_{t})}\mid{\cal F}_{t-1}\right], (2.4)
var[St~S~t1t1]\displaystyle{\rm var}\left[\widetilde{S_{t}}-\widetilde{S}_{t-1}\mid\mathcal{F}_{t-1}\right] =\displaystyle= S~t12[eKϵtP(2σt)e2KϵtP(σt)].\displaystyle\widetilde{S}_{t-1}^{2}\left[e^{K_{\epsilon_{t}}^{P}(2\sigma_{t})}-e^{2K_{\epsilon_{t}}^{P}(\sigma_{t})}\right]. (2.5)

2.1 Volatility and model estimation

The main complication in the estimation of ARSV models is due to the fact that t\mathcal{F}_{t} does not determine the random variable σt\sigma_{t} and hence makes impossible the writing of a likelihood function in a traditional sense. Many procedures have been developed over the years to go around this difficulty based on different techniques: moment matching [Tay86], generalized method of moments [MT90, JPR94, AS96], combinations of quasi-maximum likelihood with the Kalman filter [HRS94], simulated maximum likelihood [Dan94], MCMC [JPR94] and, more recently, hierarchical-likelihood [LN96, LN06, dCL08, WLLdC11] (abbreviated in what follows as h-likelihood). An excellent overview of some of these methods is provided in Chapter 11 of the monograph [Tay05].

In this paper we will focus on the Kalman and h-likelihood approaches since both are based on numerically efficient estimations of the volatility, a point of much importance in our developments.

State space representation and Kalman filtering. This approach is advocated by [HRS94] and consists of writing down the model (2.1) using a state space representation by setting

lt=log(|ytr|),Lt=log(σt).l_{t}=\log\left(|y_{t}-r|\right),\qquad L_{t}=\log\left(\sigma_{t}\right).

We assume in this paragraph that the innovations {ϵt}\{\epsilon_{t}\} are Gaussian. Using these variables, (2.1) can be rephrased as the observation and state equations:

{lt=Lt+ξt,ξt=log(|ϵt|),(LtαK)=ϕ(Lt1αK)+ηt,ηt=12wt,\left\{\begin{array}[]{rcl}l_{t}&=&L_{t}+\xi_{t},\qquad\xi_{t}=\log\left(|\epsilon_{t}|\right),\\ (L_{t}-\alpha_{K})&=&\phi(L_{t-1}-\alpha_{K})+\eta_{t},\qquad\eta_{t}=\frac{1}{2}w_{t},\end{array}\right. (2.6)

where αK=E[Lt]=12E[bt]=γ2(1ϕ)\alpha_{K}=E[L_{t}]=\frac{1}{2}E[b_{t}]=\frac{\gamma}{2(1-\phi)}. Kalman filtering cannot be directly applied to (2.6) due to the non-Gaussian character of the innovations ξt=log(|ϵt|)\xi_{t}=\log\left(|\epsilon_{t}|\right) hence, the approximation in [HRS94] consists of using this procedure by thinking of ξt\xi_{t} as a Gaussian variable with mean μξ:=E[ξt]=0.63518\mu_{\xi}:=E[\xi_{t}]=-0.63518 and variance σξ2=Var[ξt]=π2/8\sigma^{2}_{\xi}={\rm Var}[\xi_{t}]=\pi^{2}/8. The Kalman filter provides an algorithm to iteratively produce one-step ahead linear forecasts lt,1:=P[lt|t1]l_{t,1}:=P[l_{t}|\mathcal{F}_{t-1}]; given that lt=log(σt)+log(|ϵt|)l_{t}=\log(\sigma_{t})+\log\left(|\epsilon_{t}|\right), these forecasts can be used to produce predictable estimations σtk\sigma_{t}^{k} of the volatilities σt\sigma_{t} by setting log(σtk):=lt1,1μξ\log\left(\sigma_{t}^{k}\right):=l_{t-1,1}-\mu_{\xi} and hence

σtk=exp(lt1,1μξ).\sigma_{t}^{k}=\exp(l_{t-1,1}-\mu_{\xi}). (2.7)

The hh-likelihood approach. This method [LN96, LN06, dCL08, WLLdC11] consists of carrying out a likelihood estimation while considering the volatilities as unobserved parameters that are part of the optimization problem. More specifically, let α:=(γ,ϕ,σw)\alpha:=(\gamma,\phi,\sigma_{w}) be the parameters vector, b=(b1,,bT)b=(b_{1},\ldots,b_{T}) the vector that contains the unobserved bt:=log(σt2)b_{t}:=\log(\sigma_{t}^{2}), and let zt:=ytrz_{t}:=y_{t}-r; if f(zt|bt)f(z_{t}|b_{t}) and f(b|α)f(b|\alpha) denote the corresponding probability density functions, the associated hh-(log)likelihood is defined as

h(z;b,α)\displaystyle h(z;b,\alpha) =\displaystyle= t=1Tlogf(zt|bt)+logf(b|α)\displaystyle\sum_{t=1}^{T}\log f(z_{t}|b_{t})+\log f(b|\alpha) (2.8)
=\displaystyle= 12t=1T(zt2exp(bt)+bt+1σw2(btγϕbt1)2+logσw2).\displaystyle-\frac{1}{2}\sum_{t=1}^{T}\left(z_{t}^{2}\exp(-b_{t})+b_{t}+\frac{1}{\sigma_{w}^{2}}(b_{t}-\gamma-\phi b_{t-1})^{2}+\log\sigma_{w}^{2}\right).

The references listed above provide numerically efficient procedures to maximize (2.8) with respect to the variables bb and α\alpha once the sample zz has been specified. An idea of much importance in our context is that, once the coefficients α\alpha have been estimated, this procedure has a natural one-step ahead forecast of the volatility associated, similar to the one obtained using the Kalman filter. The technique operates using the variables bt:=log(σt2)b_{t}:=\log(\sigma_{t}^{2}) alternating prediction and updating; more specifically, we initialize the algorithm using the mean of the unique stationary solution of (2.1), that is, b0:=γ/(1ϕ)b_{0}:=\gamma/(1-\phi). We use the second equation in (2.1) to produce a linear prediction b1pb_{1p} of b1b_{1} based on the value of b0b_{0}, namely b1p=γ+ϕb0b_{1p}=\gamma+\phi b_{0}. Once the quote z1z_{1} is available, the h-likelihood approach can be used to update this forecast to the value b1ub_{1u} by solving the optimization problem

b1u=argminb1(z12exp(b1)+b1+1σw2(b1γϕb0)2).b_{1u}=\mathop{\rm arg\,min}_{b_{1}\in\mathbb{R}}\left(z_{1}^{2}\exp(-b_{1})+b_{1}+\frac{1}{\sigma_{w}^{2}}(b_{1}-\gamma-\phi b_{0})^{2}\right).

This value can be used in its turn to produce a forecast b2pb_{2p} for b2b_{2} by setting b2p=γ+ϕb1ub_{2p}=\gamma+\phi b_{1u}. If we continue this pattern we obtain a succession of prediction/updating steps given by the expressions

btp\displaystyle b_{tp} =\displaystyle= γ+ϕb(t1)u,\displaystyle\gamma+\phi b_{(t-1)u},
btu\displaystyle b_{tu} =\displaystyle= argminbt(zt2exp(bt)+bt+1σw2(btγϕb(t1)u)2).\displaystyle\mathop{\rm arg\,min}_{b_{t}\in\mathbb{R}}\left(z_{t}^{2}\exp(-b_{t})+b_{t}+\frac{1}{\sigma_{w}^{2}}(b_{t}-\gamma-\phi b_{(t-1)u})^{2}\right).

These forecasts can be used to produce predictable estimations σth\sigma_{t}^{h} of the volatilities σt\sigma_{t} by setting

σth:=exp(12btp).\sigma_{t}^{h}:=\exp\left(\frac{1}{2}b_{tp}\right).
Remark 2.2

Even though both the Kalman and the hh-likelihood methods constitute adequate estimation and volatility forecasting methods, the hh-likelihood technique has a much wider range of applicability for it is not subjected to the rigidity of the state space representation and hence can be generalized to stochastic volatility models with more complex link functions than the second expression in (2.1), to situations where the innovations are non-Gaussian, or there is a dependence between {ϵt}\{\epsilon_{t}\} and {wt}\{w_{t}\}.

2.2 Equivalent martingale measures for stochastic volatility models

Any technique for the pricing and hedging of options based on no-arbitrage theory requires the use of an equivalent measure for the probability space used for the modeling of the underlying asset under which its discounted prices are martingales with respect to the filtration generated by the observables. Measures with this property are usually referred to as risk-neutral or simply martingale measures and the Radon-Nikodym derivative that links this measure with the historical or physical one is called a pricing kernel.

A number of martingale measures have been formulated in the literature in the context of GARCH-like time discrete processes with predictable conditional volatility; see [BEK+11] for a good comparative account of many of them. Those constructions do not generalize to the SV context mainly due to the fact that the volatility process {σt}\{\sigma_{t}\} is not uniquely determined by the price process {St}\{S_{t}\} and hence it is not predictable with respect to the filtration ={t}\mathcal{F}=\{\mathcal{F}_{t}\} generated by {St}\{S_{t}\}.

In this piece of work we will explore two solutions to this problem that, when applied to option pricing and hedging, yields a good combination of theoretical computability and numerical efficiency. The first one has to do with the so called minimal martingale measure and the second approach is based on an approximation inspired in the Extended Girsanov Principle [EM98].

The minimal martingale measure. As we will see in the next section, this measure is particularly convenient when using local risk minimization with respect to the physical measure [FS86, FS91, FS02] as a pricing/hedging technique, for it provides the necessary tools to interpret the associated value process as an arbitrage free price for the contingent product under consideration.

The minimal martingale measure QminQ_{{\rm min}} is a measure equivalent to PP defined by the following property: every PP-martingale ML2(Ω,P)M\in L^{2}(\Omega,P) that is strongly orthogonal to the discounted price process S~\widetilde{S}, is also a QminQ_{{\rm min}}-martingale. The following result spells out the specific form that QminQ_{{\rm min}} takes when it comes to the ARSV models.

Theorem 2.3

Consider the price process S={S0,S1,,ST}S=\{S_{0},S_{1},\ldots,S_{T}\} associated to the ARSV model given by the expression (2.1). In this setup, the minimal martingale measure is determined by the Radon-Nikodym derivative dQmin/dPdQ_{{\rm min}}/dP that is obtained by evaluating at time TT the PP-martingale {Zt}t{1,,T}\{Z_{t}\}_{t\in\{1,\ldots,T\}} defined by

Zt:=k=1t(1+(eKϵtP(σk)1)(eσkϵkeKϵtP(σk))e2KϵtP(σk)eKϵtP(2σk)).Z_{t}:=\prod_{k=1}^{t}\left(1+\frac{\left(e^{K_{\epsilon_{t}}^{P}(\sigma_{k})}-1\right)\left(e^{\sigma_{k}\epsilon_{k}}-e^{K_{\epsilon_{t}}^{P}(\sigma_{k})}\right)}{e^{2K_{\epsilon_{t}}^{P}(\sigma_{k})}-e^{K_{\epsilon_{t}}^{P}(2\sigma_{k})}}\right). (2.9)

Proof.   By Corollaries 10.28 and 10.29 and Theorem 10.30 in [FS02], the minimal martingale measure, when it exists, is unique and is determined by the Radon-Nikodym derivative obtained by evaluating at TT the PP-martingale

Zt:=k=1t1+λk(Y~kY~k1),whereλk:=Ek1P[S~kS~k1]vark1(S~kS~k1),Z_{t}:=\prod_{k=1}^{t}1+\lambda_{k}\left(\widetilde{Y}_{k}-\widetilde{Y}_{k-1}\right),\quad\mbox{where}\quad\lambda_{k}:=-\frac{E^{P}_{k-1}\left[\widetilde{S}_{k}-\widetilde{S}_{k-1}\right]}{{\rm var}_{k-1}\left(\widetilde{S}_{k}-\widetilde{S}_{k-1}\right)}, (2.10)

and Y~k\widetilde{Y}_{k} is the martingale part in the Doob decomposition of S~k\widetilde{S}_{k} with respect to PP. Hence we have

Y~kY~k1=S~kEk1P[S~k]=S~k1(eσkϵkeKϵkP(σk)).\widetilde{Y}_{k}-\widetilde{Y}_{k-1}=\widetilde{S}_{k}-E^{P}_{k-1}\left[\widetilde{S}_{k}\right]=\widetilde{S}_{k-1}\left(e^{\sigma_{k}\epsilon_{k}}-e^{K^{P}_{\epsilon_{k}}(\sigma_{k})}\right). (2.11)

The equality (2.9) follows from (2.10), (2.11), (2.4), and (2.5).  \blacksquare

Remark 2.4

The measure QminQ_{{\rm min}} obtained by using (2.9) is in general signed. Indeed, as the random variable σkϵk\sigma_{k}\epsilon_{k} is k\mathcal{F}_{k}-adapted and KϵkP(σk)K_{\epsilon_{k}}^{P}(\sigma_{k}) is k\mathcal{F}_{k}-predictable, the term (eσkϵkeKϵkP(σk))\left(e^{\sigma_{k}\epsilon_{k}}-e^{K^{P}_{\epsilon_{k}}(\sigma_{k})}\right) can take arbitrarily negative values that can force ZtZ_{t} to become negative. We will see in our numerical experiments that even though this is in general the case, negative occurrences are extremely unlikely for the usual parameter values that one encounters in financial time series. Consequently, the bias introduced by censoring paths that yield negative Radon-Nikodym derivatives and using QminQ_{{\rm min}} as a well-defined positive measure is not noticeable.

Remark 2.5

The quantity KϵtP(σt)=logEP[eσtϵt|t1]K_{\epsilon_{t}}^{P}(\sigma_{t})=\log E^{P}\left[e^{\sigma_{t}\epsilon_{t}}|{\cal F}_{t-1}\right] in (2.9) is in general extremely difficult to compute. Lemma 2.1 allows us to rewrite it as a conditional expectation of a function that depends only on σt\sigma_{t} and hence it is reasonable to use the estimations σtk\sigma_{t}^{k} and σth\sigma_{t}^{h} introduced in Section 2.1 to approximate KϵtP(σt)K_{\epsilon_{t}}^{P}(\sigma_{t}). For example, if the innovations {ϵt}\{\epsilon_{t}\} are Gaussian and we hence have LϵtP(z)=z2/2L_{\epsilon_{t}}^{P}(z)=z^{2}/2, we will approximate

KϵtP(σt)=logEt1P[eσt22]K_{\epsilon_{t}}^{P}(\sigma_{t})=\log E^{P}_{t-1}\left[e^{\frac{\sigma_{t}^{2}}{2}}\right]

by (σtk)2/2(\sigma_{t}^{k})^{2}/2 or (σth)2/2(\sigma_{t}^{h})^{2}/2 depending on the technique (Kalman or hh-likelihood, respectively) used to estimate the conditional volatility.

The Extended Girsanov Principle and the mean correcting martingale measure. This construction has been introduced in [EM98] as an extension in discrete time and for multivariate processes of the classical Girsanov Theorem. This measure is designed so that when the process is considered with respect to it, its dynamical behavior coincides with that of its martingale component under the original historical probability measure. These martingale measures are widely used in the GARCH context (see for example [BEK+11, Ort10] and references therein).

In this paragraph we will work with models that are slightly more general than (2.1) in the sense that we will allow for predictable trend terms and generalized innovations, that is, our ARSV model will take the form

{yt=mt+σtϵt,{ϵt}IID(0,1)bt=γ+ϕbt1+wt,{wt}IID(0,σw2)\left\{\begin{array}[]{rcl}y_{t}&=&m_{t}+\sigma_{t}\epsilon_{t},\qquad\{\epsilon_{t}\}\sim{\rm IID}(0,1)\\ b_{t}&=&\gamma+\phi b_{t-1}+w_{t},\qquad\{w_{t}\}\sim{\rm IID}(0,\sigma_{w}^{2})\end{array}\right. (2.12)

with mtm_{t} an t1\mathcal{F}_{t-1}-measurable random variable.

Theorem 2.6

Given the model specified by (2.12), let fσtϵtPf_{\sigma_{t}\epsilon_{t}}^{P} be the conditional probability density function under PP of the random variable σtϵt\sigma_{t}\epsilon_{t} given t1\mathcal{F}_{t-1} and let {Nt}\{N_{t}\} be the stochastic discount factors defined by:

Nt:=fσtϵtP(σtϵt+mtr+logEt1P[eσtϵt])fσtϵtP(σtϵt).N_{t}:=\frac{f^{P}_{\sigma_{t}\epsilon_{t}}(\sigma_{t}\epsilon_{t}+m_{t}-r+\log E^{P}_{t-1}\left[e^{\sigma_{t}\epsilon_{t}}\right])}{f^{P}_{\sigma_{t}\epsilon_{t}}(\sigma_{t}\epsilon_{t})}.

The process Zt:=k=1tNkZ_{t}:=\prod_{k=1}^{t}N_{k}, is a (t,P)(\mathcal{F}_{t},P)-martingale such that EP[Zt]=1E^{P}[Z_{t}]=1 and ZTZ_{T} defines in the ARSV setup the martingale measure QEGPQ_{{\rm EGP}} associated to Extended Girsanov Principle via the identity ZT=dQEGP/dPZ_{T}=dQ_{{\rm EGP}}/dP.

Remark 2.7

As σt\sigma_{t} and ϵt\epsilon_{t} are independent processes, Rohatgi’s formula yields:

fσtϵtP(z)=1|x|fσtP(x)fϵtP(zx)𝑑x.f^{P}_{\sigma_{t}\epsilon_{t}}(z)=\int_{-\infty}^{\infty}\frac{1}{|x|}f^{P}_{\sigma_{t}}(x)f^{P}_{\epsilon_{t}}\left(\frac{z}{x}\right)dx.

Unfortunately, there is no closed form expression for this integral even in the Gaussian setup. Though ad hoc numerical have been developed to treat it (see for instance [GLD04]), the computation of the stochastic discount factors NtN_{t} for each time step tt may prove to be a computationally heavy task.

The point that we just raised in the remark leads us to introduce yet another martingale measure that mimics the good analytical properties of the Extended Girsanov Principle in the GARCH case but that, in the ARSV setup, only shares approximately its other defining features. In order to present this construction we define the market price of risk process:

ρt=mt+KϵtP(σt)rσt,\rho_{t}=\frac{m_{t}+K_{\epsilon_{t}}^{P}(\sigma_{t})-r}{\sigma_{t}}, (2.13)

to which we can associate what we will call the stochastic discount factor:

Nt(ϵt,ρt)=ftP(ϵt+ρt)ftP(ϵt),N_{t}(\epsilon_{t},\rho_{t})=\frac{f_{t}^{P}(\epsilon_{t}+\rho_{t})}{f_{t}^{P}(\epsilon_{t})},

where ftPf_{t}^{P} is the conditional probability density function of the innovations {ϵt}\{\epsilon_{t}\} with respect to the measure PP given t1\mathcal{F}_{t-1}.

Theorem 2.8

Using the notation that we just introduced, we have that

(i)

The process Zt:=k=1tNkZ_{t}:=\prod_{k=1}^{t}N_{k}, is a (t,P)(\mathcal{F}_{t},P)-martingale such that EP[Zt]=1E^{P}[Z_{t}]=1.

(ii)

ZTZ_{T} defines an equivalent measure QmcQ_{{\rm mc}} such that ZT=dQmc/dPZ_{T}=dQ_{{\rm mc}}/dP under which the discounted price process {S0,S1,,ST}\left\{S_{0},S_{1},\ldots,S_{T}\right\} determined by (2.12) is a martingale. By an abuse of language we will refer to QmcQ_{{\rm mc}} as the ARSV mean correcting martingale measure.

3 Local risk minimization for ARSV options

In this section we use the local risk minimization pricing/hedging technique developed by Föllmer, Schweizer, and Sondermann (see [FS86, FS91, Sch01], and references therein) as well as the different measures and volatility estimation techniques introduced in the previous section to come up with prices and hedging ratios for European options that have an ARSV process as model for the underlying asset.

As we will see, this technique provides simultaneous expressions for prices and hedging ratios that, even though require Monte Carlo computations in most cases, admit convenient interpretations based on the notion of hedging error minimization that can be adapted to various models for the underlying asset. This hedging approach has been studied in [Ort10] in the context of GARCH models with Gaussian innovations and in [BO11] for more general innovations. Additionally, the technique can be tuned in order to accommodate different prescribed hedging frequencies and hence adapts very well to realistic situations encountered by practitioners. In the negative side, like any other quadratic method, local risk minimization penalizes equally shortfall and windfall hedging errors, which may sometimes lead to inadequate hedging decisions.

3.1 Generalized trading strategies and local risk minimization

We now briefly review the necessary concepts on pricing by local risk minimization that we will needed in the sequel. The reader is encouraged to check with Chapter 10 of the excellent monograph [FS02] for a self-contained and comprehensive presentation of the subject.

As we have done so far, we will denote by StS_{t} the price of the underlying asset at time tt. The symbol rtr_{t} denotes the continuously composed risk-free interest rate paid on the currency of the underlying in the period that goes from time t1t-1 to tt; we will assume that {rt}\{r_{t}\} is a predictable process. Denote by

Rt:=j=0trj.R_{t}:=\sum_{j=0}^{t}r_{j}.

The price at time tt of the riskless asset S0S^{0} such that S00=1S_{0}^{0}=1, is given by St0=eRtS_{t}^{0}={\rm e}^{R_{t}}.

Let now H(ST)H(S_{T}) be a European contingent claim that depends on the terminal value of the risky asset StS_{t}. In the context of an incomplete market, it will be in general impossible to replicate the payoff HH by using a self-financing portfolio. Therefore, we introduce the notion of generalized trading strategy, in which the possibility of additional investment in the riskless asset throughout the trading periods up to expiry time TT is allowed. All the following statements are made with respect to a fixed filtered probability space (Ω,P,,{t}t{0,,T})(\Omega,P,\mathcal{F},\{\mathcal{F}_{t}\}_{t\in\{0,\ldots,T\}}):

  • A generalized trading strategy is a pair of stochastic processes (ξ0,ξ)(\xi^{0},\xi) such that {ξt0}t{0,,T}\{\xi^{0}_{t}\}_{t\in\{0,\ldots,T\}} is adapted and {ξt}t{1,,T}\{\xi_{t}\}_{t\in\{1,\ldots,T\}} is predictable. The associated value process VV of (ξ0,ξ)(\xi^{0},\xi) is defined as

    V0:=ξ00, and Vt:=ξt0St0+ξtSt,t1.V_{0}:=\xi_{0}^{0},\quad\mbox{ and }\quad V_{t}:=\xi_{t}^{0}\cdot S^{0}_{t}+\xi_{t}\cdot S_{t},\quad t\geq 1.
  • The gains process GG of the generalized trading strategy (ξ0,ξ)(\xi^{0},\xi) is given by

    G0:=0and Gt:=k=1tξk(SkSk1),t=0,,T.G_{0}:=0\quad\mbox{and }\quad G_{t}:=\sum_{k=1}^{t}\xi_{k}\cdot(S_{k}-S_{k-1}),\quad t=0,\ldots,T.
  • The cost process CC is defined by the difference

    Ct:=VtGt,t=0,,T.C_{t}:=V_{t}-G_{t},\quad t=0,\ldots,T.

These processes have discounted versions V~t\widetilde{V}_{t}, G~t\widetilde{G}_{t}, and C~t\widetilde{C}_{t} defined as:

V~t:=VteRt,G~t:=k=1tξk(S~kS~k1),andC~t:=V~tG~t.\widetilde{V}_{t}:=V_{t}{\rm e}^{-R_{t}},\qquad\widetilde{G}_{t}:=\sum_{k=1}^{t}\xi_{k}\cdot(\widetilde{S}_{k}-\widetilde{S}_{k-1}),\quad\mbox{and}\quad\widetilde{C}_{t}:=\widetilde{V}_{t}-\widetilde{G}_{t}.

Assume now that both HH and the {Sn}n{0,,T}\{S_{n}\}_{n\in\{0,\ldots,T\}} are in L2(Ω,P)L^{2}(\Omega,P). A generalized trading strategy is called admissible for HH whenever it is in L2(Ω,P)L^{2}(\Omega,P) and its associated value process is such that

VT=H,Pa.s.,Vt,GtL2(Ω,P),for eacht.V_{T}=H,\quad{\rm P}\ {\it a.s.},\quad\quad V_{t},G_{t}\in L^{2}(\Omega,P),\quad\mbox{for each}\ t.

The hedging technique via local risk minimization consists of finding the strategies (ξ^0,ξ^)(\widehat{\xi}^{0},\widehat{\xi}) that minimize the local risk process

Rt(ξ0,ξ):=EP[(C~t+1C~t)2t],t=0,,T1,R_{t}(\xi^{0},\xi):=E^{P}\left[(\widetilde{C}_{t+1}-\widetilde{C}_{t})^{2}\mid\mathcal{F}_{t}\right],\quad t=0,\ldots,T-1, (3.1)

within the set of admissible strategies (ξ0,ξ)(\xi^{0},\xi). More specifically, the admissible strategy (ξ^0,ξ^)(\widehat{\xi}^{0},\widehat{\xi}) is called local risk-minimizing if

Rt(ξ^0,ξ^)Rt(ξ0,ξ),Pa.s.R_{t}(\widehat{\xi}^{0},\widehat{\xi})\leq R_{t}(\xi^{0},\xi),\quad{\rm P}\ {\it a.s.}

for all tt and each admissible strategy (ξ0,ξ)(\xi^{0},\xi). It can be shown that [FS02, Theorem 10.9] an admissible strategy is local risk-minimizing if and only if the discounted cost process is a PP-martingale and it is strongly orthogonal to S~\widetilde{S}, in the sense that covt(S~t+1S~t,C~t+1C~t)=0{\rm cov}_{t}(\widetilde{S}_{t+1}-\widetilde{S}_{t},\widetilde{C}_{t+1}-\widetilde{C}_{t})=0, PP-a.s., for any t=0,,T1t=0,\ldots,T-1. More explicitly, whenever a local risk-minimizing technique exists, it is fully determined by the backwards recursions:

VT\displaystyle V_{T} =\displaystyle= H,\displaystyle H, (3.2)
ξt+1\displaystyle\xi_{t+1} =\displaystyle= covP(V~t+1,S~t+1S~tt)varP(S~t+1S~tt),\displaystyle\frac{{\rm cov}^{P}(\widetilde{V}_{t+1},\widetilde{S}_{t+1}-\widetilde{S}_{t}\mid\mathcal{F}_{t})}{{\rm var}^{P}(\widetilde{S}_{t+1}-\widetilde{S}_{t}\mid\mathcal{F}_{t})}, (3.3)
V~t\displaystyle\widetilde{V}_{t} =\displaystyle= EP[V~t+1t]ξt+1EP[(S~t+1S~t)t].\displaystyle E^{P}\left[\widetilde{V}_{t+1}\mid\mathcal{F}_{t}\right]-\xi_{t+1}E^{P}\left[\left(\widetilde{S}_{t+1}-\widetilde{S}_{t}\right)\mid\mathcal{F}_{t}\right]. (3.4)

The initial investment V0V_{0} determined by these relations will be referred to as the local risk minimization option price associated to the measure PP.

3.2 Local risk minimization with respect to a martingale measure

As we have explicitly indicated in expressions (3.2)–(3.4), the option values and the hedging ratios that they provide depend on a previously chosen probability measure and a filtration. In the absence of external sources of information, the natural filtration {t}\{\mathcal{F}_{t}\} to be used is the one generated by the price process. Regarding the choice of the probability measure, the first candidate that should be considered is the physical or historical probability associated to the price process due to the fact that from a risk management perspective this is the natural measure that should be used in order to construct the local risk (3.1).

In practice, the use of the physical probability encounters two major difficulties: on one hand, when (3.2)–(3.4) are written down with respect to this measure, the resulting expressions are convoluted and numerically difficult to estimate due to the high variance of the associated Monte Carlo estimators; a hint of this difficulty in the GARCH situation can be seen in Proposition 2.5 of [Ort10]. Secondly, unless the discounted prices are martingales with respect to the physical probability measure or there is a minimal martingale measure available, the option prices that result from this technique cannot be interpreted as arbitrage free prices.

These reasons lead us to explore the local risk minimization strategy for martingale measures and, more specifically for the martingale measures introduced in Section 2.2. Given that the trend terms that separate the physical measure from being a martingale measure are usually very small when dealing with daily or weekly financial returns, it is expected that the inaccuracy committed by using a conveniently chosen martingale measure will be smaller than the numerical error that we would face in the Monte Carlo evaluation of the expressions corresponding to the physical measure; see [Ort10, Proposition 3.2] for an argument in this direction in the GARCH context. In any case, the next section will be dedicated to carry out an empirical comparison between the hedging performances obtained using the different measures in Section 2.2, as well as the various volatility estimation techniques that we described and that are necessary to use them.

The following proposition is proved using a straightforward recursive argument in expressions (3.2)–(3.4) combined with the martingale hypothesis in its statement.

Proposition 3.1

Let QQ be an equivalent martingale measure for the price process {St}\{S_{t}\} and H(ST)H(S_{T}) be a European contingent claim that depends on the terminal value of the risky asset StS_{t}. The local risk minimizing strategy with respect to the measure QQ is determined by the recursions:

VT\displaystyle V_{T} =\displaystyle= H,\displaystyle H, (3.5)
ξt+1\displaystyle\xi_{t+1} =\displaystyle= 1Σt+12EtQ[e(RT+Rt)H(ST)(St+1ert+1St)],\displaystyle\frac{1}{\Sigma_{t+1}^{2}}E^{Q}_{t}\left[e^{-(R_{T}+R_{t})}H(S_{T})\left(S_{t+1}e^{-r_{t+1}}-S_{t}\right)\right], (3.6)
Vt\displaystyle V_{t} =\displaystyle= EtQ[e(RTRt)H(ST)],\displaystyle E^{Q}_{t}\left[e^{-(R_{T}-R_{t})}H(S_{T})\right], (3.7)

where

Σt+12:=varQ(S~t+1S~tt)=e2RtEtQ[St+12e2rt+1St2]=e2RtvarQ(St+1ert+1t).\Sigma_{t+1}^{2}:={\rm var}^{Q}(\widetilde{S}_{t+1}-\widetilde{S}_{t}\mid\mathcal{F}_{t})=e^{-2R_{t}}E^{Q}_{t}\left[S_{t+1}^{2}e^{-2r_{t+1}}-S_{t}^{2}\right]=e^{-2R_{t}}{\rm var}^{Q}(S_{t+1}{\rm e}^{-r_{t+1}}\mid\mathcal{F}_{t}). (3.8)
Remark 3.2

When expression (3.7) is evaluated at t=0t=0 it yields the initial investment V0V_{0} necessary to setup the generalized local risk minimizing trading strategy that fully replicates the derivative HH and coincides with the arbitrage free price for HH that results from using QQ as a pricing measure. Obviously, this connection only holds when local risk minimization is carried out with respect to a martingale measure.

Remark 3.3

Local risk-minimizing trading strategies computed with respect to a martingale measure QQ also minimize [FS02, Proposition 10.34] the so called remaining conditional risk, defined as the process RtQ(ξ0,ξ):=Et[(C~TC~t)2]R^{Q}_{t}(\xi^{0},\xi):=E_{t}[(\widetilde{C}_{T}-\widetilde{C}_{t})^{2}], t=0,,Tt=0,\ldots,T; this is in general not true outside the martingale framework (see [Sch01, Proposition 3.1] for a counterexample). Analogously, local risk minimizing strategies are also variance-optimal, that is, they minimize EQ[(H~V0G~T)2]E^{Q}\left[\left(\widetilde{H}-V_{0}-\widetilde{G}_{T}\right)^{2}\right] (see [FS02, Proposition 10.37]). This is particularly relevant in the ARSV context in which a standard sufficient condition (see [FS02, Theorem 10.40]) that guarantees that local risk minimization with respect to the physical measure implies variance optimality does not hold; we recall that this condition demands for the deterministic character of the quotient

βt:=(Et1P[StSt1])2vart1P[StSt1].\beta_{t}:=\frac{\left(E^{P}_{t-1}\left[S_{t}-S_{t-1}\right]\right)^{2}}{{\rm var}_{t-1}^{P}\left[S_{t}-S_{t-1}\right]}.

Indeed, expressions (2.4) and (2.5), together with Lemma 2.3 imply that in our situation

βt:=e2KϵtP(σt)[eKϵtP(2σt)e2KϵtP(σt)],\beta_{t}:=\frac{e^{2K_{\epsilon_{t}}^{P}(\sigma_{t})}}{\left[e^{K_{\epsilon_{t}}^{P}(2\sigma_{t})}-e^{2K_{\epsilon_{t}}^{P}(\sigma_{t})}\right]},

which is obviously not deterministic.

Remark 3.4

The last equality in (3.8) is of much importance when using Monte Carlo simulations to evaluate (3.5)–(3.7) and suggests the correct estimator that must be used in practice: if we generate under the measure QQ a set of NN price paths all of which start at StS_{t} at time tt and take values St+11,,St+1NS_{t+1}^{1},\ldots,S_{t+1}^{N} at time t+1t+1, then in view of the last equality in (3.8) we write

Σt+12=e2RtvarQ(St+1ert+1t)e2RtNi=1N(St+1iert+1St)2.\Sigma_{t+1}^{2}=e^{-2R_{t}}{\rm var}^{Q}(S_{t+1}{\rm e}^{-r_{t+1}}\mid\mathcal{F}_{t})\simeq\frac{e^{-2R_{t}}}{N}\sum_{i=1}^{N}\left(S_{t+1}^{i}{\rm e}^{-r_{t+1}}-S_{t}\right)^{2}.

We warn the reader that an estimator for Σt+12\Sigma_{t+1}^{2} based on the Monte Carlo evaluation of

e2RtEtQ[St+12e2rt+1St2]e^{-2R_{t}}E^{Q}_{t}\left[S_{t+1}^{2}e^{-2r_{t+1}}-S_{t}^{2}\right]

would have much more variance and once inserted in the denominator of (3.6) would produce unacceptable results.

Remark 3.5

The Monte Carlo estimation of the expressions (3.5)–(3.7) requires the generation of price paths with respect to the martingale measure QQ. In the GARCH context this can be easily carried out for a variety of pricing kernels by rewriting the process in connection with the new equivalent measure in terms of new innovations (see for example [Dua95, Ort10, GCI10, BEK+11]). This method can be combined with modified Monte Carlo estimators that enforce the martingale condition and that have a beneficial variance reduction effect like, for example, the Empirical Martingale Simulation technique introduced in [DS98, DGS01]. Unfortunately, this is difficult to carry out in our context and, more importantly, it does not produce good numerical results. Given that all the pricing measures introduced in the previous section are constructed out of the physical one, the best option in our setup consists of carrying out the path simulation with respect to the physical measure and computing the expectations in (3.5)–(3.7) using the corresponding Radon-Nikodym derivative. More specifically, let QQ be an equivalent martingale measure that is obtained out of the physical measure PP by constructing a Radon-Nikodym derivative of the form:

dQdP=k=1TNk,\frac{dQ}{dP}=\prod_{k=1}^{T}N_{k},

such that the process {Zt}\{Z_{t}\} defined by Zt:=k=1tNkZ_{t}:=\prod_{k=1}^{t}N_{k} is a PP-martingale that satisfies EP[Zt]=1E^{P}\left[Z_{t}\right]=1; notice that all the measures introduced in the previous section are of that form. In that setup we define the process {Ft}\{F_{t}\} by Ft:=i=tTNiF_{t}:=\prod_{i=t}^{T}N_{i} that allows us to rewrite QQ-expectations in terms of PP-expectations, that is, for any T\mathcal{F}_{T} measurable function ff and any t{0,,T}t\in\left\{0,\ldots,T\right\},

EQ[ft]=EP[Ft+1ft].E^{Q}\left[f\mid\mathcal{F}_{t}\right]=E^{P}\left[F_{t+1}f\mid\mathcal{F}_{t}\right].

The proof of this equality is a straightforward consequence of [FS02, Proposition A.11] and of the specific way in which the measure QQ is constructed.

3.3 Local risk minimization and changes in the hedging frequency

As we already pointed out one of the major advantages of the local risk minimization hedging scheme when compared to other sensitivity based methods, is its adaptability to prescribed changes in the hedging frequency. Indeed, suppose that the life of the option HH with maturity in TT time steps is partitioned into identical time intervals of duration jj; this assumption implies the existence of an integer kk such that kj=Tkj=T.

We now want to set up a local risk minimizing replication strategy for HH in which hedging is carried out once every jj time steps. We will denote by ξt+j\xi_{t+j} the hedging ratio at time tt that presupposes that the next hedging will take place at time t+jt+j. The value of such ratios will be obtained by minimizing the jj-spaced local risk process:

Rtj(ξ0,ξ):=EP[(C~t+jjC~tj)2t],t=0,j,2j,,(k1)j=Tj,R_{t}^{j}(\xi^{0},\xi):=E^{P}\left[(\widetilde{C}_{t+j}^{j}-\widetilde{C}_{t}^{j})^{2}\mid\mathcal{F}_{t}\right],\quad t=0,j,2j,\ldots,(k-1)j=T-j,

where {C~tj}\{\widetilde{C}_{t}^{j}\} is a cost process constructed out of value and gains processes, {Vtj}\{V_{t}^{j}\} and {Gtj}\left\{G_{t}^{j}\right\} that only take into account the prices of the underlying assets at time steps t=0,j,2j,,kj=Tt=0,j,2j,\ldots,kj=T, in particular, given an integer ll such that t=ljt=lj

G~tj:=r=1lξrj(S~rjS~(r1)j).\widetilde{G}_{t}^{j}:=\sum_{r=1}^{l}\xi_{rj}\cdot(\widetilde{S}_{rj}-\widetilde{S}_{(r-1)j}).

A straightforward modification of the argument in [FS02, Theorem 10.9] proves that the solution of this local risk minimization problem with modified hedging frequency with respect to a martingale measure is given by the expressions:

VTj\displaystyle V_{T}^{j} =\displaystyle= H,\displaystyle H, (3.9)
ξt+j\displaystyle\xi_{t+j} =\displaystyle= e(RTRt)EtQ[H(ST)(St+je(Rt+jRt)St)]EtQ[St+j2e2(Rt+jRt)St2],\displaystyle{\rm e}^{-(R_{T}-R_{t})}\frac{E^{Q}_{t}\left[H(S_{T})\left(S_{t+j}e^{-(R_{t+j}-R_{t})}-S_{t}\right)\right]}{E^{Q}_{t}\left[S_{t+j}^{2}{\rm e}^{-2(R_{t+j}-R_{t})}-S_{t}^{2}\right]}, (3.10)
Vtj\displaystyle V_{t}^{j} =\displaystyle= EtQ[e(RTRt)H(ST)],\displaystyle E^{Q}_{t}\left[e^{-(R_{T}-R_{t})}H(S_{T})\right], (3.11)

for any t=0,j,2j,,(k1)j=Tjt=0,j,2j,\ldots,(k-1)j=T-j. As a follow up to what we pointed out in the remark 3.4, the denominator in (3.10) should be computed by first noticing that

EtQ[St+j2e2(Rt+jRt)St2]=vartQ[St+je(Rt+jRt)],E^{Q}_{t}\left[S_{t+j}^{2}{\rm e}^{-2(R_{t+j}-R_{t})}-S_{t}^{2}\right]={\rm var}_{t}^{Q}\left[S_{t+j}{\rm e}^{-(R_{t+j}-R_{t})}\right],

and then using the appropriate Monte Carlo estimator for the variance, that is,

varQ(St+1ert+1t)1Ni=1N(St+jie(Rt+jiRt)St)2,{\rm var}^{Q}(S_{t+1}{\rm e}^{-r_{t+1}}\mid\mathcal{F}_{t})\simeq\frac{1}{N}\sum_{i=1}^{N}\left(S_{t+j}^{i}{\rm e}^{-(R_{t+j}^{i}-R_{t})}-S_{t}\right)^{2}, (3.12)

where St+j1,,St+jNS_{t+j}^{1},\ldots,S_{t+j}^{N} are NN realizations of the price process under QQ at time t+jt+j, using paths that have all the same origin StS_{t} at time tt. If the interest rate process {rt}\left\{r_{t}\right\} is constant and equal to rr, then (3.12) obviously reduces to

varQ(St+1ert+1t)1Ni=1N(St+jiejrSt)2.{\rm var}^{Q}(S_{t+1}{\rm e}^{-r_{t+1}}\mid\mathcal{F}_{t})\simeq\frac{1}{N}\sum_{i=1}^{N}\left(S_{t+j}^{i}{\rm e}^{-jr}-S_{t}\right)^{2}.

3.4 Local risk minimization and the minimal martingale measure

Local risk-minimization requires picking a particular probability measure in the problem. We also saw that it is only when discounted prices are martingales with respect to that measure that the obtained value process for the option in question coincides with the arbitrage free price that one would obtain by using that martingale measure as a pricing kernel. This observation particularly concerns the physical probability which one would take as the first candidate to use this technique since it is the natural measure to be used when quantifying the local risk.

The solution of this problem is one of the motivations for the introduction of the minimal martingale measure QminQ_{{\rm min}} that we defined in Section 2.2. Indeed, it can be proved that (see Theorem 10.22 in [FS02]) the value process of the local risk-minimizing strategy with respect to the physical measure coincides with arbitrage free price for HH obtained by using QminQ_{{\rm min}} as a pricing kernel. More specifically, if VtPV_{t}^{P} is the value process for HH obtained out of formulas (3.2)–(3.4) using the physical measure then

VtP=EtQmin[e(RTRt)H(ST)].V_{t}^{P}=E^{Q_{{\rm min}}}_{t}\left[e^{-(R_{T}-R_{t})}H(S_{T})\right]. (3.13)

We emphasize that, as we pointed out in Remark 2.4, the minimal martingale measure in the ARSV setup is in general signed. Nevertheless, the occurrences of negative Radon-Nikodym derivatives are extremely unlikely, at least when dealing with Gaussian innovations, which justifies its use for local risk minimization.

We also underline that even though the value processes obtained when carrying out local risk minimization with respect to the physical and the minimal martingale measures are identical, the hedges are in general not the same and consequently so are the hedging errors. In the next proposition we provide a relation that links both hedges and identify situations in which they coincide. Before we proceed we need to introduce the notion of global (hedging) risk process: it can be proved that once a probability measure PP has been fixed, if there exists a local risk-minimizing strategy (ξ0,ξ)(\xi^{0},\xi) with respect to it, then it is unique (see [FS02, Proposition 10.9]) and the discounted payoff H~\widetilde{H} can be decomposed as (see [FS02, Corollary 10.14])

H~=V0+G~T+L~T,\widetilde{H}=V_{0}+\widetilde{G}_{T}+\widetilde{L}_{T}, (3.14)

with G~t\widetilde{G}_{t} the discounted gains process associated to (ξ0,ξ)(\xi^{0},\xi) and L~t:=C~tC0\widetilde{L}_{t}:=\widetilde{C}_{t}-C_{0}, t=0,,Tt=0,\ldots,T a sequence {L~t}t{0,,T}\{\widetilde{L}_{t}\}_{t\in\{0,\ldots,T\}} that we will call the (discounted) global (hedging) risk process. {L~t}t{0,,T}\{\widetilde{L}_{t}\}_{t\in\{0,\ldots,T\}} is a square integrable PP-martingale that satisfies L0=0L_{0}=0 and that is strongly orthogonal to S~\widetilde{S} in the sense that

covP((L~t+1L~t)(S~t+1S~t)t)=0for anyt=0,,T1.{\rm cov}^{P}((\widetilde{L}_{t+1}-\widetilde{L}_{t})(\widetilde{S}_{t+1}-\widetilde{S}_{t})\mid\mathcal{F}_{t})=0\quad\mbox{for any}\quad t=0,\ldots,T-1.

The decomposition (3.14) shows that L~T\widetilde{L}_{T} measures how far H~\widetilde{H} is from the terminal value of the self-financing portfolio uniquely determined by the initial investment V0V_{0} and the trading strategy given by {ξt}\{\xi_{t}\} (see [LL08, Proposition 1.1.3]).

Proposition 3.6

Let H(ST)H(S_{T}) be a European contingent claim that depends on the terminal value of the risky asset StS_{t} and let {ξtQmin}t=1T\{\xi_{t}^{Q_{{\rm min}}}\}_{t=1}^{T} and {ξtP}t=1T\{\xi_{t}^{P}\}_{t=1}^{T} be the local risk minimizing hedges associated to the minimal martingale measure QminQ_{{\rm min}} and the physical measure PP, respectively. Let {L~tP}t=0T\left\{\widetilde{L}_{t}^{P}\right\}_{t=0}^{T} be the associated PP-global risk process. Then, for any t{1,,T}t\in\left\{1,\ldots,T\right\}:

ξtQmin=ξtP+Et1Qmin[L~tP(S~tS~t1)]vart1Qmin[S~tS~t1].\xi_{t}^{Q_{{\rm min}}}=\xi_{t}^{P}+\frac{E^{Q_{{\rm min}}}_{t-1}\left[\widetilde{L}_{t}^{P}(\widetilde{S}_{t}-\widetilde{S}_{t-1})\right]}{{\rm var}_{t-1}^{Q_{{\rm min}}}\left[\widetilde{S}_{t}-\widetilde{S}_{t-1}\right]}. (3.15)

If the processes {L~tP}t=0T\left\{\widetilde{L}_{t}^{P}\right\}_{t=0}^{T} and {S~t}t=0T\left\{\widetilde{S}_{t}\right\}_{t=0}^{T} are either QminQ_{{\rm min}}–strongly orthogonal or QminQ_{{\rm min}}–independent, then {ξtQmin}t=1T={ξtP}t=1T\{\xi_{t}^{Q_{{\rm min}}}\}_{t=1}^{T}=\{\xi_{t}^{P}\}_{t=1}^{T}.

4 Empirical study

The goal of this section is comparing the hedging performances obtained by implementing the local risk minimization scheme using the different volatility estimation techniques and martingale measures introduced in Section 2. In order to enrich the discussion, we will add to the list two standard sensitivity based hedging methods that provide good results in other contexts, namely standard Black-Scholes [BS72] and an adaptation of Duan’s static hedge [Dua95] to the ARSV context.

The way in which we proceed with the comparison consists of taking a particular ARSV model as price paths generating process and then carrying out the replication of a European call option according to the prescribed list of hedging methods for various moneyness, strikes and hedging frequencies. At the end of the life of each of this option, the square hedging error is recorded. This task is carried out for a number of independently generated price paths which allows us to estimate a mean square hedging error that we will use to compare the different hedging techniques.

The price generating ARSV process. The price process chosen for our simulations is obtained by taking the model that has as associated log-returns one of the ARSV prescriptions studied in [WLLdC11]. It consists of taking the following parameters in (2.1):

r=0.1/252,γ=0.821,ϕ=0.9,σw=0.675.r=0.1/252,\qquad\gamma=-0.821,\qquad\phi=0.9,\qquad\sigma_{w}=0.675.

The use of formulas (2.2) shows that the resulting log-returns are very leptokurtic (kurtosis=33.00{\rm kurtosis}=33.00) and volatile (the stationary variance is 9.011049.01\cdot 10^{-4} which amounts to an annualized volatility of 47%). The intentionality behind this choice is exacerbating the defects of the different hedging methods in order to better compare them.

The hedging methods. Given the price paths generated by the ARSV process detailed in the previous paragraph, we will carry out the replication of European call options that have them as underlying asset using local risk minimization with respect to the two martingale measures introduced in Section 2.2, namely:

  • ARSV mean correcting martingale measure introduced in Theorem 2.8. The hedging method is abbreviated as LRM-MCMM in the figures.

  • Minimal martingale measure. Denoted by QminQ_{{\rm min}} and introduced in Theorem 2.3. The hedging method is abbreviated as LRM-MMM in the figures. This measure may be signed in the ARSV context but this happens with extremely low probability (see Remark 2.4); in our simulations we did not encounter this situation even a single time. Local risk minimization with respect to this measure yields the same value process as the physical probability but not the same hedging ratios; Proposition 3.6 spells out the difference and provides sufficient conditions for them to coincide.

The construction of all these measures requires conditional expectations of functions of the conditional variance of the log-returns with respect to the filtration generated by the price process. As this filtration does not determine the conditional volatilities, those expectations cannot be easily computed; we hence proceed by estimating the conditional volatilities via Kalman filtering and a prediction-update routine based on the hh-likelihood approach, both briefly described in Section 2.1. Different volatility estimation techniques lead to different pricing/hedging kernels and hence to different hedging performances that we compare in our simulations.

We will complete the comparison exercise with two more methods:

  • Black-Scholes: we forget that the price process is generated via an ARSV model and we handle the hedging using the standard Black-Scholes delta [BS72] as if the underlying was a realization of a log-normal process with constant drift and volatility given by the drift and marginal volatility (given by the expression (2.2)) of the ARSV model specified above.

  • Duan’s static delta hedge: this is a generalization of the Black-Scholes delta hedge that has been introduced in [Dua95, Corollary 2.4] in the GARCH context. In that result the author computes the derivative of a European call price (the delta put can be readily obtained out of the call-put parity) with respect to the price of the underlying in an arbitrary incomplete situation in which that option price is simply obtained as the expectation of the discounted payoff with respect to a given pricing kernel QQ, that is:

    Vt=EQ[e(RTRt)H(ST)t].V_{t}=E^{Q}\left[e^{-(R_{T}-R_{t})}H(S_{T})\mid\mathcal{F}_{t}\right].

    In those circumstances:

    ξt+1Duan:=VtSt=e(RTRt)EQ[STSt𝟏STKt].\xi^{{\rm Duan}}_{t+1}:=\frac{\partial V_{t}}{\partial S_{t}}=e^{-(R_{T}-R_{t})}E^{Q}\left[\frac{S_{T}}{S_{t}}\boldsymbol{1}_{S_{T}\geq K}\mid\mathcal{F}_{t}\right].

    In our empirical study we will use this hedging technique using the various pricing kernels previously introduced. A dynamical refinement of this hedge can be obtained by considering additional dependences of the value process on the underlying price that may occur (like in our case) through the volatility process [GR99, BO11]. This will not be considered in our study.

Numerical specifications. The comparisons in performance will be carried out by hedging batches of European call options with different maturities and three different moneyness S0/KS_{0}/K of 1.11, 1, and 0.90, and for a number of independent price paths. The initial price S0S_{0} of the underlying is always equal to 100100. We will assign to each hedging experiment a prescribed hedging frequency. The mean square hedging error associated to each individual hedging method will be computed by taking the mean of the terminal square hedging errors for each of the price paths. In the particular case of the local risk minimization method, where the hedges are obtained by computing the expressions (3.9)–(3.11), the conditional expectations that they contain will be estimated via Monte Carlo evaluations based on the use of 25002500 price paths.

The hedging exercises. We now explain in detail the different hedging experiments that we have carried out.

  • Exercise 1: all the hedging techniques explained in the paper are used. The four maturities considered are 66, 88, 1010, and 1212 time steps. Hedging is carried out daily. The computation of the mean square hedging error is based on 10001000 price paths. The first thing that we notice is that, even for these short maturities, the performance of Duan’s static hedge is very poor. This is specific to the ARSV situation for in the GARCH case this technique provides very competitive results [Ort10] even in the presence of processes excited by leptokurtic innovations [BO11]. Additionally, the Black-Scholes hedging scheme provides remarkably good results, specially for in and at the money options. Regarding the local risk minimization strategy, the best results are obtained when using a combination of a Kalman based estimation of the conditional volatility and the minimal martingale measure QminQ_{{\rm min}}.

    Refer to caption
    Figure 4.1: Experiment 1. All the hedging techniques explained in the paper are used. The four maturities considered are 66, 88, 1010, and 1212 time steps. Hedging is carried out daily. The computation of the mean square hedging error is based on 10001000 price paths.
  • Exercise 2: we have retaken the specifications of Exercise 11, but this time around we have eliminated from the list of techniques under examination Duan’s static hedge due to its poor performance. The four maturities considered are 1010, 2020, 3030, and 4040 time steps. Hedging is carried out every 1010 days. The main difference with the previous exercise is the degradation of the performance of the Black-Scholes scheme in comparison with local risk minimization. This is due to the fact that, unlike Black-Scholes, local risk minimization admits adjustment to the change of hedging frequency by using the formulas (3.9)–(3.11). Regarding local risk minimization it is still the combination of the minimal martingale measure with Kalman based estimation for the volatilities that performs the best.

    Refer to caption
    Figure 4.2: Experiment 2. The four maturities considered are 1010, 2020, 3030, and 4040 time steps. Hedging is carried out every 1010 days. The computation of the mean square hedging error is based on 10001000 price paths.
  • Exercise 3: we experiment with longer maturities and smaller hedging frequencies. We consider maturities of 2020, 4040, 6060, 8080, 100100, and 120120 time steps with hedging carried out every 2020 days. The computation of the mean square hedging error is based on 600600 price paths. The novelty in these experiments with respect to the preceding ones is that in some instances the volatility estimation via hh–likelihood seems more appropriate than Kalman’s method in this setup.

    Refer to caption
    Figure 4.3: Experiment 6. The hedging techniques used are spelled out in the legend. Six maturities are considered: 4040, 6060, 8080, and 100100 time steps. Hedging is carried out every 2020 days. The computation of the mean square hedging error is based on 600600 price paths.

5 Conclusions

In this work we have reported on the applicability of the local risk minimization pricing/hedging technique in the handling of European options that have an auto-regressive stochastic volatility (ARSV) model as their underlying asset. The method has been implemented using the following martingale measures:

  • The minimal martingale measure: even though it is in general signed, the probability of occurrence of negative Radon-Nikodym derivatives is extremely low and hence can be used to serve our purposes.

  • Mean correcting martingale measure: it is a measure that shares the same expression and good analytical and numerical properties as the measure obtained in the GARCH context out of the Extended Girsanov Principle. The reason to use this proxy instead of the genuine Extended Girsanov Principle is that in the ARSV case, this procedure produces a measure that is numerically expensive to evaluate.

An added difficulty when putting this scheme into practice and that is specific to ARSV models, has to do with the need of estimating the volatility that, even though in this situation is not determined by the observable price process, it is necessary in order to construct the pricing/hedging kernels. Two techniques are used to achieve that: Kalman filtering and hierarchical likelihood (hh-likelihood). Both methods are adequate volatility estimation techniques, however the hh-likelihood technique has a much wider range of applicability for it is not subjected to the rigidity of the state space representation necessary for Kalman and hence can be used for stochastic volatility models with complex link functions or when innovations are non-Gaussian.

We have carried out a numerical study to compare the hedging performance of the methods based on different measures and volatility estimation techniques. Overall, the most appropriate technique seems to be the choice that combines the minimal martingale measure with a Kalman filtering based estimation of the volatility. hh-likelihood based volatility estimation gains pertinence as maturities become longer and the hedging frequency diminishes.

6 Appendix

6.1 Proof of Lemma 2.1

Let ZZ be an arbitrary t1\mathcal{F}_{t-1} measurable random variable. As ϵt\epsilon_{t} is independent of t1\mathcal{F}_{t-1} and by hypothesis it is also independent of uu, the joint law Pϵt,u,ZP_{\epsilon_{t},u,Z} of the random variable (ϵt,u,Z):(Ω,P)3(\epsilon_{t},u,Z):(\Omega,P)\rightarrow\mathbb{R}^{3} can be written as the product Pϵt,u,Z(x,y,z)=Pϵt(x)Pu,Z(y,z)P_{\epsilon_{t},u,Z}(x,y,z)=P_{\epsilon_{t}}(x)P_{u,Z}(y,z). Hence, using Fubini’s Theorem we have

EP[euϵtZ]\displaystyle E^{P}\left[e^{u\epsilon_{t}}Z\right] =\displaystyle= eyxz𝑑Pϵt(x)𝑑Pu,Z(y,z)=[eyx𝑑Pϵt(x)]z𝑑Pu,Z(y,z)\displaystyle\int\int e^{yx}zdP_{\epsilon_{t}}(x)dP_{u,Z}(y,z)=\int\left[\int e^{yx}dP_{\epsilon_{t}}(x)\right]zdP_{u,Z}(y,z) (6.1)
=\displaystyle= eLϵt(y)z𝑑Pu,Z(y,z)=EP[eLϵt(u)Z].\displaystyle\int e^{L_{\epsilon_{t}}(y)}zdP_{u,Z}(y,z)=E^{P}\left[e^{L_{\epsilon_{t}}(u)}Z\right].

At the same time, by the definition of KϵtPK_{\epsilon_{t}}^{P}, for any t1\mathcal{F}_{t-1} measurable random variable ZZ, the equality (6.1) implies that:

EP[eKϵtP(u)Z]=EP[euϵtZ]=EP[eLϵt(u)Z].E^{P}\left[e^{K_{\epsilon_{t}}^{P}(u)}Z\right]=E^{P}\left[e^{u\epsilon_{t}}Z\right]=E^{P}\left[e^{L_{\epsilon_{t}}(u)}Z\right].

As eKϵtP(u)e^{K_{\epsilon_{t}}^{P}(u)} is t1\mathcal{F}_{t-1} measurable and ZZ arbitrary, the almost sure uniqueness of the conditional expectation implies that

eKϵtP(u)=EP[eLϵt(u)t1],e^{K_{\epsilon_{t}}^{P}(u)}=E^{P}\left[e^{L_{\epsilon_{t}}(u)}\mid\mathcal{F}_{t-1}\right],

as required.  \blacksquare

The result that we just proved is a particular case of the following Lemma that we state for reference. The proof follows the same pattern as in the previous paragraphs.

Lemma 6.1

Let (Ω,,P)\left(\Omega,\mathcal{F},P\right) be a probability space and \mathcal{B}\subset\mathcal{F} a sub-σ\sigma-algebra. Let XX and YY be two random variables such that XX is simultaneously independent of YY and \mathcal{B}. For any measurable function Φ:2\Phi:\mathbb{R}^{2}\rightarrow\mathbb{R} define the mapping F:F:\mathbb{R}\rightarrow\mathbb{R} by F(y)=E[Φ(X,y)]F(y)=E\left[\Phi(X,y)\right]. Then,

E[Φ(X,Y)]=E[F(Y)].E\left[\Phi(X,Y)\mid\mathcal{B}\right]=E[F(Y)\mid\mathcal{B}].

6.2 Proof of Theorem 2.6

The proof is a straightforward consequence of Theorem 3.1 in [EM98]. Indeed, it suffices to identify in the ARSV case the two main ingredients necessary to carry out the Extended Girsanov Principle, namely the one period excess discounted return μt\mu_{t} defined by:

eμt:=Et1P[S~tS~t1]=emtrEt1P[eσtϵt]=emtrK,e^{\mu_{t}}:=E^{P}_{t-1}\left[\frac{\widetilde{S}_{t}}{\widetilde{S}_{t-1}}\right]=e^{m_{t}-r}E^{P}_{t-1}\left[e^{\sigma_{t}\epsilon_{t}}\right]=\frac{e^{m_{t}-r}}{K},

with K:=1/Et1P[eσtϵt]K:=1/E^{P}_{t-1}\left[e^{\sigma_{t}\epsilon_{t}}\right], and the process {Wt}\{W_{t}\} uniquely determined by the relation:

Wt:=S~tS~t1eμt=Keσtϵt.W_{t}:=\frac{\widetilde{S}_{t}}{\widetilde{S}_{t-1}}e^{-\mu_{t}}=Ke^{\sigma_{t}\epsilon_{t}}.

The Theorem follows from expression (3.8) in [EM98].  \blacksquare

6.3 Proof of Theorem 2.8

(i) We start by noticing that since ϵt\epsilon_{t} is independent of both ρt\rho_{t} and t1\mathcal{F}_{t-1}, by Lemma 6.1,

Et1P[Nt(ϵt,ρt)]=Et1P[F(ρt)],E^{P}_{t-1}\left[N_{t}(\epsilon_{t},\rho_{t})\right]=E^{P}_{t-1}\left[F(\rho_{t})\right], (6.2)

where the real function F:F:\mathbb{R}\rightarrow\mathbb{R} is defined by

F(y)=EP[Nt(ϵt,y)]=fP(x+y)fP(x)fP(x)𝑑x=fP(x+y)𝑑x=fP(x)𝑑x=1,F(y)=E^{P}\left[N_{t}(\epsilon_{t},y)\right]=\int\frac{f^{P}(x+y)}{f^{P}(x)}f^{P}(x)dx=\int f^{P}(x+y)dx=\int f^{P}(x)dx=1,

which substituted in (6.2) yields

Et1P[Nt(ϵt,ρt)]=1.E^{P}_{t-1}\left[N_{t}(\epsilon_{t},\rho_{t})\right]=1. (6.3)

This equality proves immediately the martingale property for {Zt}\{Z_{t}\}. Indeed,

Et1P[Zt]=Et1P[Nt]Zt1=Zt1.E^{P}_{t-1}\left[Z_{t}\right]=E^{P}_{t-1}\left[N_{t}\right]Z_{t-1}=Z_{t-1}.

Finally, as {Zt}\{Z_{t}\} is a PP-martingale

EP[Zt]=EP[Z1]=EP[N1]=1.E^{P}\left[Z_{t}\right]=E^{P}\left[Z_{1}\right]=E^{P}\left[N_{1}\right]=1.

(ii) ZTZ_{T} is by construction non-negative and E[ZT]=(ZT>0)=1E[Z_{T}]=\mathbb{P}(Z_{T}>0)=1. This guarantees (see, for example, Remarks after Theorem 4.2.1 in [LL08]) that QmcQ_{{\rm mc}} is a probability measure equivalent to PP.

We now start proving that QmcQ_{{\rm mc}} is a martingale measure by showing that:

Et1Qmc1[StSt1]=Et1P[emt+σtϵtNt].E^{Q^{1}_{{\rm mc}}}_{t-1}\left[\frac{S_{t}}{S_{t-1}}\right]=E^{P}_{t-1}\left[e^{m_{t}+\sigma_{t}\epsilon_{t}}N_{t}\right]. (6.4)

Indeed,

Et1Qmc1[StSt1]\displaystyle E^{Q^{1}_{{\rm mc}}}_{t-1}\left[\frac{S_{t}}{S_{t-1}}\right] =\displaystyle= Et1Qmc1[emt+σtϵt]=1Et1P[ZT]Et1P[ZTemt+σtϵt]=1Zt1Et1P[ZTemt+σtϵt]\displaystyle E^{Q^{1}_{{\rm mc}}}_{t-1}\left[e^{m_{t}+\sigma_{t}\epsilon_{t}}\right]=\frac{1}{E^{P}_{t-1}\left[Z_{T}\right]}E^{P}_{t-1}\left[Z_{T}e^{m_{t}+\sigma_{t}\epsilon_{t}}\right]=\frac{1}{Z_{t-1}}E^{P}_{t-1}\left[Z_{T}e^{m_{t}+\sigma_{t}\epsilon_{t}}\right]
=\displaystyle= 1Zt1Et1P[ET1P[ZTemt+σtϵt]]=1Zt1Et1P[ET1P[ZT]emt+σtϵt]\displaystyle\frac{1}{Z_{t-1}}E^{P}_{t-1}\left[E^{P}_{T-1}\left[Z_{T}e^{m_{t}+\sigma_{t}\epsilon_{t}}\right]\right]=\frac{1}{Z_{t-1}}E^{P}_{t-1}\left[E^{P}_{T-1}\left[Z_{T}\right]e^{m_{t}+\sigma_{t}\epsilon_{t}}\right]
=\displaystyle= 1Zt1Et1P[ZT1emt+σtϵt]==1Zt1Et1P[Ztemt+σtϵt]\displaystyle\frac{1}{Z_{t-1}}E^{P}_{t-1}\left[Z_{T-1}e^{m_{t}+\sigma_{t}\epsilon_{t}}\right]=\cdots=\frac{1}{Z_{t-1}}E^{P}_{t-1}\left[Z_{t}e^{m_{t}+\sigma_{t}\epsilon_{t}}\right]
=\displaystyle= 1Zt1Et1P[NtZt1emt+σtϵt]=Et1P[emt+σtϵtNt].\displaystyle\frac{1}{Z_{t-1}}E^{P}_{t-1}\left[N_{t}Z_{t-1}e^{m_{t}+\sigma_{t}\epsilon_{t}}\right]=E^{P}_{t-1}\left[e^{m_{t}+\sigma_{t}\epsilon_{t}}N_{t}\right].

Now, as ϵt\epsilon_{t} is independent of both σt\sigma_{t} and t1\mathcal{F}_{t-1}, we can use Lemma 6.1 to prove that

Et1P[emt+σtϵtNt]=Et1P[emtσtρt+Lϵt(σt)].E^{P}_{t-1}\left[e^{m_{t}+\sigma_{t}\epsilon_{t}}N_{t}\right]=E^{P}_{t-1}\left[e^{m_{t}-\sigma_{t}\rho_{t}+L_{\epsilon_{t}}(\sigma_{t})}\right]. (6.5)

Indeed, we can write

Et1P[emt+σtϵtNt]\displaystyle E^{P}_{t-1}\left[e^{m_{t}+\sigma_{t}\epsilon_{t}}N_{t}\right] =\displaystyle= emtEt1P[eσtϵtNt(ϵt,ρt)]=emtEt1P[F(σt,ρt)],\displaystyle e^{m_{t}}E^{P}_{t-1}\left[e^{\sigma_{t}\epsilon_{t}}N_{t}(\epsilon_{t},\rho_{t})\right]=e^{m_{t}}E^{P}_{t-1}\left[F(\sigma_{t},\rho_{t})\right], (6.6)

where the function F:2F:\mathbb{R}^{2}\rightarrow\mathbb{R} is defined by

F(y,z)\displaystyle F(y,z) =\displaystyle= EP[eyϵtNt(ϵt,z)]=eyxfP(x)fP(x+z)fP(x)𝑑x=ey(sz)fP(s)𝑑s\displaystyle E^{P}\left[e^{y\epsilon_{t}}N_{t}(\epsilon_{t},z)\right]=\int e^{yx}f^{P}(x)\frac{f^{P}(x+z)}{f^{P}(x)}dx=\int e^{y(s-z)}f^{P}(s)ds
=\displaystyle= EP[ey(ϵtz)]=eyzeLϵtP(y),\displaystyle E^{P}\left[e^{y(\epsilon_{t}-z)}\right]=e^{-yz}e^{L_{\epsilon_{t}}^{P}(y)},

which substituted in (6.6) yields (6.5). Finally, if we insert in (6.5) the explicit expression (2.13) that defines the market price of risk, we obtain that

Et1Qmc1[StSt1]=Et1P[emtσtρt+Lϵt(σt)]=erKϵtP(σt)Et1P[eLϵtP(σt)]=er,E^{Q^{1}_{{\rm mc}}}_{t-1}\left[\frac{S_{t}}{S_{t-1}}\right]=E^{P}_{t-1}\left[e^{m_{t}-\sigma_{t}\rho_{t}+L_{\epsilon_{t}}(\sigma_{t})}\right]=e^{r-K_{\epsilon_{t}}^{P}(\sigma_{t})}E^{P}_{t-1}\left[e^{L^{P}_{\epsilon_{t}}(\sigma_{t})}\right]=e^{r},

as required. The last equality in the previous expression follows from Lemma 2.1.  \blacksquare

6.4 Proof of Proposition 3.6

Consider the unique decomposition (3.14) of the discounted payoff H~\widetilde{H} as a sum of the discounted PP-gains and PP-risk processes:

H~=V0+k=1TξkP(S~kS~k1)+L~TP.\widetilde{H}=V_{0}+\sum_{k=1}^{T}\xi_{k}^{P}\cdot\left(\widetilde{S}_{k}-\widetilde{S}_{k-1}\right)+\widetilde{L}_{T}^{P}. (6.7)

We recall that the minimal martingale measure is characterized by the property that every PP-martingale that is strongly orthogonal to the discounted price process {S~t}t=0T\left\{\widetilde{S}_{t}\right\}_{t=0}^{T}, is also a QminQ_{{\rm min}}-martingale. Hence, as the PP-martingale {L~tP}t=0T\left\{\widetilde{L}_{t}^{P}\right\}_{t=0}^{T} is PP-strongly orthogonal to {S~t}t=0T\left\{\widetilde{S}_{t}\right\}_{t=0}^{T}, it is therefore a QminQ_{{\rm min}}-martingale. Having this in mind, as well as (3.13), we take conditional expectations Et1QminE^{Q_{{\rm min}}}_{t-1} on both sides of the decomposition (6.7) and we obtain:

V~tP=V0+k=1tξkP(S~kS~k1)+L~tP.\widetilde{V}_{t}^{P}=V_{0}+\sum_{k=1}^{t}\xi_{k}^{P}\cdot\left(\widetilde{S}_{k}-\widetilde{S}_{k-1}\right)+\widetilde{L}_{t}^{P}.

As V~tP\widetilde{V}_{t}^{P} is the local risk minimizing value process for both the PP and the QminQ_{{\rm min}} measures, that is V~tP=V~tQmin\widetilde{V}_{t}^{P}=\widetilde{V}_{t}^{Q_{{\rm min}}}, if we multiply both sides of by (S~tS~t1)\left(\widetilde{S}_{t}-\widetilde{S}_{t-1}\right) and take conditional expectations Et1QminE^{Q_{{\rm min}}}_{t-1} we have:

Et1Qmin[V~tQmin(S~tS~t1)]\displaystyle E^{Q_{{\rm min}}}_{t-1}\left[\widetilde{V}_{t}^{Q_{{\rm min}}}\left(\widetilde{S}_{t}-\widetilde{S}_{t-1}\right)\right] =\displaystyle= Et1Qmin[V0(S~tS~t1)]+k=1tEt1Qmin[ξkP(S~kS~k1)(S~tS~t1)]\displaystyle E^{Q_{{\rm min}}}_{t-1}\left[V_{0}\left(\widetilde{S}_{t}-\widetilde{S}_{t-1}\right)\right]+\sum_{k=1}^{t}E^{Q_{{\rm min}}}_{t-1}\left[\xi_{k}^{P}\cdot\left(\widetilde{S}_{k}-\widetilde{S}_{k-1}\right)\left(\widetilde{S}_{t}-\widetilde{S}_{t-1}\right)\right]
+Et1Qmin[L~tP(S~tS~t1)]\displaystyle+E^{Q_{{\rm min}}}_{t-1}\left[\widetilde{L}_{t}^{P}\left(\widetilde{S}_{t}-\widetilde{S}_{t-1}\right)\right]
=\displaystyle= ξtPEt1Qmin[(S~tS~t1)2]+Et1Qmin[L~tP(S~tS~t1)].\displaystyle\xi_{t}^{P}E^{Q_{{\rm min}}}_{t-1}\left[\left(\widetilde{S}_{t}-\widetilde{S}_{t-1}\right)^{2}\right]+E^{Q_{{\rm min}}}_{t-1}\left[\widetilde{L}_{t}^{P}\left(\widetilde{S}_{t}-\widetilde{S}_{t-1}\right)\right].

If we divide both sides of this equality by Et1Qmin[(S~tS~t1)2]E^{Q_{{\rm min}}}_{t-1}\left[\left(\widetilde{S}_{t}-\widetilde{S}_{t-1}\right)^{2}\right] and we use the relation (3.3) we obtain that

ξtQmin=ξtP+Et1Qmin[L~tP(S~tS~t1)]vart1Qmin[S~tS~t1],\xi_{t}^{Q_{{\rm min}}}=\xi_{t}^{P}+\frac{E^{Q_{{\rm min}}}_{t-1}\left[\widetilde{L}_{t}^{P}(\widetilde{S}_{t}-\widetilde{S}_{t-1})\right]}{{\rm var}_{t-1}^{Q_{{\rm min}}}\left[\widetilde{S}_{t}-\widetilde{S}_{t-1}\right]},

as required. Finally, we notice that

Et1Qmin[(L~tPL~t1P)(S~tS~t1)]=Et1Qmin[L~tP(S~tS~t1)]Et1Qmin[L~t1P(S~tS~t1)]=Et1Qmin[L~tP(S~tS~t1)],E^{Q_{{\rm min}}}_{t-1}\left[\left(\widetilde{L}_{t}^{P}-\widetilde{L}_{t-1}^{P}\right)(\widetilde{S}_{t}-\widetilde{S}_{t-1})\right]=E^{Q_{{\rm min}}}_{t-1}\left[\widetilde{L}_{t}^{P}(\widetilde{S}_{t}-\widetilde{S}_{t-1})\right]-E^{Q_{{\rm min}}}_{t-1}\left[\widetilde{L}_{t-1}^{P}(\widetilde{S}_{t}-\widetilde{S}_{t-1})\right]=E^{Q_{{\rm min}}}_{t-1}\left[\widetilde{L}_{t}^{P}(\widetilde{S}_{t}-\widetilde{S}_{t-1})\right],

hence, either the QminQ_{{\rm min}}-independence or the orthogonality of {L~tP}t=0T\left\{\widetilde{L}_{t}^{P}\right\}_{t=0}^{T} and {S~t}t=0T\left\{\widetilde{S}_{t}\right\}_{t=0}^{T} imply that Et1Qmin[L~tP(S~tS~t1)]=0E^{Q_{{\rm min}}}_{t-1}\left[\widetilde{L}_{t}^{P}(\widetilde{S}_{t}-\widetilde{S}_{t-1})\right]=0 and hence {ξtQmin}t=1T={ξtP}t=1T\{\xi_{t}^{Q_{{\rm min}}}\}_{t=1}^{T}=\{\xi_{t}^{P}\}_{t=1}^{T}.  \blacksquare

References

  • [AS96] T. G. Andersen and B. E. Sørensen. GMM estimation of a stochastic volatility model: A Monte Carlo study. Journal of Business and Economic Statistics, 14:328–352, 1996.
  • [BEK+11] Alex Badescu, Robert J. Elliott, Reg Kulperger, Jarkko Miettinen, and Tak Kuen Siu. A comparison of pricing kernels for GARCH option pricing with generalized hyperbolic distributions. IJTAF, 14(5):669–708, 2011.
  • [BO11] Alex Badescu and Juan-Pablo Ortega. Hedging garch options with generalized innovations. In preparation, 14(5):669–708, 2011.
  • [Bol86] Tim Bollerslev. Generalized autoregressive conditional heteroskedasticity. J. Econometrics, 31(3):307–327, 1986.
  • [BS72] F. Black and M. Scholes. The valuation of option contracts and a test of market efficiency. J. Finance, 27:399–417, 1972.
  • [CPnR04] M. Angeles Carnero, D. Peña, and E. Ruiz. Persistence and kurtosis in GARCH and stochastic volatility models. Journal of Financial Econometrics, 2(2):319–342, 2004.
  • [Dan94] J. Daníelsson. Stochastic volatility in asset prices: estimation with simulated maximum likelihood. Journal of Econometrics, 64:375–400, 1994.
  • [dCL08] Joan del Castillo and Youngjo Lee. GLM-methods for volatility models. Stat. Model., 8(3):263–283, 2008.
  • [DGS01] Jin-Chuan Duan, Geneviève Gauthier, and Jean-Guy Simonato. Asymptotic distribution of the ems option price estimator. Mgmt. Sci., 47(8):1122–1132, 2001.
  • [DS98] Jin-Chuan Duan and Jean-Guy Simonato. Empirical martingale simulation of asset prices. Mgmt. Sci., 44(9):1218–1233, 1998.
  • [Dua95] Jin-Chuan Duan. The GARCH option pricing model. Math. Finance, 5(1):13–32, 1995.
  • [EM98] Robert J. Elliott and Dilip B. Madan. A discrete time equivalent martingale measure. Math. Finance, 8(2):127–152, 1998.
  • [Eng82] Robert F. Engle. Autoregressive conditional heteroscedasticity with estimates of the variance of United Kingdom inflation. Econometrica, 50(4):987–1007, 1982.
  • [FS86] Hans Föllmer and Dieter Sondermann. Hedging of nonredundant contingent claims. In Contributions to mathematical economics, pages 205–223. North-Holland, Amsterdam, 1986.
  • [FS91] Hans Föllmer and Martin Schweizer. Hedging of contingent claims under incomplete information. In Applied stochastic analysis (London, 1989), volume 5 of Stochastics Monogr., pages 389–414. Gordon and Breach, New York, 1991.
  • [FS02] Hans Föllmer and Alexander Schied. Stochastic finance, volume 27 of de Gruyter Studies in Mathematics. Walter de Gruyter & Co., Berlin, 2002. An introduction in discrete time.
  • [GCI10] D. Guegan, C. Chorro, and F. Ielpo. Option pricing for garch-type models with generalized hyperbolic innovations. Preprint, 2010.
  • [GLD04] Andrew G. Glen, Lawrence M. Leemis, and John H. Drew. Computing the distribution of the product of two continuous random variables. Comput. Statist. Data Anal., 44(3):451–464, 2004.
  • [GR99] R. Garcia and E. Renault. Note on hedging in ARCH-type option pricing models. Math. Finance, 8(8):153–161, 1999.
  • [HN00] S. L. Heston and S. Nandi. A closed-form garch option valuation model. The Review of Financial Studies, 13(3):585–625, 2000.
  • [HRS94] A. C. Harvey, E. Ruiz, and N. Shephard. Multivariate stochastic variance models. Review of Economic Studies, 61:247–264, 1994.
  • [JPR94] E. Jacquier, N.G. Polson, and P.E. Rossi. Bayesian analysis of stochastic volatility models. Journal of Business and Economic Statistics, 12:371–417, 1994.
  • [LL08] Damien Lamberton and Bernard Lapeyre. Introduction to stochastic calculus applied to finance. Chapman & Hall/CRC Financial Mathematics Series. Chapman & Hall/CRC, Boca Raton, FL, second edition, 2008.
  • [LN96] Y. Lee and J. A. Nelder. Hierarchical generalized linear models. J. Roy. Statist. Soc. Ser. B, 58(4):619–678, 1996. With discussion.
  • [LN06] Youngjo Lee and John A. Nelder. Double hierarchical generalized linear models. J. Roy. Statist. Soc. Ser. C, 55(2):139–185, 2006.
  • [Mer76] R. Merton. Option pricing when underlying stock returns are discontinuous. J. Finan. Econ., 3:125–144, 1976.
  • [MT90] A. Melino and S. M. Turnbull. Pricing foreign currency options with stochastic volatility. Journal of Econometrics, 45:239–265, 1990.
  • [Ort10] J.-P. Ortega. GARCH options via local risk minimization. Quantitative Finance, 2010.
  • [Sch01] M. Schweizer. A guided tour through quadratic hedging approaches. In Option pricing, interest rates and risk management, Handbook of Mathematical Finance, pages 538–574. Cambridge University Press, Cambridge, 2001.
  • [Tay86] Stephen J. Taylor. Modelling Financial Time Series. John Wiley & Sons, Chichester, 1986.
  • [Tay05] Stephen J. Taylor. Asset Price Dynamics, Volatility, and Prediction. Princeton University Press, Princeton, 2005.
  • [WLLdC11] Johan Lim Woojoo Lee, Youngjo Lee, and Joan del Castillo. The hierarchical-likelihood approach to autoregressive stochastic volatility models. Computational Statistics and Data Analysis, 55(55):248–260, 2011.