This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.


A Class of Explicit optimal contracts in the face of shutdown

Jessica Martin111jessica.martin@insa-toulouse.fr INSA de Toulouse
IMT UMR CNRS 5219
Université de Toulouse
135 Avenue de Rangueil 31077 Toulouse Cedex 4 France
Stéphane Villeneuve222stephane.villeneuve@tse-fr.eu
The authors acknowledge funding from ANR PACMAN and from ANR under grant ANR-17-EUR-0010 (Investissements d’Avenir program). The second author gratefully thanks the FdR-SCOR “Chaire Marché des risques et création de valeurs”.
Université Toulouse 1 Capitole, (TSE-TSMR)
2 Rue du Doyen Gabriel Marty
31000 Toulouse
Abstract

What type of delegation contract should be offered when facing a risk of the magnitude of the pandemic we are currently experiencing and how does the likelihood of an exogenous early termination of the relationship modify the terms of a full-commitment contract? We study these questions by considering a dynamic principal-agent model that naturally extends the classical Holmström-Milgrom setting to include a risk of default whose origin is independent of the inherent agency problem. We obtain an explicit characterization of the optimal wage along with the optimal action provided by the agent. The optimal contract is linear by offering both a fixed share of the output which is similar to the standard shutdown-free Holmström-Milgrom model and a linear prevention mechanism that is proportional to the random lifetime of the contract. We then tweak the model to add a possibility for risk mitigation through investment and study its optimality.

Keywords:  Principal-Agent problems, default risk, Hamilton-Jacobi Bellman equations.

1  Introduction

Without seeking to oppose public health and economic growth, there is no doubt that the management of the Covid crisis had serious consequences on entire sectors of the economy. The first few months of 2020 will go down in world history as a period of time characterized by massive layoffs, forced closures of non-essential companies, disruption of cross-border transportation whilst populations were subject to lockdown and/or social distancing measures and hospitals and the medical world struggled to get a grasp on the Sars-Cov-2 pandemic. Whilst the immediate priority was saving lives, decongesting hospitals and preventing the spread of the disease, many extraordinary economic support measures were taken to help businesses and individuals stay afloat during these unprecedented times and in the hope of tempering the economic crisis that would follow. Although the world has lived through many crises over the past centuries, from several Panics in the 1800s and the Great Depression of the 1930s to the more recent Financial Crisis of 2008, never before has the global economy as a whole come to such a standstill due to an external event. Such large shutdown risks do not only materialize during pandemics but throughout many major other large events. The massive bushfires that affected Australia towards the end of 2019, temporarily halting agriculture, construction activity and tourism in some areas of the country are another recent example. As we begin to see a glimpse of hope for a way out through vaccination, the focus is turning to building the world of tomorrow with the idea that we must learn to live with such risks. This paper tries to make its contribution by focusing on a simple microeconomic issue. In a world subject to moral hazard, how can we agree to an incentive contract whose obligations could be made impossible or at least very difficult because of the occurrence of a risk of the nature of the Covid19 pandemic? Including such a shutdown risk-sharing in contracts seems crucial going forward for at least two reasons. First, it is not certain that public authority will be able to continue to take significant economic support measures to insure the partners of a contract if the frequency of such global risks were to increase. On the other hand, the private insurance market does not offer protection against the risk of a pandemic which makes pooling too difficult. It therefore seems likely that we will have to turn to an organized form of risk sharing between the contractors.

Economic theory has a well-developed set of tools to analyze incentive and risk-sharing problems using expected-utility theory. Most of the now abundant literature related to dynamic contracting through a Principal-Agent model has, so far, mostly been based on continuously governed (eg. Brownian motion) output-processes. This was the case of the foundational work of Holmström and Milgrom [12] and many of its many extensions such as those of Schattler and Sung in [20] and [21], and the more recent contributions of Sannikov in [19], and Cvitanic et al [5] and [6]. However some relatively recent works have introduced jump processes into continuous time contracting. Biais et al. were the first to do so in [3] where they study optimal contracting between an insurance company and a manager whose effort can reduce the occurence of an underlying accident. In a similar vein, work by Capponi and Frei [4] also used a jump-diffusion driven outcome process in order to include the possibility for accidents to negatively affect revenue. Here, we extend the classical Holmström and Milgrom [12] framework to include a shutdown risk. We do not claim that this model with CARA preferences is general enough to come up with robust economic facts, but it has the remarkable advantage of being explicitly creditworthy, which allows us to find an explicit optimal contract that disentangles the incentives from external risk-sharing and allows us to understand the sensitivity of the optimal contract to the different exogenous parameters of the model. Our work uses a jump-diffusion process too and presents some structural similarities with [4] : both models consider a risk averse principal and agent with exponential utility and reach an explicit characterization of the optimal wage. However, Capponi and Frei combine the continuous part of the diffusion and the accident jump process additively and they are able to allow prevention through intensity control of the jump process. This makes sense as many accidents are preventable through known measures. Our framework uses a different form of jump diffusion to enable the shutdown event to completely stop revenue generation in a continuous-time setting. This is done by building on a standard continuous Brownian-motion based output process. The main novelty is the multiplicative effect of the jump risk : upon the arrival of the risk, the whole of the output process comes to a halt. As an extension, we allow the halt to no longer be a complete fatality : production may continue at a degraded level through an investment by the principal. From a methodological viewpoint, our reasoning uses a now standard method in dynamic contracting based on [19] and [6] which consists in transforming both the first-best and second-best problems into classical Markovian control problems. The solution to these control problems can be characterized through a Hamilton-Jacobi-Bellman equation. Quite remarkably, this equation has, in our context, an explicit solution that is closely linked to a so-called Bernouilli ODE which facilitates many extensions.

To the best of our knowledge, this paper is the first to explicitly introduce a default in a dynamic Principal-agent framework, in both a first-best (also called full Risk-Sharing) and second-best (also called Moral Hazard) setting. A key feature of our study is that the shape of the optimal contract is linear. More precisely, the agent’s compensation is the sum of two functions: the first is linear with respect to the output and serves to give the incentives, while the second is linear with respect to the effective duration of the contract and serves to share the default risk. While the linear incentive part of the contract is in line with the existing literature on continuous-time Principal-Agent problems without default under exponential utilities, the risk-sharing part deserves some clarification. The contract exposes both agents to a risk of exogenous interruption but it has two different regimes that are determined by an explicit relation between the risk-aversions and the agent’s effort cost. Under the first regime, the agent is more sensitive to the risk of default than the principal. In this case, the principal deposits on the date 0 a positive amount onto an escrow account whose balance will then decrease over time at a constant rate. It is crucial to observe that the later the default arrives, the more the amount in the escrow account decreases to a point where it may even become negative. If the default occurs, the principal transfers the remaining balance to the agent. Under the second regime, the principal is more sensitive to the risk of default. In this case, the principal deposits a negative amount into the escrow account, which now grows at a constant rate and symmetrical reasoning applies. This linearity contrasts with the optimum obtained in [4] as the additive contribution of their jump process to revenue generation leads to a sub-linear wage. This result is coherent with the paper by Hoffman and Pfeil [11] which proves that, in line with the empirical studies by Bertrand and Mullainathan [2], the agent must be rewarded or punished for a risk that is beyond his control.
Finally, this paper also explicitly characterizes the optimal contract when a possibility for shutdown risk mitigation exists at a cost. Such a possibility is coherent with agency-free external risk: prevention is not possible, at least on a short-term or medium-term time scale. At best the principal may be able to invest to mitigate its effects. Crucially we find that in many circumstances, investing is not optimal for the principal. When it is, it is only optimal up until some cutoff time related to a balance between the cost of investment, the agent’s rents and possible remaining gain.

The rest of the document is structured as follows. In Section 2, we present the model and the Principal-Agent problems that we consider. In Section 3, we analyse the first-best case where the principal observes the agent’s effort. Then in Section 4, we give our main results and analysis. In Section 5, we extend our model to include a possibility for mitigation upon a halt.

2  The Model

The model is inherited from the classical work of Holmström and Milgrom [12]. A principal contracts with an agent to manage a project she owns. The agent influences the project’s profitability by exerting an unobservable effort. For a fixed effort policy, the output process is still random and the idiosyncratic uncertainty is modeled by a Brownian motion.
We assume that the contract matures at time T>0T>0 and both principal and agent are risk-averse with CARA preferences. The departure from the classical model is as follows: we assume the project is facing some external risk that could partially or totally interrupt the production at some random time τ\tau. The probability distribution of τ\tau is assumed to be independent of the Brownian motion that drives the uncertainty of the output process and also independent of the agent’s actions. Finally, we assume that the contract offers a transfer WW at time TT from the principal to the agent that is a functional of the output process.

Probability setup

Let T>0T>0 be some fixed time horizon. The key to modeling our Principal-Agent problems under an agency-free external risk of default is the simultaneous presence over the interval [0,T][0,T] of a continuous random process and a jump process as well as the ability to extend the standard mathematical techniques used for dynamic contracting to this mixed setting. Thus, we shall deal with two kinds of information : the information from the output process, denoted as 𝔽=(t)t0\mathbb{F}=({\cal F}_{t})_{t\geq 0} and the information from the default time, i.e. the knowledge of the time where the default occurred in the past, if the default has appeared. This construction is not new and occurs frequently in mathematical finance333We refer the curious reader to the two important references [1] and [9]..

The complete probability space that we consider will be denoted as (Ω,𝒢,0)(\Omega,\mathcal{G},\mathbb{P}^{0}), with two independent stochastic processes :

  • BB a standard one-dimensional 𝔽\mathbb{F}-Brownian motion,

  • NN the right-continuous single-jump process defined as Nt=1τtN_{t}=\textbf{1}_{\tau\leq t}, tt in [0,T][0,T] where τ\tau is some positive random variable independent of BB that models the default time.

NN will also be referred to as the default indicator process. We therefore use the standard approach of progressive enlargement of filtration by considering 𝔾={𝒢t,t0}\mathbb{G}=\left\{\mathcal{G}_{t},t\geq 0\right\} the smallest complete right-continuous extension of 𝔽\mathbb{F} that makes τ\tau a 𝔾\mathbb{G}-stopping time. Because τ\tau is independent of BB, BB is a 𝔾\mathbb{G}-Brownian motion under 0\mathbb{P}^{0} according to Proposition 1.21 p 11 in [1]. We also suppose that there exists a bounded deterministic compensator of NN, Λt=0tλ(s)𝑑s\Lambda_{t}=\int_{0}^{t}\lambda(s)\,ds for some bounded function λ(.)\lambda(.) called the intensity implying that:

Mt=Nt0tλ(s)(1Ns)𝑑s,t[0,T]M_{t}=N_{t}-\int_{0}^{t}\lambda(s)(1-N_{s})ds,\quad t\in[0,T]

is a 𝔾\mathbb{G}-compensated martingale. Note that through knowledge of the function λ,\lambda, the principal and agent can compute at time 0 the probability of default happening over the contracting period [0,T][0,T]. Indeed :

(τT)=1exp(ΛT).\mathbb{P}(\tau\leq T)=1-\exp(-\Lambda_{T}).

We first suppose for computational ease that the intensity λ\lambda is a constant. We will see in Section 4.3 that our results may easily be lifted to more general deterministic compensators.

Remark 2.1.

Here we will suppose that the compensator of NN is common knowledge to both the Principal and the Agent. We could imagine settings where the Principal and Agent’s beliefs regarding the risk of default may differ : this natural extension of our work would call for analysis of the dynamic contracting problem under hidden information which is left for future research.

Principal-Agent Problem

We suppose that the agent agrees to work for the principal over a time period [0,T][0,T] and provide up to the default time a costly action (at)t[0,T](a_{t})_{t\in[0,T]} belonging to 𝒜\mathcal{A}, where 𝒜\mathcal{A} denotes the set of admissible 𝔽\mathbb{F}-predictable strategies that will be specified later on. The Principal-Agent problem models the realistic setting where the principal cannot observe the agent’s effort. As such the agent chooses his action in order to maximize his own utility. The principal must offer a wage based on the information driven by the output process up to the default time that incentivizes the agent to work efficiently and contribute positively to the output process. Mathematically, the unobservability of the agent’s behaviour is modeled through a change of measure. Under 0\mathbb{P}^{0} , we assume that the project’s profitability evolves as

Xt:=x0+0t(1Ns)𝑑Bs.X_{t}:=x_{0}+\int_{0}^{t}(1-N_{s})dB_{s}.

Thus, 0\mathbb{P}^{0} corresponds to the probability distribution of the profitability when the agent makes no effort over [0,T][0,T]. When the agent makes an effort a=(at)ta=(a_{t})_{t}, we shall assume that the project’s profitability evolves as

Xt:=x0+0tas(1Ns)𝑑s+0t(1Ns)𝑑Bsa,X_{t}:=x_{0}+\int_{0}^{t}a_{s}(1-N_{s})ds+\int_{0}^{t}(1-N_{s})dB^{a}_{s},

where BaB^{a} is a 𝔽\mathbb{F}-Brownian motion under a measure a\mathbb{P}^{a}. The agent fully observes the decomposition of the production process under a measure a\mathbb{P}^{a} whilst the principal only observes the realization of XtX_{t}. In order for the model to be consistent, the probabilities 0\mathbb{P}^{0} and a\mathbb{P}^{a} must be equivalent for all (at)t[0,Tτ](a_{t})_{t\in[0,T\wedge\tau]} belonging to 𝒜\mathcal{A}. Therefore, we introduce the following set of actions

={a=(at)t:𝔽-predictable and taking values in [A,A] for some A>0}.{\cal B}=\Big{\{}a=(a_{t})_{t}:\quad\mathbb{F}\hbox{-predictable and taking values in }[-A,A]\text{ for some }A>0\Big{\}}.

The action process in {\cal B} are uniformly bounded by some fixed constant A>0A>0 that will be assumed as large as necessary. For aa\in{\cal B}, we define a\mathbb{P}^{a} as

dad0|𝒢T=exp(0Tas(1Ns)𝑑Bs120T|as|2(1Ns)𝑑s):=LT.\frac{d\mathbb{P}^{a}}{d\mathbb{P}^{0}}|\mathcal{G}_{T}=\exp\left(\int_{0}^{T}a_{s}(1-N_{s})dB_{s}-\frac{1}{2}\int_{0}^{T}|{a_{s}}|^{2}(1-N_{s})ds\right):=L_{T}.

Because 𝔼0(LT)=1\mathbb{E}^{0}(L_{T})=1 , (Bta)t[0,T](B^{a}_{t})_{t\in[0,T]} with Bta=Bt0tas(1Ns)𝑑s,t[0,T]B^{a}_{t}=B_{t}-\int_{0}^{t}a_{s}(1-N_{s})ds,t\in[0,T] is a 𝔾\mathbb{G}-Brownian motion under a\mathbb{P}^{a} according to Proposition 3.6 c) p 55 in [1]. It is key to note that if halt occurs, i.e. if τT\tau\leq T, then the production process is halted before TT meaning that : Xtτa=Xta,t[0,T]X^{a}_{t\wedge\tau}=X_{t}^{a},\quad t\in[0,T]. Let us then observe that an action a=(at)ta=(a_{t})_{t} of {\cal B} can be extended to a 𝔾\mathbb{G}-predictable process (a~t)t[0,T](\tilde{a}_{t})_{t\in[0,T]} by setting a~t=at11tτ\tilde{a}_{t}=a_{t}1\!\!1_{t\leq\tau}.

The cost of effort for the agent is modeled through a quadratic cost function : κ(a):=κa22,\kappa(a):=\kappa\frac{a^{2}}{2}, for κ>0\kappa>0 some fixed parameter. As a reward for the agent’s effort, the principal pays him a wage WW at time TT. WW is assumed to be a 𝒢Tτ\mathcal{G}_{T\wedge\tau} random variable which means that the payment at time TT in case of an early default is known at time τ\tau. The principal and the agent are considered to be risk averse and risk aversion is modeled through two CARA utility functions :

UP(x):=exp(γPx)andUA(x):=exp(γAx),U_{P}(x):=-\exp(-\gamma_{P}x)\;\text{and}\;U_{A}(x):=-\exp(-\gamma_{A}x),

where γP>0\gamma_{P}>0 and γA>0\gamma_{A}>0 are two fixed constants modeling the principal’s and the agent’s risk aversion.

In this setting and for any given wage WW, the agent maximizes his own utility and solves :

V0A(W)=supa𝔼a[UA(W0Tκ(as(1Ns))𝑑s)].V_{0}^{A}(W)=\sup_{a\in\mathcal{B}}\mathbb{E}^{a}\left[U_{A}\left(W-\int_{0}^{T}\kappa(a_{s}(1-N_{s}))ds\right)\right]. (2.1)

A wage WW is said to be incentive compatible if there exists an action policy a(W)a^{*}(W)\in{\cal B} that maximises (2.1) and thus satisfies

V0A(W)=𝔼a(W)[UA(W0Tκ(as(W)(1Ns))𝑑s)].V_{0}^{A}(W)=\mathbb{E}^{a^{*}(W)}\left[U_{A}\left(W-\int_{0}^{T}\kappa(a_{s}^{*}(W)(1-N_{s}))ds\right)\right].

When the principal is able to offer an incentive compatible wage WW, she knows what the agent’s best reply will be. As such the principal establishes a set 𝒜(W)\mathcal{A}^{*}(W)\subset\mathcal{B} of best replies for the agent for any incentive compatible WW. Therefore, the first task is to characterize the set of incentive-compatible wages 𝒲IC{\cal W}_{IC}. Only then may the principal consider maximizing his own utility by solving :

supW𝒲ICsupa𝒜(W)𝔼a(W)[UP(XTa(W)W)]\sup_{W\in\mathcal{W}_{IC}}\;\sup_{a^{*}\in\mathcal{A}^{*}(W)}\;\mathbb{E}^{{a^{*}}(W)}\left[U_{P}\left(X^{a^{*}{(W)}}_{T}-W\right)\right] (2.2)

under the participation constraint

𝔼a(W)[UA(W0Tκ(as(W)(1Ns))𝑑s)]UA(yPC),\mathbb{E}^{{a^{*}}(W)}\left[U_{A}\left(W-\int_{0}^{T}\kappa(a_{s}^{*}(W)(1-N_{s}))ds\right)\right]\geq U_{A}(y_{PC}), (2.3)

where yPCy_{PC} is a monetary reservation utility for the agent.

Remark 2.2.

Problem (2.2) has been thoroughly analyzed in a setting where the output process may not default (see the pioneer papers [12], [20] ). Setting κ=1\kappa=1 for simplicity, the optimal action is constant and given by :

a=γP+1γP+γA+1,a^{*}=\frac{\gamma_{P}+1}{\gamma_{P}+\gamma_{A}+1},

and the optimal wage is linear in the output:

W=yPC+aXT+(γA12(a)2)T.W=y_{PC}+a^{*}X_{T}+\left(\frac{\gamma_{A}-1}{2}({a^{*}})^{2}\right)T.

We may naturally expect to encounter an extension of these results in this setting.

3  Optimal First-best contracting

We begin with analysis of the first-best benchmark (the full Risk-Sharing problem) which leads to a simple optimal sharing rule. Of course this problem is not the most realistic when it comes to modeling dynamic contracting situations. However it provides a benchmark to which we can compare the more realistic Moral Hazard situation. Indeed, the principal’s utility in the full Risk-Sharing problem is the best that the principal will ever be able to obtain in a contracting situation as he may observe (and it is thus assumed that he may dictate) the agent’s action.

To write the first-best problem, we assume that both the principal and the agent observe the variations of the same production process (Xta)t[0,T](X_{t}^{a})_{t\in[0,T]} under 0\mathbb{P}^{0}:

Xta:=x0+0tas(1Ns)ds+0t(1Ns)dBs.t[0,T]X^{a}_{t}:=x_{0}+\int_{0}^{t}a_{s}(1-N_{s})ds+\int_{0}^{t}(1-N_{s})dB_{s}.\quad t\in[0,T] (3.1)

The agent is guaranteed a minimum value of expected utility through the participation constraint :

𝔼[UA(W0Tκ(as(1Ns))𝑑s)]UA(yPC),\mathbb{E}\left[U_{A}\left(W-\int_{0}^{T}\kappa(a_{s}(1-N_{s}))ds\right)\right]\geq U_{A}(y_{PC}),\\ (3.2)

but has no further say on the wage or action. Consider the admissible set :

𝒜PC:={(W,a) such that W is 𝒢Tτ measurable with 𝔼[exp(2γAW)]<+,(at)t, and (3.2)is satisfied}.\mathcal{A}_{PC}:=\left\{(W,a)\hbox{ such that }W\text{ is }\mathcal{G}_{T\wedge\tau}\text{ measurable with }\mathbb{E}\left[\exp(-2\gamma_{A}W)\right]<+\infty,(a_{t})_{t}\in{\cal B},\,\text{ and }(\ref{eq:PCFB})\;\text{is satisfied}\right\}.

The full Risk-Sharing problem involves maximizing the principal’s utility across 𝒜PC\mathcal{A}_{PC} :

sup(W,a)𝒜PC𝔼[UP(XTaW)].\sup_{(W,a)\in\mathcal{A}_{PC}}\quad\mathbb{E}\left[U_{P}\left(X^{a}_{T}-W\right)\right]. (3.3)

Tackling the Participation Constraint

A first step to optimal contracting in this first-best setting involves answering the following question: can we characterize the set 𝒜PC{\cal A}_{PC}? Following the standard route, we will first establish a necessary condition. For a given pair (W,a)𝒜PC(W,a)\in{\cal A}_{PC}, let us introduce the agent’s continuation utility (Ut(W,a))t(U^{(W,a)}_{t})_{t} as follows:

Ut(W,a):=𝔼t[UA(WtTκ(as(1Ns))𝑑s)],U^{(W,a)}_{t}:=\mathbb{E}_{t}\left[U_{A}\left(W-\int_{t}^{T}\kappa(a_{s}(1-N_{s}))ds\right)\right],

where we use the shorthand notation : 𝔼t[.]:=𝔼[.|𝒢t].\mathbb{E}_{t}[.]:=\mathbb{E}[.|\mathcal{G}_{t}]. We may write the Agent’s continuation value process as the product :

Ut(W,a)=t(W,a)𝒟t(W,a),U^{(W,a)}_{t}=\mathcal{M}_{t}^{(W,a)}\mathcal{D}_{t}^{(W,a)},

where :

t(W,a):=𝔼t[UA(W0Tκ(as(1Ns))𝑑s)]and𝒟t(W,a):=exp(γA0tκ(as(1Ns))𝑑s).\mathcal{M}_{t}^{(W,a)}:=\mathbb{E}_{t}\left[U_{A}\left(W-\int_{0}^{T}\kappa(a_{s}(1-N_{s}))ds\right)\right]\quad\text{and}\quad\mathcal{D}_{t}^{(W,a)}:=\exp\left(-\gamma_{A}\int_{0}^{t}\kappa(a_{s}(1-N_{s}))ds\right).

Observe that for any admissible pair (W,a)𝒜PC(W,a)\in{\cal A}_{PC}, the process =(t(W,a))t\mathcal{M}=(\mathcal{M}_{t}^{(W,a)})_{t} is a 𝔾\mathbb{G}-square integrable martingale. According to the Martingale Representation Theorem for 𝔾\mathbb{G}-martingales (see [1], Theorem 3.12 p. 60), there exists some predictable pair (zs,ls)(z_{s},l_{s}) in 2×2\mathbb{H}^{2}\times\mathbb{H}^{2}, where 2\mathbb{H}^{2} is the set of 𝔽\mathbb{F}-predictable processes ZZ with 𝔼[0T|Zt|2𝑑t]<+\mathbb{E}\left[\int_{0}^{T}|Z_{t}|^{2}dt\right]<+\infty, such that :

t(W,a):=0(W,a)+0tzs(1Ns)𝑑Bs+0tls(1Ns)𝑑Ms.\mathcal{M}_{t}^{(W,a)}:=\mathcal{M}_{0}^{(W,a)}+\int_{0}^{t}z_{s}(1-N_{s})dB_{s}+\int_{0}^{t}l_{s}(1-N_{s})dM_{s}.

Integration by parts yields the dynamic of UU, noting that 𝒟\mathcal{D} has finite variation :

dUt(W,a)\displaystyle dU_{t}^{(W,a)} =γAκ(at(1Ns))Ut(W,a)dt+𝒟t(W,a)zt(1Ns)dBt+𝒟t(W,a)lt(1Ns)dMt.\displaystyle=-\gamma_{A}\kappa(a_{t}(1-N_{s}))U_{t}^{(W,a)}dt+\mathcal{D}_{t}^{(W,a)}z_{t}(1-N_{s})dB_{t}+\mathcal{D}_{t}^{(W,a)}l_{t}(1-N_{s})dM_{t}.

Setting Zt(W,a):=𝒟t(W,a)zt2Z_{t}^{(W,a)}:=\mathcal{D}_{t}^{(W,a)}z_{t}\in\mathbb{H}^{2} and Kt(W,a):=𝒟t(W,a)lt2K_{t}^{(W,a)}:=\mathcal{D}_{t}^{(W,a)}l_{t}\in\mathbb{H}^{2}, we obtain:

dUt(W,a)\displaystyle dU_{t}^{(W,a)} =γAκ(at(1Ns))Ut(W,a)dt+Zt(W,a)(1Ns)dBt+Kt(W,a)(1Ns)dMt.\displaystyle=-\gamma_{A}\kappa(a_{t}(1-N_{s}))U_{t}^{(W,a)}dt+Z_{t}^{(W,a)}(1-N_{s})dB_{t}+K_{t}^{(W,a)}(1-N_{s})dM_{t}.

By construction, we have that UT(W,a)=UA(W)U_{T}^{(W,a)}=U_{A}(W). It follows that (Ut(W,a),Zt(W,a),Kt(W,a))\left(U_{t}^{(W,a)},Z_{t}^{(W,a)},K_{t}^{(W,a)}\right) is a solution to the BSDE:

dUt(W,a)=Zt(W,a)(1Ns)dBtKt(W,a)(1Ns)dMt+γAκ(at(1Ns))Ut(W,a)dt,\displaystyle-dU_{t}^{(W,a)}=-Z_{t}^{(W,a)}(1-N_{s})dB_{t}-K_{t}^{(W,a)}(1-N_{s})dM_{t}+\gamma_{A}\kappa(a_{t}(1-N_{s}))U_{t}^{(W,a)}dt, (3.4)

with UT(W,a)=UA(W).U_{T}^{(W,a)}=U_{A}(W). Therefore, (3.2) is satisfied if and only if U0(W,a)UA(yPC)U_{0}^{(W,a)}\geq U_{A}(y_{PC}).

Remark 3.1.

Let 𝕊2\mathbb{S}^{2} be the set of 𝔾\mathbb{G}-adapted RCLL processes UU such that

𝔼[sup0tT|Ut|2]<+.\mathbb{E}[\sup_{0\leq t\leq T}|U_{t}|^{2}]<+\infty.

Through Proposition 2.6 of [7], the solution to (3.4) is unique in (𝕊2×2×2)(\mathbb{S}^{2}\times\mathbb{H}^{2}\times\mathbb{H}^{2}). Indeed, the driver g(ω,U)=γAκ(at(1Nt))Ug(\omega,U)=\gamma_{A}\kappa(a_{t}(1-N_{t}))U is uniformly Lipschitz in UU because (at)t(a_{t})_{t} is bounded and the terminal condition is in L2L^{2}.

To sum up, we have the following necessary condition for admissibility.

Lemma 3.1.

If (W,a)𝒜PC(W,a)\in{\cal A}_{PC} then there exists a unique solution (Ut(W,a),Zt(W,a),Kt(W,a))\left(U_{t}^{(W,a)},Z_{t}^{(W,a)},K_{t}^{(W,a)}\right) in (𝕊2×2×2)(\mathbb{S}^{2}\times\mathbb{H}^{2}\times\mathbb{H}^{2}) to the BSDE (3.4) such that U0(W,a)UA(yPC)U_{0}^{(W,a)}\geq U_{A}(y_{PC}).

To obtain a sufficient condition, we introduce, for π=(y0,a,β,H)××2×2\pi=(y_{0},a,\beta,H)\in\mathbb{R}\times{\cal B}\times\mathbb{H}^{2}\times\mathbb{H}^{2}, the wage process (Wtπ)t(W_{t}^{\pi})_{t} defined as

Wtπ\displaystyle W_{t}^{\pi} :=y0+0tβs(1Ns)dBs+0tHs(1Ns)dMs+0t{γA2βs2(1Ns)+κ(as(1Ns))\displaystyle:=y_{0}+\int_{0}^{t}\beta_{s}(1-N_{s})dB_{s}+\int_{0}^{t}H_{s}(1-N_{s})dM_{s}+\int_{0}^{t}\left\{\frac{\gamma_{A}}{2}\beta_{s}^{2}(1-N_{s})+\kappa(a_{s}(1-N_{s}))\right.
λγA[exp(γAHs)1+γAHs](1Ns)}ds,\displaystyle\left.\frac{\lambda}{\gamma_{A}}[\exp(-\gamma_{A}H_{s})-1+\gamma_{A}H_{s}](1-N_{s})\right\}ds, (3.5)

and consider the set

Γ:={(y0,a,β,H)××2×2 such that y0yPC and 𝔼[exp(2γAWTπ)]<+.}.\Gamma:=\left\{(y_{0},a,\beta,H)\in\mathbb{R}\times{\cal B}\times\mathbb{H}^{2}\times\mathbb{H}^{2}\text{ such that }y_{0}\geq y_{PC}\text{ and }\mathbb{E}\left[\exp(-2\gamma_{A}W_{T}^{\pi})\right]<+\infty.\right\}.

We have the following result.

Lemma 3.2.

For any πΓ\pi\in\Gamma, the pair (WTπ,a)(W_{T}^{\pi},a) belongs to 𝒜PC{\cal A}_{PC}.

Proof.

We apply Itô’s formula to the process Ytπ=UA(Wtπ)Y_{t}^{\pi}=U_{A}(W_{t}^{\pi}) to obtain

dYtπ=γAYtπβt(1Nt)dBt+Ytπ(eγAHt1)(1Nt)dMtγAκ(at(1Nt))Ytπdt.dY_{t}^{\pi}=-\gamma_{A}Y_{t}^{\pi}\beta_{t}(1-N_{t})\,dB_{t}+Y_{t}^{\pi}\left(e^{-\gamma_{A}H_{t}}-1\right)(1-N_{t})\,dM_{t}-\gamma_{A}\kappa(a_{t}(1-N_{t}))Y_{t}^{\pi}\,dt.

Moreover, because πΓ\pi\in\Gamma, YTπ=UA(WTπ)Y_{T}^{\pi}=U_{A}(W_{T}^{\pi}) is square-integrable. Remark 3.1 yields the triplet (Ytπ,γAYtπβt,Ytπ(eγAHt1))\left(Y_{t}^{\pi},-\gamma_{A}Y_{t}^{\pi}\beta_{t},Y_{t}^{\pi}(e^{-\gamma_{A}H_{t}}-1)\right) is the unique solution in (𝕊2×2×2)(\mathbb{S}^{2}\times\mathbb{H}^{2}\times\mathbb{H}^{2}) to BSDE (3.4) with terminal condition UA(WTπ)U_{A}(W_{T}^{\pi}) when πΓ\pi\in\Gamma. Therefore,

Y0π=UA(y0)=𝔼[UA(WTπ0Tκ(as(1Ns))𝑑s)]UA(yPC),Y_{0}^{\pi}=U_{A}(y_{0})=\mathbb{E}\left[U_{A}\left(W_{T}^{\pi}-\int_{0}^{T}\kappa(a_{s}(1-N_{s}))\,ds\right)\right]\geq U_{A}(y_{PC}),

and thus (3.2) is satisfied. ∎

Remark 3.2.

The admissible contracts are essentially the terminal values of the controlled processes (3.1) for πΓ\pi\in\Gamma. The difficulty is that we do not know how to characterize the β\beta and HH processes that guarantee that π\pi belongs to Γ\Gamma. Nevertheless, it is easy to check by a standard application of the Gronwall lemma that if β\beta and HH are bounded then πΓ\pi\in\Gamma. This last observation will prove to be crucial in the explicit resolution of our problem.

First-best Dynamic contracting

Using Lemma 3.2, the full Risk-Sharing problem under default writes as the Markovian control problem :

VPFB:=supπ=(y0,a,Z,K)Γ𝔼[UP(XT(x0,a)WTπ)],V_{P}^{FB}:=\underset{\pi=(y_{0},a,Z,K)\in\Gamma}{\text{sup}}\mathbb{E}\left[U_{P}\left(X_{T}^{(x_{0},a)}-W_{T}^{\pi}\right)\right], (3.6)

where Xt(x0,a)X_{t}^{(x_{0},a)} is given by :

dXs(x0,a)=as(1Ns)ds+(1Ns)dBs,dX^{(x_{0},a)}_{s}=a_{s}(1-N_{s})ds+(1-N_{s})dB_{s},

with X0(x0,a)=x0X_{0}^{(x_{0},a)}=x_{0} and the wage process is given by :

dWsπ=Zs(1Ns)dBs\displaystyle dW_{s}^{\pi}=Z_{s}(1-N_{s})dB_{s} +Ks(1Ns)dMs\displaystyle+K_{s}(1-N_{s})dM_{s}
+{γA2Zs2(1Ns)+κ(as(1Ns))+λγA[exp(γAKs)1+γAKs](1Ns)}ds,\displaystyle+\left\{\frac{\gamma_{A}}{2}Z_{s}^{2}(1-N_{s})+\kappa(a_{s}(1-N_{s}))+\frac{\lambda}{\gamma_{A}}[\exp(-\gamma_{A}K_{s})-1+\gamma_{A}K_{s}](1-N_{s})\right\}ds,

with W0π=y0W_{0}^{\pi}=y_{0}.
We have the following key theorem for the first-best problem.

Theorem 3.1.

Let at=1κ,Zt=γPγP+γA,a^{*}_{t}=\dfrac{1}{\kappa},Z^{*}_{t}=\dfrac{\gamma_{P}}{\gamma_{P}+\gamma_{A}}, and let :

Kt=1γP+γAlog(Φ0(t)),K^{*}_{t}=\frac{1}{\gamma_{P}+\gamma_{A}}\log(\Phi_{0}(t)),

where :

Φ0(t):=(c1+c2c1exp(c1γAγP+γA(Tt))c2c1)γP+γAγA,\Phi_{0}(t):=\left(\frac{c_{1}+c_{2}}{c_{1}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\frac{c_{2}}{c_{1}}\right)^{\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}}, (3.7)

with :

c1:=γP2γA2(γP+γA)γP2κλγP+γAγAandc2:=λγP+γAγA.c_{1}:=\frac{\gamma_{P}^{2}\gamma_{A}}{2(\gamma_{P}+\gamma_{A})}-\frac{\gamma_{P}}{2\kappa}-\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\quad\text{and}\quad c_{2}:=\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}.\\

Then π=(yPC,a,Z,K)Γ\pi^{*}=(y_{PC},a^{*},Z^{*},K^{*})\in\Gamma parameterizes the optimal contract (WTπ,a)(W^{\pi^{*}}_{T},a^{*}) for the first-best problem.

The rest of this subsection is dedicated to the proof of this Theorem. We first make the following observation. As XX remains constant after τ\tau, the principal has no further decision to make after the default time. Thus, its value function is constant and equal to UP(xy)U_{P}(x-y) on the interval [τ,T][\tau,T].
We now focus on the control part of the problem (i.e. computation of the optimal control triplet π~=(a,Z,K)\tilde{\pi}=(a,Z,K) for a given pair (x0,y0)(x_{0},y_{0})). To do so, we follow the dynamic programming approach developed in [17], Section 4 to define the value function

V(0,x0,y0)=supπ~Γ~𝔼[UP(XTaWTπ~)(1NT)+0TUP(XtaWtπ~)λeλt𝑑t],V(0,x_{0},y_{0})=\sup_{\tilde{\pi}\in\tilde{\Gamma}}\mathbb{E}\left[U_{P}(X_{T}^{a}-W_{T}^{\tilde{\pi}})(1-N_{T})+\int_{0}^{T}U_{P}(X_{t}^{a}-W_{t}^{\tilde{\pi}})\lambda e^{-\lambda t}\,dt\right], (3.8)

where

Γ~={π~×2×2},\tilde{\Gamma}=\left\{\tilde{\pi}\in\mathcal{B}\times\mathbb{H}^{2}\times\mathbb{H}^{2}\right\},

Because Γ×Γ~\Gamma\subset\mathbb{R}\times\tilde{\Gamma}, we have

VPFBsupy0yPCV(0,x0,y0).V_{P}^{FB}\leq\sup_{y_{0}\geq y_{PC}}V(0,x_{0},y_{0}).

According to stochastic control theory, the Hamilton-Jacobi-Bellman equation associated to the stochastic control problem (3.8) is the following (see [16]):

tv(t,x,y)+supa,Z,K{xv(t,x,y)a+yv(t,x,y)[γA2Z2+κ(a)+λγA[exp(γAK)1]]\displaystyle\partial_{t}v(t,x,y)+\sup_{a,Z,K}\left\{\partial_{x}v(t,x,y)a+\partial_{y}v(t,x,y)\left[\frac{\gamma_{A}}{2}Z^{2}+\kappa(a)+\frac{\lambda}{\gamma_{A}}[\exp(-\gamma_{A}K)-1]\right]\right.
+λ[UP(xyK)v0(t,x,y)]+yyv(t,x,y)Z22+12xxv(t,x,y)+xyv(t,x,y)Z}=0,\displaystyle\left.+\lambda\left[U_{P}(x-y-K)-v_{0}(t,x,y)\right]+\partial_{yy}v(t,x,y)\frac{Z^{2}}{2}+\frac{1}{2}\partial_{xx}v(t,x,y)+\partial_{xy}v(t,x,y)Z\right\}=0, (3.9)

with the boundary condition :

v(T,x,y)=UP(xy).v(T,x,y)=U_{P}(x-y).

It happens that the HJB equation (3.2) is explicitly solvable by exploiting the separability property of the exponential utility function.

Lemma 3.3.

The function v(t,x,y)=UP(xy)Φ0(t)v(t,x,y)=U_{P}(x-y)\Phi_{0}(t) with :

Φ0(t)=(c1+c2c1exp(c1γAγP+γA(Tt))c2c1)γP+γAγA,\Phi_{0}(t)=\left(\frac{c_{1}+c_{2}}{c_{1}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\frac{c_{2}}{c_{1}}\right)^{\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}},

where :

c1=γP2γA2(γP+γA)γP2κλγP+γAγAandc2=λγP+γAγA,c_{1}=\frac{\gamma_{P}^{2}\gamma_{A}}{2(\gamma_{P}+\gamma_{A})}-\frac{\gamma_{P}}{2\kappa}-\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\quad\text{and}\quad c_{2}=\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}},

solves (in the classical sense) the HJB partial differential equation (3.2).
Furthermore at=1κ,Zt=γPγP+γAa^{*}_{t}=\dfrac{1}{\kappa},Z^{*}_{t}=\dfrac{\gamma_{P}}{\gamma_{P}+\gamma_{A}} and Kt=1γP+γAlog(Φ0(t))K^{*}_{t}=\dfrac{1}{\gamma_{P}+\gamma_{A}}\log(\Phi_{0}(t)) are the optimal controls.

Proof.

We search for a solution to Equation (3.2) for a vv of the form :

v(t,x,y)=UP(xy)Φ0(t),v(t,x,y)=U_{P}(x-y)\Phi_{0}(t),

with Φ0\Phi_{0} a positive mapping. Such a vv satisfies (3.2) if and only if Φ0(t)\Phi_{0}(t) solves the PDE :

Φ0(t)+infa,Z,K{γPΦ0(t)a+γPΦ0(t)(γA2Z2+κ(a)+λγA{exp(γAK)1})\displaystyle\Phi_{0}^{\prime}(t)+\inf_{a,Z,K}\left\{-\gamma_{P}\Phi_{0}(t)a+\gamma_{P}\Phi_{0}(t)\left(\frac{\gamma_{A}}{2}Z^{2}+\kappa(a)+\frac{\lambda}{\gamma_{A}}\left\{\exp(-\gamma_{A}K)-1\right\}\right)\right.
+γP2Φ0(t)Z22+γP22Φ0(t)γP2Φ0(t)Z+λ(exp(γPK)Φ0(t))}=0,\displaystyle\left.+\gamma_{P}^{2}\Phi_{0}(t)\frac{Z^{2}}{2}+\frac{\gamma_{P}^{2}}{2}\Phi_{0}(t)-\gamma_{P}^{2}\Phi_{0}(t)Z+\lambda\left(\exp(\gamma_{P}K)-\Phi_{0}(t)\right)\right\}=0,

with the boundary condition Φ0(T)=1.\Phi_{0}(T)=1. As Φ0\Phi_{0} is a positive mapping, the infimum is well defined. We derive the following first order conditions that must be satisfied by the optimal controls :

{γPΦ0(t)=γPκaΦ0(t)γPΦ0(t)Z(γA+γP)=γP2Φ0(t)γPΦ0(t)λexp(γAK)=γPλexp(γPK),\begin{cases}\gamma_{P}\Phi_{0}(t)=\gamma_{P}\kappa a\Phi_{0}(t)\\ \gamma_{P}\Phi_{0}(t)Z(\gamma_{A}+\gamma_{P})=\gamma_{P}^{2}\Phi_{0}(t)\\ \gamma_{P}\Phi_{0}(t)\lambda\exp(-\gamma_{A}K)=\gamma_{P}\lambda\exp(\gamma_{P}K),\end{cases}

equating to :

a=1κ,Z=γPγP+γA,K=log(Φ0(t))γP+γA.a^{*}=\frac{1}{\kappa},\quad Z^{*}=\frac{\gamma_{P}}{\gamma_{P}+\gamma_{A}},\quad K^{*}=\frac{\log(\Phi_{0}(t))}{\gamma_{P}+\gamma_{A}}.

It follows that :

infa,Z,K{γPΦ0(t)a+γPΦ0(t)(γA2Z2+κ(a)+λγA{exp(γAK)1})\displaystyle\inf_{a,Z,K}\left\{-\gamma_{P}\Phi_{0}(t)a+\gamma_{P}\Phi_{0}(t)\left(\frac{\gamma_{A}}{2}Z^{2}+\kappa(a)+\frac{\lambda}{\gamma_{A}}\left\{\exp(-\gamma_{A}K)-1\right\}\right)\right.
+γP2Φ0(t)Z22+γP22Φ0(t)γP2Φ0(t)Z+λ(exp(γPK)Φ0(t))}\displaystyle\left.+\gamma_{P}^{2}\Phi_{0}(t)\frac{Z^{2}}{2}+\frac{\gamma_{P}^{2}}{2}\Phi_{0}(t)-\gamma_{P}^{2}\Phi_{0}(t)Z+\lambda\left(\exp(\gamma_{P}K)-\Phi_{0}(t)\right)\right\}
=γPΦ0(t)a+γPΦ0(t)(γA2Z2+κ(a)+λγA{exp(γAK)1})\displaystyle=-\gamma_{P}\Phi_{0}(t)a^{*}+\gamma_{P}\Phi_{0}(t)\left(\frac{\gamma_{A}}{2}{Z^{*}}^{2}+\kappa(a^{*})+\frac{\lambda}{\gamma_{A}}\left\{\exp(-\gamma_{A}K^{*})-1\right\}\right)
+γP2Φ0(t)Z22+γP22Φ0(t)γP2Φ0(t)Z+λ(exp(γPK)Φ0(t))\displaystyle+\gamma_{P}^{2}\Phi_{0}(t)\frac{{Z^{*}}^{2}}{2}+\frac{\gamma_{P}^{2}}{2}\Phi_{0}(t)-\gamma_{P}^{2}\Phi_{0}(t)Z^{*}+\lambda\left(\exp(\gamma_{P}K^{*})-\Phi_{0}(t)\right)
=Φ0(t)γP2γA2(γP+γA)terms with ZΦ0(t)γP2κterms with aλγP+γAγAΦ0(t)+λγP+γAγAΦ0(t)γPγP+γA.terms with K\displaystyle=\underbrace{\Phi_{0}(t)\frac{\gamma_{P}^{2}\gamma_{A}}{2(\gamma_{P}+\gamma_{A})}}_{\text{terms with }Z^{*}}\underbrace{-\Phi_{0}(t)\frac{\gamma_{P}}{2\kappa}}_{\text{terms with }a^{*}}\underbrace{-\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\Phi_{0}(t)+\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\Phi_{0}(t)^{\frac{\gamma_{P}}{\gamma_{P}+\gamma_{A}}}.}_{\text{terms with }K^{*}}

We may inject this expression back into the PDE on Φ0\Phi_{0}. Doing so yields the following Bernoulli equation :

Φ0(t)+c1Φ0(t)+c2Φ0(t)γPγP+γA=0,Φ0(T)=1,\Phi^{\prime}_{0}(t)+c_{1}\Phi_{0}(t)+c_{2}\Phi_{0}(t)^{\frac{\gamma_{P}}{\gamma_{P}+\gamma_{A}}}=0,\quad\Phi_{0}(T)=1,

where

c1=γP2γA2(γP+γA)γP2κλγP+γAγAandc2=λγP+γAγA.c_{1}=\frac{\gamma_{P}^{2}\gamma_{A}}{2(\gamma_{P}+\gamma_{A})}-\frac{\gamma_{P}}{2\kappa}-\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\quad\text{and}\quad c_{2}=\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}.

The unique solution to this equation is (see for instance [22]) :

Φ0(t)=(c1+c2c1exp(c1γAγP+γA(Tt))c2c1)γP+γAγA,\Phi_{0}(t)=\left(\frac{c_{1}+c_{2}}{c_{1}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\frac{c_{2}}{c_{1}}\right)^{\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}},

and the result follows. ∎

Proof of Theorem 3.1.

The value function v(t,x,y)=UP(xy)Φ0(t)v(t,x,y)=U_{P}(x-y)\Phi_{0}(t) is a classical solution to the HJB equation (3.2). A standard verification theorem yields that v=Vv=V. Through Lemma 3.3, the optimal controls for the full Risk-Sharing problem are :

at=1κ,Zt=γPγP+γAandKt=1γP+γAlog(Φ0(t)),a^{*}_{t}=\dfrac{1}{\kappa},Z^{*}_{t}=\dfrac{\gamma_{P}}{\gamma_{P}+\gamma_{A}}\;\text{and}\;K^{*}_{t}=\dfrac{1}{\gamma_{P}+\gamma_{A}}\log(\Phi_{0}(t)),

with Φ0\Phi_{0} as defined in Lemma 3.3. These controls are free of yy and it follows that :

V(0,x0,y0)=E[UP(XT(x0,a)WT(y0,a,Z,K))],V(0,x_{0},y_{0})=E\left[U_{P}\left(X_{T}^{(x_{0},a^{*})}-W_{T}^{(y_{0},a^{*},Z^{*},K^{*})}\right)\right],

is a decreasing function of y0y_{0}. Thus we obtain

supy0yPCV0(0,x0,y0)=E[UP(XT(x0,a)WT(yPC,a,Z,K))].\sup_{y_{0}\geq y_{PC}}V_{0}(0,x_{0},y_{0})=E\left[U_{P}\left(X_{T}^{(x_{0},a^{*})}-W_{T}^{(y_{PC},a^{*},Z^{*},K^{*})}\right)\right].

Finally, we observe that the optimal controls are bounded and thus Remark (3.2) yields π=(yPC,a,Z,K)Γ\pi^{*}=(y_{PC},a^{*},Z^{*},K^{*})\in\Gamma. As a consequence,

supy0yPCV(0,x0,y0)=𝔼[UP(XT(x0,a)WT(yPC,a,Z,K))]VPFB.\sup_{y_{0}\geq y_{PC}}V(0,x_{0},y_{0})=\mathbb{E}\left[U_{P}\left(X_{T}^{(x_{0},a)}-W_{T}^{(y_{PC},a^{*},Z^{*},K^{*})}\right)\right]\leq V_{P}^{FB}.

Because the reverse inequality holds, the final result follows. ∎

4  Optimal contracting under shutdown risk

Main results

The following is dedicated to our main result for the Moral Hazard problem. We shall state our main theorem with the explicit optimal contract before turning to some analysis of the effect of the shutdown on dynamic contracting. In the case of moral hazard, one is forced to make a stronger assumption about the nature of a contract. This stronger hypothesis will naturally appear to justify the martingale optimality principle. In our setting, a contract is a 𝔾Tτ\mathbb{G}_{T\wedge\tau} measurable random variable WW such that for every β\beta\in\mathbb{R}, we have

𝔼[exp(βW)]<+.\mathbb{E}\left[\exp(\beta W)\right]<+\infty.

A first step to optimal contracting involves answering the preliminary question: can we characterize incentive compatible wages and if so what is the related optimal action for the agent? The characterization of incentive compatible contracts relies on the martingale optimality principle (see [13] and [18]) that we recall below.

Lemma 4.1 (Martingale Optimality Principle).

Given a contract WW, consider a family of stochastic processes Ra(W):=(Rta)t[0,T]R^{a}(W):=(R_{t}^{a})_{t\in[0,T]} indexed by aa in \mathcal{B} that satisfies :

  1. 1.

    RTa=UA(W0Tκ(as(1Ns))𝑑s)R_{T}^{a}=U_{A}(W-\int_{0}^{T}\kappa(a_{s}(1-N_{s}))ds) for any aa in \mathcal{B}

  2. 2.

    R.aR^{a}_{.} is a a\mathbb{P}^{a}-supermartingale for any aa in \mathcal{B}

  3. 3.

    R0aR^{a}_{0} is independent of aa.

  4. 4.

    There exists aa^{*} in \mathcal{B} such that RaR^{a^{*}} is a a\mathbb{P}^{a^{*}}-martingale.

Then,

R0a=𝔼a[UA(W0Tκ(as)𝑑s)]𝔼a[UA(W0Tκ(as(1Ns))𝑑s)],R_{0}^{a^{*}}=\mathbb{E}^{a^{*}}\left[U_{A}(W-\int_{0}^{T}\kappa(a^{*}_{s})ds)\right]\geq\mathbb{E}^{a}\left[U_{A}(W-\int_{0}^{T}\kappa(a_{s}(1-N_{s}))ds)\right],

meaning that aa^{*} is the optimal agent’s action in response to the contract WW.

We will construct such a family following the standard route. Consider a given contract WW, we define the family Ra(W):=(Rta)t[0,T]R^{a}(W):=(R_{t}^{a})_{t\in[0,T]} by

Rta:=exp(γA(Yt(W)0tκ(as(1Ns))𝑑s)),R_{t}^{a}:=-\exp\left(-\gamma_{A}\left(Y_{t}(W)-\int_{0}^{t}\kappa(a_{s}(1-N_{s}))ds\right)\right),

where (Y(W),Z(W),K(W))(Y(W),Z(W),K(W)) in (𝕊2×2×2)(\mathbb{S}^{2}\times\mathbb{H}^{2}\times\mathbb{H}^{2}) is the unique solution of the following BSDE under 0\mathbb{P}^{0}

Yt(W)=WtTf(Zs(W),Ks(W))(1Ns)𝑑stTZs(W)(1Ns)𝑑BstTKs(W)(1Ns)𝑑Ms,Y_{t}(W)=W-\int_{t}^{T}f(Z_{s}(W),K_{s}(W))(1-N_{s})\,ds-\int_{t}^{T}Z_{s}(W)(1-N_{s})\,dB_{s}-\int_{t}^{T}K_{s}(W)(1-N_{s})\,dM_{s}, (4.1)

with

f(z,k):=12γAz2+λk+λγA(eγAk1)+infa{κ(a)az}.f(z,k):=\frac{1}{2}\gamma_{A}z^{2}+\lambda k+\frac{\lambda}{\gamma_{A}}(e^{-\gamma_{A}k}-1)+\inf_{a\in\mathcal{B}}\left\{\kappa(a)-az\right\}.
Remark 4.1.

The theoretical justification of the well-posedness of the BSDE (4.1) deserves some comments. The first results were obtained in [14] and [8] when the contract WW is assumed to be bounded. The necessary extension in our model when WW admits an exponential moment has been treated recently in the paper [15].

By construction, RTa=UA(W0Tκ(as(1Ns))𝑑s)R_{T}^{a}=U_{A}(W-\int_{0}^{T}\kappa(a_{s}(1-N_{s}))ds) for any aa in \mathcal{B}. Moreover, R0a=Y0(W)R^{a}_{0}=Y_{0}(W) is independent of the agent’s action aa. We compute the variations of RaR^{a} and obtain :

=γARsaZs(1Ns)dBs+Rsa(eγAKs1)(1Ns)dMs\displaystyle=-\gamma_{A}R_{s}^{a}Z_{s}(1-N_{s})dB_{s}+R_{s}^{a}(e^{-\gamma_{A}K_{s}}-1)(1-N_{s})dM_{s}
+RsaγA{12γAZs2f(Zs,Ks)+κ(as(1Ns))+λKs+λγA(eγAKs1)}(1Ns)ds.\displaystyle+R_{s}^{a}\gamma_{A}\left\{\frac{1}{2}\gamma_{A}Z_{s}^{2}-f(Z_{s},K_{s})+\kappa(a_{s}(1-N_{s}))+\lambda K_{s}+\frac{\lambda}{\gamma_{A}}(e^{-\gamma_{A}K_{s}}-1)\right\}(1-N_{s})ds.
=γARsaZs(1Ns)dBsa+Rsa(eγAKs1)(1Ns)dMs\displaystyle=-\gamma_{A}R_{s}^{a}Z_{s}(1-N_{s})dB_{s}^{a}+R_{s}^{a}(e^{-\gamma_{A}K_{s}}-1)(1-N_{s})dM_{s}
+RsaγA{12γAZs2f(Zs,Ks)+κ(as(1Ns))+λKs+λγA(eγAKs1)asZs}(1Ns)ds.\displaystyle+R_{s}^{a}\gamma_{A}\left\{\frac{1}{2}\gamma_{A}Z_{s}^{2}-f(Z_{s},K_{s})+\kappa(a_{s}(1-N_{s}))+\lambda K_{s}+\frac{\lambda}{\gamma_{A}}(e^{-\gamma_{A}K_{s}}-1)-a_{s}Z_{s}\right\}(1-N_{s})ds.

Thus RaR^{a} is a a\mathbb{P}^{a}-super-martingale for every aa in \mathcal{B}, the function

a(z)=A11zκA+zκ11κAzκA+A11zκAa^{*}(z)=-A1\!\!1_{z\leq-\kappa A}+\frac{z}{\kappa}1\!\!1_{-\kappa A\leq z\leq\kappa A}+A1\!\!1_{z\geq\kappa A}

is a unique minimizer for ff and RaR^{a^{*}} is a a\mathbb{P}^{a^{*}}-martingale. As a consequence, every contract WW is incentive compatible which a unique best reply a(Z(W))a^{*}(Z(W)). Finally, a contract WW satisfies the participation constraint if and only if Y0(W)yPCY_{0}(W)\geq y_{PC}.

Relying on the idea of Sannikov [19] and its recent theoretical justification by Cvitanic, Possamai and Touzi [6], we will consider the agent promised wage Y(W)Y(W) as a state variable to embed the principal’s problem into the class of Markovian problems, by considering the sensitivities of the agent’s promised wage Z(W)Z(W) and K(W)K(W) as control variables. For π=(y0,Z,K)[yPC;+)×2×2\pi=(y_{0},Z,K)\in[y_{PC};+\infty)\times\mathbb{H}^{2}\times\mathbb{H}^{2}, we define under 0\mathbb{P}^{0}, the control process called the agent continuation value

Wt(y0,Z,K)=y0+0tZs(1Ns)𝑑Bs+0tKs(1Ns)𝑑Ms+0tf(Zs,Ks)(1Ns)𝑑s.W^{(y_{0},Z,K)}_{t}=y_{0}+\int_{0}^{t}Z_{s}(1-N_{s})dB_{s}+\int_{0}^{t}K_{s}(1-N_{s})dM_{s}+\int_{0}^{t}f(Z_{s},K_{s})(1-N_{s})\,ds. (4.2)

Under :=(a(Z))\mathbb{P}^{*}:=\mathbb{P}^{(a^{*}(Z))}, we thus have

Wt(y0,Z,K)\displaystyle W^{(y_{0},Z,K)}_{t} =y0+0tZs(1Ns)𝑑Bs+0tKs(1Ns)𝑑Ms\displaystyle=y_{0}+\int_{0}^{t}Z_{s}(1-N_{s})dB_{s}^{*}+\int_{0}^{t}K_{s}(1-N_{s})dM_{s} (4.3)
+0t{γA2Zs2+κ(a(Zs))+λγA[exp(γAKs)1+γAKs]}(1Ns)𝑑s\displaystyle+\int_{0}^{t}\left\{\frac{\gamma_{A}}{2}Z_{s}^{2}+\kappa(a^{*}(Z_{s}))+\frac{\lambda}{\gamma_{A}}[\exp(-\gamma_{A}K_{s})-1+\gamma_{A}K_{s}]\right\}(1-N_{s})ds
=y0+0tZs(1Ns)𝑑Bs+0tKs(1Ns)𝑑Ns\displaystyle=y_{0}+\int_{0}^{t}Z_{s}(1-N_{s})dB_{s}^{*}+\int_{0}^{t}K_{s}(1-N_{s})dN_{s}
+0t{γA2Zs2+κ(a(Zs))+λγA[exp(γAKs)1]}(1Ns)𝑑s\displaystyle+\int_{0}^{t}\left\{\frac{\gamma_{A}}{2}Z_{s}^{2}+\kappa(a^{*}(Z_{s}))+\frac{\lambda}{\gamma_{A}}[\exp(-\gamma_{A}K_{s})-1]\right\}(1-N_{s})ds

Now, we consider the set

ζ={π=(Z,K)2×2 such that β,𝔼[exp(βWT(y,Z,K))]<+ for y}.\zeta=\left\{\pi=(Z,K)\in\mathbb{H}^{2}\times\mathbb{H}^{2}\text{ such that }\forall\beta\in\mathbb{R},\,\mathbb{E}\left[\exp(\beta W_{T}^{(y,Z,K)})\right]<+\infty\text{ for }y\in\mathbb{R}\right\}.

By construction, WT(y,π)W^{(y,\pi)}_{T} is a contract that satisfies the participation constraint for every πζ\pi\in\zeta and yyPCy\geq y_{PC}. Moreover, by the well-posedness of the BSDE (4.1) , every contract WW that satisfies the participation constraint can be written WT(Y0(W),Z(W),K(W))W^{(Y_{0}(W),Z(W),K(W))}_{T} with π(W)=(Z(W),K(W))ζ\pi(W)=(Z(W),K(W))\in\zeta. Therefore, the problem of the principal can now be rewritten as the following optimisation problem

VP:=supyypcv(0,x,y),V_{P}:=\sup_{y\geq y_{pc}}v(0,x,y),

where

v(0,x,y)=supπζ𝔼[UP(XTτWTτπ)]v(0,x,y)=\sup_{\pi\in\zeta}\mathbb{E}^{*}\left[U_{P}(X_{T\wedge\tau}-W_{T\wedge\tau}^{\pi})\right] (4.4)

To characterize the optimal contract, we will proceed analogously as in the full risk sharing case by constructing a smooth solution to the HJB equation associated to the Markov control problem (4.4) given by

0=tv(t,x,y)+infZ,K{xv(t,x,y)Zκ+yv(t,x,y)[γA2Z2+κ(a(Z))+λγA[exp(γAK)1]]\displaystyle 0=\partial_{t}v(t,x,y)+\inf_{Z,K}\left\{\partial_{x}v(t,x,y)\dfrac{Z}{\kappa}+\partial_{y}v(t,x,y)\left[\frac{\gamma_{A}}{2}Z^{2}+\kappa(a^{*}(Z))+\frac{\lambda}{\gamma_{A}}[\exp(-\gamma_{A}K)-1]\right]\right.
+λ[UP(xyK)v(t,x,y)]+yyv0(t,x,y)Z22+12xxv(t,x,y)+xyv(t,x,y)Z},\displaystyle\left.+\lambda\left[U_{P}(x-y-K)-v(t,x,y)\right]+\partial_{yy}v_{0}(t,x,y)\frac{Z^{2}}{2}+\frac{1}{2}\partial_{xx}v(t,x,y)+\partial_{xy}v(t,x,y)Z\right\}, (4.5)
Lemma 4.2.

Assume the constant AA in the definition of the set of admissible efforts {\cal B} satisfies

A>γP+κ1κ(γP+γA)+1.A>\dfrac{\gamma_{P}+\kappa^{-1}}{\kappa(\gamma_{P}+\gamma_{A})+1}\,.

Then, the function UP(xy)Φ0(t),U_{P}(x-y)\Phi_{0}(t), with

Φ0(t)=(c1+c2c1exp(c1γAγP+γA(Tt))c2c1)γP+γAγA,\Phi_{0}(t)=\left(\frac{c_{1}+c_{2}}{c_{1}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\frac{c_{2}}{c_{1}}\right)^{\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}},

where

c1=γP2γA2(γP+γA+κ1)γPκ1(γP+κ1)2(γP+γA+κ1)λγP+γAγAandc2=λγP+γAγA.c_{1}=\frac{\gamma_{P}^{2}\gamma_{A}}{2(\gamma_{P}+\gamma_{A}+\kappa^{-1})}-\frac{\gamma_{P}\kappa^{-1}(\gamma_{P}+\kappa^{-1})}{2{(\gamma_{P}+\gamma_{A}+\kappa^{-1})}}-\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\quad\text{and}\quad c_{2}=\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}.

solves in the classical sense the HJB equation (5.2). In particular Zt=γP+κ1γP+γA+κ1Z^{*}_{t}=\dfrac{\gamma_{P}+\kappa^{-1}}{\gamma_{P}+\gamma_{A}+\kappa^{-1}} and Kt=1γP+γAlog(Φ0(t)),K^{*}_{t}=\dfrac{1}{\gamma_{P}+\gamma_{A}}\log(\Phi_{0}(t)),

Proof.

Because the assumption on AA implies a(z)=z/κa^{*}(z)=z/\kappa, the proof of this lemma is a direct adaptation of the proof of Lemma 3.3 to which we refer the reader. ∎

We are in a position to prove the main result of this section

Theorem 4.1.

We have the following explicit characterizations of the optimal contracts. Let AA as in the Lemma 4.2 and let Zt=γP+κ1γP+γA+κ1Z^{*}_{t}=\dfrac{\gamma_{P}+\kappa^{-1}}{\gamma_{P}+\gamma_{A}+\kappa^{-1}} and Kt=1γP+γAlog(Φ0(t)),K^{*}_{t}=\frac{1}{\gamma_{P}+\gamma_{A}}\log(\Phi_{0}(t)), where Φ0\Phi_{0} is defined as in (3.7) with the constants :

c1:=γP2γA2(γP+γA+κ1)γPκ1(γP+κ1)2(γP+γA+κ1)λγP+γAγAandc2:=λγP+γAγA.c_{1}:=\frac{\gamma_{P}^{2}\gamma_{A}}{2(\gamma_{P}+\gamma_{A}+\kappa^{-1})}-\frac{\gamma_{P}\kappa^{-1}(\gamma_{P}+\kappa^{-1})}{2{(\gamma_{P}+\gamma_{A}+\kappa^{-1})}}-\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\quad\text{and}\quad c_{2}:=\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}.\\

Then (yPC,Z,K)(y_{PC},Z^{*},K^{*}) parametrizes the optimal wage for the Moral Hazard problem. The Agent performs the optimal action Zκ\frac{Z^{*}}{\kappa}.

Proof.

Because the function UP(xy)Φ0(t)U_{P}(x-y)\Phi_{0}(t) is a classical solution to the HJB equation (5.2) and the optimal controls are bounded and free of yy, we proceed analogously as in the proof of Theorem 3.1. Finally, we have to prove that the optimal wage W=YT(yPC,Z,K)W^{*}=Y_{T}^{(y_{PC},Z^{*},K^{*})} admits exponential moments to close the loop. According to (4.3), we have

W=yPC+ZBTτ+12(γA+1κ)(Z)2(Tτ)+Kτ1τT+0TλγA[exp(γAKs)1](1Ns)𝑑s.W^{*}=y_{PC}+Z^{*}B^{*}_{T\wedge\tau}+\frac{1}{2}\left(\gamma_{A}+\frac{1}{\kappa}\right)(Z^{*})^{2}(T\wedge\tau)+K^{*}_{\tau_{-}}\textbf{1}_{\tau\leq T}+\int_{0}^{T}\frac{\lambda}{\gamma_{A}}[\exp(-\gamma_{A}K^{*}_{s})-1](1-N_{s})ds.

Because (Bt)t(B^{*}_{t})_{t} is a Brownian motion and KtK^{*}_{t} is deterministic, it is straightforward to check that WW^{*} admits exponential moments. ∎

Model analysis

The optimal contract includes two components. One is linear in the output with an incentivizing slope that is similar to the classical optimal contract found in [12]. This is necessary to implement a desirable level of effort. The other is unrelated to the incentives but linked to the shutdown risk sharing. It is key to observe that this second term is nonzero even if the shutdown risk does not materialize before the termination of the contract.
The characterization of the optimal contracts in Theorem 4.1 sparks an immediate observation: the two parties only need to be committed to the contracting agreement up until TτT\wedge\tau. Therefore in this simple model, using an expected-utility related reasoning and without considering mechanisms such as employment law, the occurence of the agency-free external risk, halting production, leads to early contract terminations. This is in line with what actually happened during the Covid pandemic. Indeed in the USA and in eight weeks of the pandemic, 36.5 million people applied for unemployment insurance. In more protective economies, mass redundancies were only prevented through the instauration of furlough type schemes allowing private employees’ wages to temporarily be paid by gouvernements. This phenomena makes fundamental sense : a principal whose output process is completely halted cannot enforce the agent to work hard because she has no revenue to provide the incentives. Let’s focus on the second term:

Kτ1τT+0TλγA[exp(γAKs)1](1Ns)𝑑s,K^{*}_{\tau_{-}}\textbf{1}_{\tau\leq T}+\int_{0}^{T}\frac{\lambda}{\gamma_{A}}[\exp(-\gamma_{A}K^{*}_{s})-1](1-N_{s})ds, (4.6)

Understanding the effect of these extra terms is crucial to fully understand the sharing of the agency-free shutdown risk. First, we show that the sign of the control KK^{*} is constant.

Lemma 4.3.

Let c1c_{1} and c2c_{2} be the relevant constants given in Theorem 4.1 then the optimal control (Kt)t[0,T](K^{*}_{t})_{t\in[0,T]} can be expressed as :

Kt=1γAlog(𝔼[exp(γAγP+γA(c1+c2)((Tt)τ))])t[0,T].K_{t}^{*}=\frac{1}{\gamma_{A}}\log\left(\mathbb{E}\left[\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})((T-t)\wedge\tau)\right)\right]\right)\quad t\in[0,T]. (4.7)
Proof.

We have that :

Kt=1γP+γAlog(Φ0(t)),K^{*}_{t}=\dfrac{1}{\gamma_{P}+\gamma_{A}}\log(\Phi_{0}(t)),

with

Φ0(t)=(c1+c2c1exp(c1γAγP+γA(Tt))c2c1)γP+γAγA.\Phi_{0}(t)=\left(\frac{c_{1}+c_{2}}{c_{1}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\frac{c_{2}}{c_{1}}\right)^{\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}}.

The aim here is to link this expression for Φ0\Phi_{0} to that of an expected value. As such, we consider the following expected value that decomposes as shown :

𝔼[exp(γAγP+γA(c1+c2)((Tt)τ))]\displaystyle\mathbb{E}\left[\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})((T-t)\wedge\tau)\right)\right]
=𝔼[exp(γAγP+γA(c1+c2)(Tt))1τ>Tt]+𝔼[exp(γAγP+γA(c1+c2)τ)1τTt].\displaystyle=\mathbb{E}\left[\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})(T-t)\right)\textbf{1}_{\tau>T-t}\right]+\mathbb{E}\left[\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})\tau\right)\textbf{1}_{\tau\leq T-t}\right].

Using c2=γP+γAγAλc_{2}=\dfrac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\lambda, the first term of the expected value rewrites as follows :

𝔼[exp(γAγP+γA(c1+c2)(Tt))1τ>Tt]\displaystyle\mathbb{E}\left[\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})(T-t)\right)\textbf{1}_{\tau>T-t}\right] =exp(γAγP+γA(c1+c2)(Tt))exp(λ(Tt))\displaystyle=\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})(T-t)\right)\exp\left(-\lambda(T-t)\right)
=exp(c1γAγP+γA(Tt)).\displaystyle=\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right).

It remains to compute the second term. We obtain :

𝔼[exp(γAγP+γA(c1+c2)τ)1τTt]\displaystyle\mathbb{E}\left[\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})\tau\right)\textbf{1}_{\tau\leq T-t}\right] =0Ttλexp(γAγP+γA(c1+c2)s)exp(λs)𝑑s\displaystyle=\int_{0}^{T-t}\lambda\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})s\right)\exp\left(-\lambda s\right)ds
=0Ttλexp(γAγP+γA(c1+c2)sλs)𝑑s\displaystyle=\int_{0}^{T-t}\lambda\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})s-\lambda s\right)ds
=0Ttλexp(γAγP+γAc1s)𝑑s\displaystyle=\int_{0}^{T-t}\lambda\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}c_{1}s\right)ds
=[λc1γP+γAγAexp(γAγP+γAc1s)]0Tt\displaystyle=\left[\frac{\lambda}{c_{1}}\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}c_{1}s\right)\right]_{0}^{T-t}
=[c2c1exp(γAγP+γAc1s)]0Tt\displaystyle=\left[\frac{c_{2}}{c_{1}}\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}c_{1}s\right)\right]_{0}^{T-t}
=c2c1exp(c1γAγP+γA(Tt))c2c1.\displaystyle=\frac{c_{2}}{c_{1}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\frac{c_{2}}{c_{1}}.

Combining both terms we reach the final expression :

𝔼[exp(γAγP+γA(c1+c2)((Tt)τ))]\displaystyle\mathbb{E}\left[\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})((T-t)\wedge\tau)\right)\right] =c1+c2c1exp(c1γAγP+γA(Tt))c2c1.\displaystyle=\frac{c_{1}+c_{2}}{c_{1}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\frac{c_{2}}{c_{1}}.

Therefore we identify that :

Φ0(t)=(𝔼[exp(γAγP+γA(c1+c2)((Tt)τ))])γP+γAγA.\Phi_{0}(t)=\left(\mathbb{E}\left[\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})((T-t)\wedge\tau)\right)\right]\right)^{\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}}.

As a consequence, we may also rewrite Kt.K^{*}_{t}. Indeed :

Kt=1γP+γAlog(Φ0(t)),K^{*}_{t}=\frac{1}{\gamma_{P}+\gamma_{A}}\log(\Phi_{0}(t)),

and with the new expression for Φ0\Phi_{0} we obtain the result :

Kt=1γAlog(𝔼[exp(γAγP+γA(c1+c2)((Tt)τ))]).K_{t}^{*}=\frac{1}{\gamma_{A}}\log\left(\mathbb{E}\left[\exp\left(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(c_{1}+c_{2})((T-t)\wedge\tau)\right)\right]\right).

Remark 4.2.

We have the same expression for the optimal control KK^{*} in the first-best case, using for c1c_{1} and c2c_{2} the relevant constants given in Theorem 3.1.

As a consequence, this alternative form for KK^{*} leads to easy analysis of the sign of the control, given in the following lemma.

Lemma 4.4.

The sign of KK^{*} over the contracting period [0,T][0,T] is constant and entirely determined by the model’s risk aversions γP\gamma_{P} and γA\gamma_{A}, and the Agent’s effort cost κ.\kappa. Indeed, the sign of KK^{*} is equal to the sign of γPγAγPκ1(κ1)2.\gamma_{P}\gamma_{A}-\gamma_{P}\kappa^{-1}-(\kappa^{-1})^{2}. Moreover, KtK^{*}_{t} varies monotonously in time, with KT=0.K^{*}_{T}=0.

Proof.

From the expression (4.7), we easily deduce that :

  • If c1+c2=0c_{1}+c_{2}=0 then Kt=0K^{*}_{t}=0 for every t[0,T]t\in[0,T],

  • if c1+c2>0c_{1}+c_{2}>0 then Kt>0K^{*}_{t}>0 for tτt\leq\tau and the function tKtt\to K^{*}_{t} decreases,

  • if c1+c2<0c_{1}+c_{2}<0 then Kt<0K^{*}_{t}<0 for tτt\leq\tau and the function tKtt\to K^{*}_{t} increases.

Replacing c1c_{1} and c2c_{2} by their relevant expressions in each case leads to the result.

Finally, we will show that the risk-sharing component of the contract is in fact linear with respect to the default time. This is a strong result of our study for which we had no ex-ante intuition.

Corollary 4.1.

The shutdown risk-sharing component of the optimal wage is linear in the default time. More precisely, the optimal wage is

W=yPC+ZBTτ+12(γA+1κ)(Z)2(Tτ)+K0(c1γP+γA+λγA)(Tτ).W^{*}=y_{PC}+Z^{*}B^{*}_{T\wedge\tau}+\frac{1}{2}\left(\gamma_{A}+\frac{1}{\kappa}\right)(Z^{*})^{2}(T\wedge\tau)+K^{*}_{0}-\left(\frac{c_{1}}{\gamma_{P}+\gamma_{A}}+\frac{\lambda}{\gamma_{A}}\right)(T\wedge\tau).
Proof.

Because KT=0K^{*}_{T}=0, the optimal wage can be written

W=yPC+ZBTτ+12(γA+1κ)(Z)2(Tτ)+f(Tτ),W^{*}=y_{PC}+Z^{*}B^{*}_{T\wedge\tau}+\frac{1}{2}\left(\gamma_{A}+\frac{1}{\kappa}\right)(Z^{*})^{2}(T\wedge\tau)+f(T\wedge\tau),

with

f(t)=Kt+0tλγA(exp(γAKs)1)𝑑s,t[0,T].f(t)=K^{*}_{t}+\int_{0}^{t}\frac{\lambda}{\gamma_{A}}\left(\exp(-\gamma_{A}K^{*}_{s})-1\right)ds,t\in[0,T].

Let us define

g(t)=c1+c2c1exp(c1γAγP+γA(Tt))c2c1.g(t)=\frac{c_{1}+c_{2}}{c_{1}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\frac{c_{2}}{c_{1}}.

We have g(t)=(c1+c2)γAγP+γAexp(c1γAγP+γA(Tt)).g^{\prime}(t)=-(c_{1}+c_{2})\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right).

Therefore,

tKt\displaystyle\frac{\partial}{\partial t}K^{*}_{t} =1γAg(t)g(t)\displaystyle=\frac{1}{\gamma_{A}}\frac{g^{\prime}(t)}{g(t)}
=1γP+γA{(c1+c2)exp(c1γAγP+γA(Tt))c1+c2c1exp(c1γAγP+γA(Tt))c2c1}\displaystyle=\frac{1}{\gamma_{P}+\gamma_{A}}\left\{\dfrac{-(c_{1}+c_{2})\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)}{\frac{c_{1}+c_{2}}{c_{1}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\frac{c_{2}}{c_{1}}}\right\}
=c1γP+γA{(c1+c2)exp(c1γAγP+γA(Tt))(c1+c2)exp(c1γAγP+γA(Tt))c2}.\displaystyle=\frac{c_{1}}{\gamma_{P}+\gamma_{A}}\left\{\dfrac{-(c_{1}+c_{2})\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)}{(c_{1}+c_{2})\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-c_{2}}\right\}.

Also

t0tλγA(exp(γAKs)1)𝑑s=λγA(exp(γAKt)1)\frac{\partial}{\partial t}\int_{0}^{t}\frac{\lambda}{\gamma_{A}}\left(\exp(-\gamma_{A}K^{*}_{s})-1\right)ds=\frac{\lambda}{\gamma_{A}}\left(\exp(-\gamma_{A}K^{*}_{t})-1\right)

and so :

t0tλγA(exp(γAKs)1)𝑑s\displaystyle\frac{\partial}{\partial t}\int_{0}^{t}\frac{\lambda}{\gamma_{A}}\left(\exp(-\gamma_{A}K^{*}_{s})-1\right)ds =λγA(exp(γAKt)1)\displaystyle=\frac{\lambda}{\gamma_{A}}\left(\exp(-\gamma_{A}K^{*}_{t})-1\right)
=λγA(1g(t)1)well-defined as g(t)>0 on [0,T]\displaystyle=\frac{\lambda}{\gamma_{A}}\left(\frac{1}{g(t)}-1\right)\quad\text{well-defined as $g(t)>0$ on $[0,T]$}
=λγA{1c1+c2c1exp(c1γAγP+γA(Tt))c2c11}\displaystyle=\frac{\lambda}{\gamma_{A}}\left\{\frac{1}{\frac{c_{1}+c_{2}}{c_{1}}\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\frac{c_{2}}{c_{1}}}-1\right\}
=c1λγA{1(c1+c2)exp(c1γAγP+γA(Tt))c2}λγA\displaystyle=\frac{c_{1}\lambda}{\gamma_{A}}\left\{\frac{1}{(c_{1}+c_{2})\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-c_{2}}\right\}-\frac{\lambda}{\gamma_{A}}

Finally, we have :

f(t)\displaystyle f^{\prime}(t) =tKt+t0tλγA(exp(γAKs)1)𝑑s\displaystyle=\frac{\partial}{\partial t}K^{*}_{t}+\frac{\partial}{\partial t}\int_{0}^{t}\frac{\lambda}{\gamma_{A}}\left(\exp(-\gamma_{A}K^{*}_{s})-1\right)ds
=c1γP+γA{(c1+c2)exp(c1γAγP+γA(Tt))(c1+c2)exp(c1γAγP+γA(Tt))c2}\displaystyle=\frac{c_{1}}{\gamma_{P}+\gamma_{A}}\left\{\dfrac{-(c_{1}+c_{2})\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)}{(c_{1}+c_{2})\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-c_{2}}\right\}
+c1λγA{1(c1+c2)exp(c1γAγP+γA(Tt))c2}λγA\displaystyle+\frac{c_{1}\lambda}{\gamma_{A}}\left\{\frac{1}{(c_{1}+c_{2})\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-c_{2}}\right\}-\frac{\lambda}{\gamma_{A}}
=c1γP+γA1(c1+c2)exp(c1γAγP+γA(Tt))c2{(c1+c2)exp(c1γAγP+γA(Tt))λγAγP+γA}\displaystyle=\frac{-c_{1}}{\gamma_{P}+\gamma_{A}}\frac{1}{(c_{1}+c_{2})\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-c_{2}}\left\{(c_{1}+c_{2})\exp\left(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}(T-t)\right)-\lambda\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}\right\}
λγA\displaystyle-\frac{\lambda}{\gamma_{A}}
=c1γP+γAλγAas c2=λγAγP+γA\displaystyle=-\frac{c_{1}}{\gamma_{P}+\gamma_{A}}-\frac{\lambda}{\gamma_{A}}\quad\text{as $c_{2}=\frac{\lambda\gamma_{A}}{\gamma_{P}+\gamma_{A}}$}
=(c1γP+γA+λγA).\displaystyle=-\left(\frac{c_{1}}{\gamma_{P}+\gamma_{A}}+\frac{\lambda}{\gamma_{A}}\right).

We may note that the default related part of the wage paid under full-Risk-Sharing also writes under this form, where we take the relevant values for K0,c1K_{0}^{*},c_{1} and c2c_{2}.

Remark 4.3.

A little algebra gives

f(t)=K0(c1γP+γA+λγA)t=K0(c1+c2)tγP+γAf(t)=K^{*}_{0}-\left(\frac{c_{1}}{\gamma_{P}+\gamma_{A}}+\frac{\lambda}{\gamma_{A}}\right)t=K^{*}_{0}-\frac{\left(c_{1}+c_{2}\right)t}{\gamma_{P}+\gamma_{A}}

The slope of ff is of opposite sign to the sign of c1+c2c_{1}+c_{2}. As a consequence,

  • If c1+c2=0c_{1}+c_{2}=0 then f(t)=0f(t)=0 for every t[0,T]t\in[0,T],

  • if c1+c2>0c_{1}+c_{2}>0 then f(0)=K0>0f(0)=K^{*}_{0}>0 for tτt\leq\tau and the function tf(t)t\to f(t) decreases,

  • if c1+c2<0c_{1}+c_{2}<0 then f(0)=K0<0f(0)=K^{*}_{0}<0 for tτt\leq\tau and the function tf(t)t\to f(t) increases.

K0>0K_{0}^{*}>0 is the extra compensation asked at the signature of the contract by an agent who is more sensitive to the shutdown risk than the principal. We can easily visualize the sign of the control K0K^{*}_{0} as a function of γP,\gamma_{P}, γA\gamma_{A} and κ\kappa. Below we plot the sign depending on the risk-aversions and fixing κ=1\kappa=1 and κ=2\kappa=2. The first two plots (Figure 9 and 2) correspond to the full Risk-Sharing case whilst Figures 3 and 4 show the sign of KK^{*} in the Moral Hazard case. The xx-axis holds the values of γA\gamma_{A} and the yy-axis the values of γP\gamma_{P}. Both risk-aversion constants are valued between 0 and 1010 and the origin is in the bottom left corner. Blue encodes a negative sign and green a positive sign.

Refer to caption
Figure 1: Sign of K0K_{0}^{*} depending on γP\gamma_{P} and γA\gamma_{A} for κ=1\kappa=1.
Refer to caption
Figure 2: Sign of K0K_{0}^{*} depending on γP\gamma_{P} and γA\gamma_{A} for κ=2\kappa=2.
Refer to caption
Figure 3: Sign of K0K_{0}^{*} depending on γP\gamma_{P} and γA\gamma_{A} for κ=1\kappa=1.
Refer to caption
Figure 4: Sign of K0K_{0}^{*} depending on γP\gamma_{P} and γA\gamma_{A} for κ=2\kappa=2.

We observe that in most situations, the sign of K0K^{*}_{0} is positive. A negative sign occurs when either the principal or the agent are close to being risk-neutral (symmetrically so in the full Risk-Sharing case but asymmetrically so in the Moral Hazard case : the sign switches from negative to positive at a much lower level of risk-aversion for the principal than the agent). Also note that increasing the agent’s effort coefficient κ\kappa decreases the level of risk-aversion for which K0K^{*}_{0} goes from positive to negative. Note that it is known by both the Principal and the Agent at time 0 whether the contract will fall into either regime.

Our key result shows that the risk of shutdown adds an extra linear term to the optimal compensation that, up to the underlying constants, has the same structure under both full Risk Sharing and under Moral Hazard. Even if the optimal contract separates the role of incentives from that of shutdown risk sharing, the amount of the insurance deposit K0K^{*}_{0} depends strongly on the magnitude of the moral hazard problem. Therefore, we may naturally quantify the effect of Moral Hazard on the risk-sharing part. With this question in mind we compare the values of K0K^{*}_{0} under Risk-Sharing and Moral Hazard for different parameter values. This may be observed in the Figures 5 to 9 below where we represent the values of K0K^{*}_{0} under Moral Hazard (red) and Risk-Sharing (blue) as a function of one of the underlying parameters (γP,γA,λ,κ,T\gamma_{P},\gamma_{A},\lambda,\kappa,T) whilst fixing the remaining 4.

These figures lead to a crucial observation : under any given set of parameters, the value of K0K^{*}_{0} under Moral Hazard is always greater than the value of K0K^{*}_{0} under Risk Sharing. Note that in some settings, given some parameters the signs of the two cases are not always the same as the sign is that of c1+c2c_{1}+c_{2} which depends on the values of γP\gamma_{P}, γA\gamma_{A} and κ\kappa in a different manner in Risk-Sharing and Moral Hazard. Of course though, as soon as K0K_{0}^{*} for Risk-Sharing is positive, K0K_{0}^{*} for Moral Hazard is positive too. We may note that although λ\lambda and TT do not impact the sign, variations in their values have an impact on the magnitude |K0||K_{0}^{*}| : the higher the risk of early termination of the contract, the lower the amount of insurance requested by the agent and the longer the duration of the contract, the higher the amount of insurance requested by the agent.

Figure 5: Values of K0K_{0}^{*} depending on γP\gamma_{P}
Refer to caption

γA=0.5,λ=1,κ=1,T=1\quad\gamma_{A}=0.5,\lambda=1,\kappa=1,T=1

Refer to caption

γA=1.5,λ=1,κ=1,T=1\quad\gamma_{A}=1.5,\lambda=1,\kappa=1,T=1

Refer to caption

γA=5,λ=1,κ=1,T=1\quad\gamma_{A}=5,\lambda=1,\kappa=1,T=1

Figure 6: Values of K0K_{0}^{*} depending on γA\gamma_{A}
Refer to caption

γP=0.5,λ=1,κ=1,T=1\quad\gamma_{P}=0.5,\lambda=1,\kappa=1,T=1

Refer to caption

γP=1,λ=1,κ=1,T=1\quad\gamma_{P}=1,\lambda=1,\kappa=1,T=1

Refer to caption

γP=3,λ=1,κ=1,T=1\quad\gamma_{P}=3,\lambda=1,\kappa=1,T=1

Figure 7: Values of K0K_{0}^{*} depending on λ\lambda
Refer to caption

γP=4,γA=4,κ=1,T=1\quad\gamma_{P}=4,\gamma_{A}=4,\kappa=1,T=1

Refer to caption

γP=1,γA=1,κ=1,T=1\quad\gamma_{P}=1,\gamma_{A}=1,\kappa=1,T=1

Figure 8: Values of K0K_{0}^{*} depending on κ\kappa
Refer to caption

γP=1,γA=1,λ=1,T=1\quad\gamma_{P}=1,\gamma_{A}=1,\lambda=1,T=1

Refer to caption

γP=1,γA=3,λ=1,T=1\quad\gamma_{P}=1,\gamma_{A}=3,\lambda=1,T=1

Refer to caption

γP=3,γA=3,λ=1,T=1\quad\gamma_{P}=3,\gamma_{A}=3,\lambda=1,T=1

Figure 9: Values of K0K_{0}^{*} depending on TT
Refer to caption

γP=1,γA=1,λ=1,κ=1\quad\gamma_{P}=1,\gamma_{A}=1,\lambda=1,\kappa=1

Refer to caption

γP=1,γA=3,λ=1,κ=1\quad\gamma_{P}=1,\gamma_{A}=3,\lambda=1,\kappa=1

Refer to caption

γP=3,γA=3,λ=1,κ=1\quad\gamma_{P}=3,\gamma_{A}=3,\lambda=1,\kappa=1

Additionally, the parameter λ\lambda does have a quantifiable effect of the wage as it affects its expected value :

𝔼(f(Tτ))=K0c1+c2γP+γA(1eλTT).\mathbb{E}(f(T\wedge\tau))=K_{0}^{*}-\frac{c_{1}+c_{2}}{\gamma_{P}+\gamma_{A}}\left(\frac{1-e^{-\lambda T}}{T}\right).

This expected value is represented below as a function of γA\gamma_{A} and γP\gamma_{P} for different values of λ\lambda (with T=1T=1 and κ=1\kappa=1 fixed). The first three figures (Figures 10, 11 and 12 ) concern the full Risk-Sharing case whilst the second set of figures (Figures 13, 14 and 15) concern the Moral Hazard setting.

Refer to caption
Figure 10: Expected value depending on γP\gamma_{P} and γA\gamma_{A} for λ=0.5\lambda=0.5.
Refer to caption
Figure 11: Expected value depending on γP\gamma_{P} and γA\gamma_{A} for λ=1\lambda=1.
Refer to caption
Figure 12: Expected value depending on γP\gamma_{P} and γA\gamma_{A} for λ=5\lambda=5.
Refer to caption
Figure 13: Expected value depending on γP\gamma_{P} and γA\gamma_{A} for λ=0.5\lambda=0.5.
Refer to caption
Figure 14: Expected value depending on γP\gamma_{P} and γA\gamma_{A} for λ=1\lambda=1.
Refer to caption
Figure 15: Expected value depending on γP\gamma_{P} and γA\gamma_{A} for λ=5\lambda=5.

The value seems to be in many cases very close to 0. As such, the agent earns, on average, a very similar wage to a "stopped" Holmstrom-Milgrom wage. However when the principal and the agent are particularly risk-averse, the expected value increases quite notably and the agent gains slightly more on average. This is in line with the papers by Hoffman and Pfeil [11] and Bertrand and Mullainathan [2] which show the agent must be rewarded for a risk that is beyond his control. Note that the simulations show that for fixed levels of risk-aversion the expected value increases as λ\lambda increases. For example for γP=γA=7\gamma_{P}=\gamma_{A}=7, in Figure 13 the expected value is approximately worth 0.0750.075 and in Figure 14 it is approximately equal to 0.150.15. As such the risk-averse agent is increasingly rewarded for the uncontrollable risk.

Finally, we can make a few further comments related to the underlying expected utilities in this new contracting setting. First, the agent’s expected utility is the same under both full Risk-Sharing and Moral Hazard. Indeed he walks away with his participation constraint :

𝔼[UA(W0Tκ(as)𝑑s)]=UA(yPC),\mathbb{E}\left[U_{A}\left(W^{*}-\int_{0}^{T}\kappa(a^{*}_{s})ds\right)\right]=U_{A}(y_{PC}),

where (W,a)(W^{*},a^{*}) designates the Risk-Sharing or Moral Hazard optimal contract under shutdown risk. Such a result also holds in the same setting without shutdown risk : the agent tracts average the same expected utility under full Risk-Sharing or Moral Hazard, with or without an underlying agency-free external risk.
When it comes to the principal we may wonder how his expected utility may be affected by the possibility of a halt in production. In particular we may question what the principal loses in not being able to observe the agent’s action under a likelihood of agency-free external risk ? To answer this we denote as V0RS(0,x,y)V_{0}^{RS}(0,x,y) the Principal’s expected utility under full Risk-Sharing and V0MH(0,x,y)V_{0}^{MH}(0,x,y) under Moral Hazard. We have that :

V0RS(0,x,y)=UP(xy)Φ0RS(0)andV0MH(0,x,y)=UP(xy)Φ0MH(0)V_{0}^{RS}(0,x,y)=U_{P}(x-y)\Phi_{0}^{RS}(0)\quad\text{and}\quad V_{0}^{MH}(0,x,y)=U_{P}(x-y)\Phi_{0}^{MH}(0)

where Φ0RS\Phi_{0}^{RS} and Φ0MH\Phi_{0}^{MH} are the related functions from Theorem 3.1 and Theorem 4.1. With these expressions we have that :

V0MH(0,x,y)V0RS(0,x,y)=Φ0MH(0)Φ0RS(0).\dfrac{V_{0}^{MH}(0,x,y)}{V_{0}^{RS}(0,x,y)}=\frac{\Phi_{0}^{MH}(0)}{\Phi_{0}^{RS}(0)}.

The Moral Hazard problem involves optimizing across a more restricted set of contracts. Therefore we know that :

V0MH(0,x,y)V0RS(0,x,y),V_{0}^{MH}(0,x,y)\leq V_{0}^{RS}(0,x,y),

and as UP(xy)<0U_{P}(x-y)<0 :

Φ0RS(0)Φ0MH(0).\Phi_{0}^{RS}(0)\leq\Phi_{0}^{MH}(0).

We may question whether this inequality leads to a big gap between the expected utilities and we answer this by plotting the ratio Φ0MH(0)Φ0RS(0)\frac{\Phi_{0}^{MH}(0)}{\Phi_{0}^{RS}(0)} for different values of λ\lambda and with T=1T=1 and κ=1\kappa=1 fixed.

Refer to caption
Figure 16: Ratio depending on γP\gamma_{P} and γA\gamma_{A} for λ=0.5\lambda=0.5.
Refer to caption
Figure 17: Ratio depending on γP\gamma_{P} and γA\gamma_{A} for λ=1\lambda=1.
Refer to caption
Figure 18: Ratio depending on γP\gamma_{P} and γA\gamma_{A} for λ=5\lambda=5.

We first observe a standard result : for low levels of risk-aversion the ratio is close to 1 and the principal does not lose much by not observing the agent’s actions. As the values of risk-aversion increase the principal loses out more and more by not being in a first best setting. This classical result comes with an observation that is specific to the presence of shutdown risk : as λ\lambda increases (and therefore as the chance of shutdown risk occurring before TT increases), the ratio stays close to 1 for higher and higher levels of risk-aversion. For example in Figure 16 when γP=γA=6\gamma_{P}=\gamma_{A}=6 we observe that Φ0MH(0)Φ0RS(0)2\frac{\Phi_{0}^{MH}(0)}{\Phi_{0}^{RS}(0)}\approx 2 yet in Figure 18 for the same levels of risk-aversion we have that Φ0MH(0)Φ0RS(0)1.\frac{\Phi_{0}^{MH}(0)}{\Phi_{0}^{RS}(0)}\approx 1. So a high possibility of some production halt occurring reduces the gap between the full Risk-Sharing contract and the Moral Hazard contract. Such a phenomena may be due to the fact that a high possibility of a production halt at some point in the time interval means that the wage process evolves on a time period that is on average shorter before stopping. There is thus less time for a significant gap to appear between the full Risk-Sharing case and the Moral Hazard case.

As this analysis comes to a close we finish this section by discussing a possibility for extension with more general deterministic compensators.

General deterministic compensators (Λt)t[0,T](\Lambda_{t})_{t\in[0,T]}

Throughout this paper, we have considered a constant compensator λ\lambda for the jump process. This choice allows for clearer calculations but it is key to note that our results extend to the case where λ\lambda is no longer constant such as :

Λt=0tλs𝑑s,\Lambda_{t}=\int_{0}^{t}\lambda_{s}ds,

with (λs)s[0,T](\lambda_{s})_{s\in[0,T]} some deterministic positive mapping such that ΛT<+\Lambda_{T}<+\infty. The proofs for the optimal contracting are simply a direct extension of the proofs of the previous sections. Of course due to the independence between BB and NN, a time dependent compensator does not induce any change to the Holmstrom-Milgrom part of the wages. Only the part related to KK^{*} is affected. We provide the details in the following.

The full Risk-Sharing problem

The optimal wage for the Risk-Sharing problem in such a setting is of the form :

Wt\displaystyle W_{t} =y+0tZs(1Ns)𝑑Bs+0tKs(1Ns)𝑑Ms\displaystyle=y^{*}+\int_{0}^{t}Z^{*}_{s}(1-N_{s})dB_{s}+\int_{0}^{t}K^{*}_{s}(1-N_{s})dM_{s}
+0t{γA2Zs2+κ(as(1Ns))+λsγA[exp(γAKs)1+γAKs]}(1Ns)𝑑s,\displaystyle+\int_{0}^{t}\left\{\frac{\gamma_{A}}{2}{Z_{s}^{*}}^{2}+\kappa(a^{*}_{s}(1-N_{s}))+\frac{\lambda_{s}}{\gamma_{A}}[\exp(-\gamma_{A}K^{*}_{s})-1+\gamma_{A}K^{*}_{s}]\right\}(1-N_{s})ds,

where :

y=yPC,at=1κ,Zt=γPγP+γAandKt=1γP+γAlog(Φ0(t)),y^{*}=y_{PC},a^{*}_{t}=\frac{1}{\kappa},Z^{*}_{t}=\frac{\gamma_{P}}{\gamma_{P}+\gamma_{A}}\quad\text{and}\quad K^{*}_{t}=\frac{1}{\gamma_{P}+\gamma_{A}}\log(\Phi_{0}(t)),

with Φ0(t)\Phi_{0}(t) solution to the Bernouilli equation :

Φ0(t)+c1(t)Φ0(t)+c2(t)Φ0(t)γPγP+γA=0,Φ0(T)=1,\Phi^{\prime}_{0}(t)+c_{1}(t)\Phi_{0}(t)+c_{2}(t)\Phi_{0}(t)^{\frac{\gamma_{P}}{\gamma_{P}+\gamma_{A}}}=0,\quad\Phi_{0}(T)=1,

where

c1(t)=γP2γA2(γP+γA)γP2κλtγP+γAγAandc2(t)=λtγP+γAγA.c_{1}(t)=\frac{\gamma_{P}^{2}\gamma_{A}}{2(\gamma_{P}+\gamma_{A})}-\frac{\gamma_{P}}{2\kappa}-\lambda_{t}\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\quad\text{and}\quad c_{2}(t)=\lambda_{t}\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}.

The Moral Hazard problem

The optimal wage in the Moral-Hazard problem is again of the form :

Wt=y+0tZs(1Ns)𝑑Bs+0tKs(1Ns)𝑑Ms+0t{12γAZs2Zs22κ+λsKs+λsγA(eγAKs1)}(1Ns)𝑑s,W_{t}=y^{*}+\int_{0}^{t}Z_{s}^{*}(1-N_{s})dB^{*}_{s}+\int_{0}^{t}K_{s}^{*}(1-N_{s})dM_{s}+\int_{0}^{t}\left\{\frac{1}{2}\gamma_{A}{Z_{s}^{*}}^{2}-\frac{{Z_{s}^{*}}^{2}}{2\kappa}+\lambda_{s}K_{s}^{*}+\frac{\lambda_{s}}{\gamma_{A}}(e^{-\gamma_{A}K_{s}^{*}}-1)\right\}(1-N_{s})ds,

where :

y=yPC,Zt=γP+κ1γP+γA+κ1andKt=1γP+γAlog(Φ0(t)),y^{*}=y_{PC},Z^{*}_{t}=\frac{\gamma_{P}+\kappa^{-1}}{\gamma_{P}+\gamma_{A}+\kappa^{-1}}\quad\text{and}\quad K^{*}_{t}=\frac{1}{\gamma_{P}+\gamma_{A}}\log(\Phi_{0}(t)),

with Φ0(t)\Phi_{0}(t) solution to the Bernouilli equation :

Φ0(t)+c1(t)Φ0(t)+c2(t)Φ0(t)γPγP+γA=0,Φ0(T)=1,\Phi^{\prime}_{0}(t)+c_{1}(t)\Phi_{0}(t)+c_{2}(t)\Phi_{0}(t)^{\frac{\gamma_{P}}{\gamma_{P}+\gamma_{A}}}=0,\quad\Phi_{0}(T)=1,

with :

c1(t)=γP2γA2(γP+γA+κ1)γPκ1(γP+κ1)2(γP+γA+κ1)λtγP+γAγAandc2(t)=λtγP+γAγA.c_{1}(t)=\frac{\gamma_{P}^{2}\gamma_{A}}{2(\gamma_{P}+\gamma_{A}+\kappa^{-1})}-\frac{\gamma_{P}\kappa^{-1}(\gamma_{P}+\kappa^{-1})}{2{(\gamma_{P}+\gamma_{A}+\kappa^{-1})}}-\lambda_{t}\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\quad\text{and}\quad c_{2}(t)=\lambda_{t}\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}.

5  Mitigating the effects of agency-free external risk

This paper has so far modeled the occurence of a halt as a complete fatality suffered by both parties in the contracting agreement. Yet the recent crisis has highlighted the ability of humans and businesses to react and adapt when faced with adversity. We now include such phenomena in the contracting setting by allowing the principal to invest upon a halt in order to continue some form of (possibly disrupted) production. This is quite a natural and realistic variant on our initial model. Indeed when faced with a period of lockdown, companies may for example invest in teleworking infrastructure so that a number of employees whose jobs are doable remotely can continue to work. Similarly, jobs that require some form of presence could continue if companies invest in protective equipment and adapt their organization. We may wonder how such a mechanism may affect optimal contracting.

Setting for mitigation

Mathematically, we consider that the production process evolves as previously up until τT\tau\wedge T. If a halt happens at some time τT\tau\leq T, we allow the principal to invest an amount i>0i>0 to continue production at a degraded level θ(0,1)\theta\in(0,1). It is assumed that the investment decision is at the principal’s convenience. It is modeled by a control DD which is a 𝔾τ{\mathbb{G}}_{\tau}-measurable random variable with values in {0,1}\{0,1\}. The θ\theta parameter is firm-specific and reflects the effectiveness of the post-shutdown reorganization.
Under the initial probability measure 0\mathbb{P}^{0}, the output process XtX_{t} evolves as

Xt=x0+0t((1Ns)+Ns11D=1)𝑑Bs.X_{t}=x_{0}+\int_{0}^{t}((1-N_{s})+N_{s}1\!\!1_{D=1})\,dB_{s}.

We recall from [10] Lemma 4.4. the decomposition of a 𝔾\mathbb{G}-adapted process ϕ\phi. There exist a 𝔽\mathbb{F}-adapted process ϕ0\phi^{0} and a family of processes (ϕt1(u),utT)(\phi^{1}_{t}(u),u\leq t\leq T) that are 𝔽t(+)\mathbb{F}_{t}\otimes{\cal B}(\mathbb{R}_{+}) measurable such that

ϕt=ϕt011t<τ+ϕt1(τ)11tτ.\phi_{t}=\phi_{t}^{0}1\!\!1_{t<\tau}+\phi^{1}_{t}(\tau)1\!\!1_{t\geq\tau}.

Contract: A contract WW is a 𝔾T\mathbb{G}_{T}-measurable random variable satisfying 𝔼(exp(2γAW))<+\mathbb{E}(\exp(-2\gamma_{A}W))<+\infty of the form

W=W011T<τ+(WT1,1(τ)11D=1+WT1,0(τ)11D=0)11τT.W=W^{0}1\!\!1_{T<\tau}+(W_{T}^{1,1}(\tau)1\!\!1_{D=1}+W^{1,0}_{T}(\tau)1\!\!1_{D=0})1\!\!1_{\tau\leq T}.

where W0W^{0} is 𝔽T\mathbb{F}_{T}-mesurable and Wt1,1(u)W_{t}^{1,1}(u) and Wt1,0(u)W_{t}^{1,0}(u) are 𝔽t(+)\mathbb{F}_{t}\otimes{\cal B}(\mathbb{R}_{+}) measurable. We will assume that WT1,0(τ)=WTτ1,0(τ)W_{T}^{1,0}(\tau)=W_{T\wedge\tau}^{1,0}(\tau) since in the absence of investment, it is no longer necessary to give incentives after τ\tau.

Effort process: In this setting, the agent will adapt his effort to the occurence of the shutdown risk. This is mathematically modeled by a 𝔾\mathbb{G}-adapted process (at)t(a_{t})_{t} in the form

at=at011t<τ+at1(τ)11tτa_{t}=a_{t}^{0}1\!\!1_{t<\tau}+a^{1}_{t}(\tau)1\!\!1_{t\geq\tau}

where a0a^{0} and a1a^{1} are respectively 𝔽\mathbb{F}-adapted and 𝔽t(+)\mathbb{F}_{t}\otimes{\cal B}(\mathbb{R}_{+}) measurable. Furthermore, we assume that the effort processes are bounded by some constant AA. We then define a\mathbb{P}^{a} as dad0|𝒢T=LTθ\frac{d\mathbb{P}^{a}}{d\mathbb{P}^{0}}|\mathcal{G}_{T}=L_{T}^{\theta}, with

LTθ=exp(0Tas0(1Ns)+θas1(τ)Ns11D=1dBs120T(as0)2(1Ns)+θ2(as1(τ))2Ns11D=1ds)L_{T}^{\theta}=\exp\left(\int_{0}^{T}a_{s}^{0}(1-N_{s})+\theta a^{1}_{s}(\tau)N_{s}1\!\!1_{D=1}dB_{s}-\frac{1}{2}\int_{0}^{T}(a_{s}^{0})^{2}(1-N_{s})+\theta^{2}(a^{1}_{s}(\tau))^{2}N_{s}1\!\!1_{D=1}ds\right)

Because the processes a0a^{0} and a1a^{1} are bounded, (Bta)t[0,T](B^{a}_{t})_{t\in[0,T]} with

Bta=Bt0t(as0(1Ns)+θas1(τ)Ns11D=1)𝑑s,t[0,T]B^{a}_{t}=B_{t}-\int_{0}^{t}(a_{s}^{0}(1-N_{s})+\theta a^{1}_{s}(\tau)N_{s}1\!\!1_{D=1})\,ds,t\in[0,T]

is a 𝔾\mathbb{G}-Brownian motion under a\mathbb{P}^{a}. Under a\mathbb{P}^{a}, the output process evolves as

Xt=x0+0t(as0(1Ns)+θas1(τ)Ns11D=1)𝑑s+0t((1Ns)+Ns11D=1)𝑑Bs.X_{t}=x_{0}+\int_{0}^{t}(a_{s}^{0}(1-N_{s})+\theta a^{1}_{s}(\tau)N_{s}1\!\!1_{D=1})\,ds+\int_{0}^{t}((1-N_{s})+N_{s}1\!\!1_{D=1})\,dB_{s}.

The Optimal contract

We first make the following observation. After τ\tau, if the default time occurs before the maturity of the contract, the principal has a binary decision to take. If she decides to not invest, she gets the value V1,0(x,y)=UP(xy)V^{1,0}(x,y)=U_{P}(x-y) where xx is the level of input and yy is the agent continuation value. On the other hand, if she decides to invest, she will face for tτt\geq\tau the moral hazard problem of Holmstrom and Milgrom for which we know the optimal contract and the associated value function

V1,1(t,x,y)=UP(xy)Φ1(t,θ)V^{1,1}(t,x,y)=U_{P}(x-y)\Phi_{1}(t,\theta)

where Φ1(t,θ)=exp(γPCinv(Tt))\Phi_{1}(t,\theta)=\exp(-\gamma_{P}C_{inv}(T-t)) and Cinv:=(γP+θ2κ)22(γP+γA+θ2κ)γP2.C_{inv}:=\dfrac{\left(\gamma_{P}+\frac{\theta^{2}}{\kappa}\right)^{2}}{2\left(\gamma_{P}+\gamma_{A}+\frac{\theta^{2}}{\kappa}\right)}-\frac{\gamma_{P}}{2}.
Because the principal has to pay a sunk cost i>0i>0 to invest, she will decide optimally to invest if and only if at τ\tau for a given (x,y)(x,y), she observes

V1,1(τ,x,y)V1,0(x,y),V^{1,1}(\tau,x,y)\geq V^{1,0}(x,y),

or equivalently Cinv(Tτ)>iC_{inv}(T-\tau)>i. Hence, if Cinv>0C_{inv}>0, the optimal control will be D=11{τ<TiCinv}.D^{*}=1\!\!1_{\{\tau<T-\frac{i}{C_{inv}}\}}. To sum up, we have

Lemma 5.1.
  1. 1.

    Investment for mitigation is never optimal upon a halt if :

    Cinv<0ori>TCinv.C_{inv}<0\quad\text{or}\quad i>TC_{inv}.
  2. 2.

    Now suppose that :

    Cinv>0andi<TCinv.C_{inv}>0\quad\text{and}\quad i<TC_{inv}.

    Mitigation is optimal up until the cutoff time tmaxt_{max} defined as :

    tmax:=TiCinv.t_{max}:=T-\frac{i}{C_{inv}}.

    Note that i<TCinvi<TC_{inv} guarantees that tmax0.t_{max}\geq 0.

We are in a position to solve the before-default principal problem. Proceeding analogously as in Section 4, the before-default value function is given by the Markovian control problem

VP=supyyPCV0(0,x,y),V_{P}=\sup_{y\geq y_{PC}}V_{0}(0,x,y),

with

V(0,x0,y0)=supπ=(Z,K)ζ𝔼[UP(XTπWTπ)(1NT)+0Tmax(V1,0(Xtπ,Wtπ),V1,1(t,Xtπ,Wtπ))λeλt𝑑t],V(0,x_{0},y_{0})=\sup_{\pi=(Z,K)\in\zeta}\mathbb{E}\left[U_{P}(X_{T}^{\pi}-W_{T}^{\pi})(1-N_{T})+\int_{0}^{T}\max(V^{1,0}(X_{t}^{\pi},W_{t}^{\pi}),V^{1,1}(t,X_{t}^{\pi},W_{t}^{\pi}))\lambda e^{-\lambda t}\,dt\right], (5.1)

and

dXt=a(Zt)(1Nt)dt+(1Nt)dBt,dX_{t}=a^{*}(Z_{t})(1-N_{t})\,dt+(1-N_{t})\,dB_{t}^{*},
dWtπ\displaystyle dW^{\pi}_{t} =Zs(1Ns)dBs+Ks(1Ns)dMs\displaystyle=Z_{s}(1-N_{s})dB_{s}^{*}+K_{s}(1-N_{s})dM_{s} (5.2)
+{γA2Zs2+κ(a(Zs))+λγA[exp(γAKs)1+γAKs]}(1Ns)ds.\displaystyle+\left\{\frac{\gamma_{A}}{2}Z_{s}^{2}+\kappa(a^{*}(Z_{s}))+\frac{\lambda}{\gamma_{A}}[\exp(-\gamma_{A}K_{s})-1+\gamma_{A}K_{s}]\right\}(1-N_{s})ds.
Remark 5.1.

To be perfectly complete, we develop in the appendix the martingale optimality principle which makes it possible to obtain the dynamics (5.2).

Theorem 5.1.

We have the following explicit characterizations of the optimal contracts. Assume the constant AA in the definition of the set of admissible efforts {\cal B} satisfies

A>γP+κ1κ(γP+γA)+1.A>\dfrac{\gamma_{P}+\kappa^{-1}}{\kappa(\gamma_{P}+\gamma_{A})+1}.

Let Zt=γP+κ1γP+γA+κ1Z^{*}_{t}=\dfrac{\gamma_{P}+\kappa^{-1}}{\gamma_{P}+\gamma_{A}+\kappa^{-1}} and

Kt=1γP+γAlog(Φ0(t)min{1,exp(γPi)Φ1(t,θ)})K^{*}_{t}=\frac{1}{\gamma_{P}+\gamma_{A}}\log\left(\frac{\Phi_{0}(t)}{\min\Big{\{}1,\exp(\gamma_{P}i)\Phi_{1}(t,\theta)\Big{\}}}\right)

with Φ0\Phi_{0} as defined above with :

c1:=γP2γA2(γP+γA+κ1)γPκ1(γP+κ1)2(γP+γA+κ1)λγP+γAγAc_{1}:=\frac{\gamma_{P}^{2}\gamma_{A}}{2(\gamma_{P}+\gamma_{A}+\kappa^{-1})}-\frac{\gamma_{P}\kappa^{-1}(\gamma_{P}+\kappa^{-1})}{2{(\gamma_{P}+\gamma_{A}+\kappa^{-1})}}-\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}

and

c2(t)=λγP+γAγAmin{1,exp(γPi)Φ1(t,θ)}γAγP+γAc_{2}(t)=\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\min\Big{\{}1,\exp(\gamma_{P}i)\Phi_{1}(t,\theta)\Big{\}}^{\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}}

Then (yPC,Z,K)(y_{PC},Z^{*},K^{*}) parametrizes the optimal wage for the Moral Hazard problem with a possibility for mitigation. The Agent performs the optimal action Zκ\dfrac{Z^{*}}{\kappa} before τ\tau and θZκ\dfrac{\theta Z^{*}}{\kappa} after τ\tau when τ<Ti/Cinv\tau<T-i/C_{inv}.

Proof.

The reasoning used to compute the optimal Moral Hazard contract very much parallels the reasoning used above in Section 4. As a consequence, we are much more brief in the following proof.
The Hamilton-Jacobi-Bellman equation associated to the value function V0V_{0} is the following :

0=tv0(t,x,y)+infZ,K{xv0(t,x,y)Zκ+yv0(t,x,y)[γA2Z2+Z22κ+λγA[exp(γAK)1]]\displaystyle 0=\partial_{t}v_{0}(t,x,y)+\inf_{Z,K}\left\{\partial_{x}v_{0}(t,x,y)\dfrac{Z}{\kappa}+\partial_{y}v_{0}(t,x,y)\left[\frac{\gamma_{A}}{2}Z^{2}+\frac{Z^{2}}{2\kappa}+\frac{\lambda}{\gamma_{A}}[\exp(-\gamma_{A}K)-1]\right]\right.
+λ[v1(t,x,y+K)v0(t,x,y)]+yyv0(t,x,y)Z22+12xxv0(t,x,y)+xyv0(t,x,y)Z},\displaystyle\left.+\lambda\left[v_{1}(t,x,y+K)-v_{0}(t,x,y)\right]+\partial_{yy}v_{0}(t,x,y)\frac{Z^{2}}{2}+\frac{1}{2}\partial_{xx}v_{0}(t,x,y)+\partial_{xy}v_{0}(t,x,y)Z\right\}, (5.3)

with the boundary condition :

v0(T,x,y)=UP(xy)v_{0}(T,x,y)=U_{P}(x-y)

where

v1(t,x,y)=UP(xy)min{1,exp(γPi)Φ1(t,θ)}.v_{1}(t,x,y)=U_{P}(x-y)\min\Big{\{}1,\exp(\gamma_{P}i)\Phi_{1}(t,\theta)\Big{\}}.
Lemma 5.2.

Assume A>γP+κ1γP+γA+κ1A>\dfrac{\gamma_{P}+\kappa^{-1}}{\gamma_{P}+\gamma_{A}+\kappa^{-1}}. The function v0(t,x,y)=UP(xy)Φ0(t),v_{0}(t,x,y)=U_{P}(x-y)\Phi_{0}(t), with

Φ0(t):=exp(c1t){exp(c1γAγP+γAT)+γAγP+γAtTc2(s)exp(γAγP+γAc1s)𝑑s}γP+γAA\Phi_{0}(t):=\exp(-c_{1}t)\left\{\exp(c_{1}\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}T)+\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}\int_{t}^{T}c_{2}(s)\exp(\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}c_{1}s)ds\right\}^{\frac{\gamma_{P}+\gamma_{A}}{A}}

where

c1=γP2γA2(γP+γA+κ1)γPκ1(γP+κ1)2(γP+γA+κ1)λγP+γAγAc_{1}=\frac{\gamma_{P}^{2}\gamma_{A}}{2(\gamma_{P}+\gamma_{A}+\kappa^{-1})}-\frac{\gamma_{P}\kappa^{-1}(\gamma_{P}+\kappa^{-1})}{2{(\gamma_{P}+\gamma_{A}+\kappa^{-1})}}-\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}

and

c2(t)=λγP+γAγAmin{1,exp(γPi)Φ1(t,θ)}γAγP+γA.c_{2}(t)=\lambda\frac{\gamma_{P}+\gamma_{A}}{\gamma_{A}}\min\Big{\{}1,\exp(\gamma_{P}i)\Phi_{1}(t,\theta)\Big{\}}^{\frac{\gamma_{A}}{\gamma_{P}+\gamma_{A}}}.

solves in the classical sense the HJB equation (5.2). In particular Zt=γP+κ1γP+γA+κ1Z^{*}_{t}=\dfrac{\gamma_{P}+\kappa^{-1}}{\gamma_{P}+\gamma_{A}+\kappa^{-1}} and Kt=1γP+γAlog(Φ0(t)min{1,exp(γPi)Φ1(t,θ)})K^{*}_{t}=\frac{1}{\gamma_{P}+\gamma_{A}}\log\left(\dfrac{\Phi_{0}(t)}{\min\Big{\{}1,\exp(\gamma_{P}i)\Phi_{1}(t,\theta)\Big{\}}}\right).

Proof.

The proof of this lemma is a direct adaptation of the proof of Lemma 4.2 to which we refer the reader. ∎

The proof of the final result relies on the regularity of v0v_{0} and a standard verification result. Because the controls are free of yy, we deduce that VP=V0(0,x,yPC).V_{P}=V_{0}(0,x,y_{PC}).

The main change brought about by investment involves the halt related control KK^{*}. Indeed the optimal ZZ^{*} in the Moral Hazard case are simply the optimal "Holmström-Milgrom" controls for the related production process. At first glance, the optimal control KK^{*} seems to be quite different from that of Theorem 4.1. However one may verify that when we are in a setting where investment is never optimal (through the criteria of Lemma 5.1), the expression for KK^{*} simplifies to exactly that of Theorem 4.1. The key to deduce this is that in such a setting, min{1,exp(γPi)Φ1(t,θ)}=1\min\Big{\{}1,\exp(\gamma_{P}i)\Phi_{1}(t,\theta)\Big{\}}=1. We may therefore focus our analysis on the effects on investment when investing may be optimal (i.e. when Cinv>0andi<TCinvC_{inv}>0\quad\text{and}\quad i<TC_{inv}). In such a setting, KK^{*} has two phases :

  • -

    before tmaxt_{max}, KK^{*} is adjusted to account for the possibility of risk mitigation

  • -

    after tmax,t_{max}, KK^{*} has the same values as without mitigation. Indeed :

    min{1,exp(γPi)Φ1(t,θ)}=1forttmax.\min\Big{\{}1,\exp(\gamma_{P}i)\Phi_{1}(t,\theta)\Big{\}}=1\quad\text{for}\quad t\geq t_{max}.

We are able to analyze the effect of different parameters and to do so represent the deterministic part of KK^{*} as a function of time in the following figures.

We fix parameters γP=κ=T=1,γA=0.5\gamma_{P}=\kappa=T=1,\gamma_{A}=0.5 : again this allows for mitigation to be optimal before some tmaxt_{max}.

Refer to caption
Figure 19: λ=0.5\lambda=0.5, i=0.1i=0.1, θ=0.9\theta=0.9
Refer to caption
Figure 20: λ=1\lambda=1, i=0.1i=0.1, θ=0.9\theta=0.9.
Refer to caption
Figure 21: λ=5\lambda=5, i=0.1i=0.1, θ=0.9\theta=0.9.
Refer to caption
Figure 22: i=0.05i=0.05, λ=1\lambda=1, θ=0.9\theta=0.9
Refer to caption
Figure 23: i=0.1i=0.1, λ=1\lambda=1, θ=0.9\theta=0.9 .
Refer to caption
Figure 24: i=0.15i=0.15, λ=1\lambda=1, θ=0.9\theta=0.9 .
Refer to caption
Figure 25: i=0.1i=0.1, λ=1\lambda=1, θ=0.85\theta=0.85
Refer to caption
Figure 26: i=0.1i=0.1, λ=1\lambda=1, θ=0.9\theta=0.9 .

We immediately observe that with mitigation, the value of KK^{*} before tmaxt_{max} and is higher than without mitigation : the possibility for mitigation shrinks the opportunities for speculation (see Figures 19 to 21) and increasingly so as the probability of a halt increases. In fact the sign of KK^{*} may now change over the duration of the contracting period : see Figure 21. Quite naturally, tmaxt_{max} varies with θ\theta and ii. Indeed it decreases as ii increases or θ\theta decreases : as the cost of investment increases and/or the level of degradation in continued production increases, more time is needed for investment for continued production to be worth it.

6  Appendix

We sketch the martingale optimality principle arising from the Agent’s problem in the investment setting. We set 𝔾2\mathbb{H}^{2}_{\mathbb{G}} is the set of 𝔾\mathbb{G}- adapted processes ZZ with 𝔼[0TZs2𝑑s]<+\mathbb{E}[\int_{0}^{T}Z_{s}^{2}ds]<+\infty and 𝕊𝔾2\mathbb{S}^{2}_{\mathbb{G}} is the set of 𝔾\mathbb{G}-predictable processes YY with cadlag paths such that ZZ with 𝔼[supt[0,T]Yt2]<+\mathbb{E}[\sup_{t\in[0,T]}Y^{2}_{t}]<+\infty. For a 𝔾τ\mathbb{G}_{\tau}-measurable random variable DD with values in {0,1}\left\{0,1\right\} that models the investment decision, we consider a contract WW as a 𝔾T\mathbb{G}_{T} measurable r.v. which can be decomposed under the form :

W=W01T<τ+(WT1,1(τ)1D=1+WT1,0(τ)1D=0)1τT,W=W^{0}\textbf{1}_{T<\tau}+\left(W^{1,1}_{T}(\tau)\textbf{1}_{D=1}+W^{1,0}_{T}(\tau)\textbf{1}_{D=0}\right)\textbf{1}_{\tau\leq T},

where W0W^{0} is 𝔽T\mathbb{F}_{T}-measurable, and Wt1,1(u)W_{t}^{1,1}(u) and Wt1,0(u)W_{t}^{1,0}(u) are 𝔽t(+)\mathbb{F}_{t}\otimes\mathcal{B}(\mathbb{R}_{+}) measurable, with in particular WT1,0(τ)=WTτ1,0(τ).W^{1,0}_{T}(\tau)=W^{1,0}_{T\wedge\tau}(\tau). Given a contract WW, the agent faces the following control problem,

supa𝔼a𝔼[UA(W0Tκ(as)𝑑s)].\sup_{a\in\mathcal{B}}\mathbb{E}^{\mathbb{P}^{a}}\mathbb{E}\left[U_{A}\left(W-\int_{0}^{T}\kappa(a_{s})ds\right)\right].

Remember that an effort process is now a 𝔾\mathbb{G}-adapted process (at)t[0,T](a_{t})_{t\in[0,T]} and consequently has the form :

at=at01t<τ+at1(τ)1tτa_{t}=a_{t}^{0}\textbf{1}_{t<\tau}+a_{t}^{1}(\tau)\textbf{1}_{t\geq\tau}

where a0a^{0} is 𝔽\mathbb{F}-adapted and a1a^{1} is 𝔽t(+)\mathbb{F}_{t}\otimes\mathcal{B}(\mathbb{R}_{+}) measurable and where both are assumed bounded by some constant AA. By convention, we still denote by \mathcal{B} the set of such effort.

Lemma 6.1.

Suppose that there exists some unique triplet (Y,Z,K)(Y,Z,K) in 𝕊𝔾2×𝔾2×𝔾2\mathbb{S}^{2}_{\mathbb{G}}\times\mathbb{H}^{2}_{\mathbb{G}}\times\mathbb{H}^{2}_{\mathbb{G}} such that :

Yt=WtTZs((1Ns)+Ns1D=1)𝑑BstTKs(1Ns)𝑑MstTf(s,Zs,Ks)𝑑s,Y_{t}=W-\int_{t}^{T}Z_{s}((1-N_{s})+N_{s}\textbf{1}_{D=1})dB_{s}-\int_{t}^{T}K_{s}(1-N_{s})dM_{s}-\int_{t}^{T}f(s,Z_{s},K_{s})ds,

where

f(s,Zs,Ks)\displaystyle f(s,Z_{s},K_{s}) =(λKs+λγA(eγAKs1))(1Ns)\displaystyle=\left(\lambda K_{s}+\frac{\lambda}{\gamma_{A}}(e^{-\gamma_{A}K_{s}}-1)\right)(1-N_{s})
+12γAZs2((1Ns)+Ns1D=1)+infa{κ(as)asZs(1Ns)θasZs1τs,D=1},\displaystyle+\frac{1}{2}\gamma_{A}Z_{s}^{2}((1-N_{s})+N_{s}\textbf{1}_{D=1})+\inf_{a\in\mathcal{B}}\left\{\kappa(a_{s})-a_{s}Z_{s}(1-N_{s})-\theta a_{s}Z_{s}\textbf{1}_{\tau\leq s,D=1}\right\},

then

Rta=UA(Yt0tκ(as)𝑑s)R_{t}^{a}=U_{A}\left(Y_{t}-\int_{0}^{t}\kappa(a_{s})ds\right)

satisfies a Martingale Optimality Principle for the Agent’s problem in this setting.

Proof.

By construction, RTa=UA(W0Tκ(as)𝑑s)R_{T}^{a}=U_{A}\left(W-\int_{0}^{T}\kappa(a_{s})ds\right) and R0aR_{0}^{a} is independent of the Agent’s action aa. As in Section 4, we compute the variations of RaR^{a} to obtain:

dRsa\displaystyle dR^{a}_{s} =γARsaZs((1Ns)+Ns1D=1)dBsa+Rsa(eγAKs1)(1Ns)dMs\displaystyle=-\gamma_{A}R_{s}^{a}Z_{s}((1-N_{s})+N_{s}\textbf{1}_{D=1})dB_{s}^{a}+R_{s}^{a}(e^{-\gamma_{A}K_{s}}-1)(1-N_{s})dM_{s}
+RsaγA(12γAZs2((1Ns)+Ns1D=1)f(s,Zs,Ks)+κ(as)+(λKs+λγA(eγAKs1))(1Ns))\displaystyle+R_{s}^{a}\gamma_{A}\left(\frac{1}{2}\gamma_{A}Z_{s}^{2}((1-N_{s})+N_{s}\textbf{1}_{D=1})-f(s,Z_{s},K_{s})+\kappa(a_{s})+(\lambda K_{s}+\frac{\lambda}{\gamma_{A}}(e^{-\gamma_{A}K_{s}}-1))(1-N_{s})\right)
+RsaγA(asZs(1Ns)+θasNs1D=1)\displaystyle+R_{s}^{a}\gamma_{A}\left(-a_{s}Z_{s}(1-N_{s})+\theta a_{s}N_{s}\textbf{1}_{D=1}\right)
=γARsaZs((1Ns)+Ns1D=1)dBsa+Rsa(eγAKs1)(1Ns)dMs\displaystyle=-\gamma_{A}R_{s}^{a}Z_{s}((1-N_{s})+N_{s}\textbf{1}_{D=1})dB_{s}^{a}+R_{s}^{a}(e^{-\gamma_{A}K_{s}}-1)(1-N_{s})dM_{s}
+RsaγA(12γAZs2((1Ns)+Ns1D=1)f(s,Zs,Ks)+κ(as0(1Ns))+κ(θas1Ns1D=1)+(λKs+λγA(eγAKs1))(1Ns))\displaystyle+R_{s}^{a}\gamma_{A}\left(\frac{1}{2}\gamma_{A}Z_{s}^{2}((1-N_{s})+N_{s}\textbf{1}_{D=1})-f(s,Z_{s},K_{s})+\kappa(a_{s}^{0}(1-N_{s}))+\kappa(\theta a_{s}^{1}N_{s}\textbf{1}_{D=1})+(\lambda K_{s}+\frac{\lambda}{\gamma_{A}}(e^{-\gamma_{A}K_{s}}-1))(1-N_{s})\right)
+RsaγA(as0Zs(1Ns)+θas1Ns1D=1)\displaystyle+R_{s}^{a}\gamma_{A}\left(-a_{s}^{0}Z_{s}(1-N_{s})+\theta a_{s}^{1}N_{s}\textbf{1}_{D=1}\right)

and therefore a\mathbb{R}^{a} is a super-martingale for every aa in \mathcal{B}. Setting :

as0(z)=A1zκ<A+A1zκ>A+zκ1AzκA{a^{0}_{s}}^{*}(z)=-A\textbf{1}_{\frac{z}{\kappa}<-A}+A\textbf{1}_{\frac{z}{\kappa}>A}+\frac{z}{\kappa}\textbf{1}_{-A\leq\frac{z}{\kappa}\leq A}

and

as1(z)=A1θzκA+A1θzκ>A+θzκ1AθzκA,{a^{1}_{s}}^{*}(z)=-A\textbf{1}_{\frac{\theta z}{\kappa}\leq-A}+A\textbf{1}_{\frac{\theta z}{\kappa}>A}+\frac{\theta z}{\kappa}\textbf{1}_{-A\leq\frac{\theta z}{\kappa}\leq A},

then

at=at01t<τ+at11tτ.a^{*}_{t}={a_{t}^{0}}^{*}\textbf{1}_{t<\tau}+{a_{t}^{1}}^{*}\textbf{1}_{t\geq\tau}.

We get that a\mathbb{R}^{a^{*}} is a a\mathbb{P}^{a^{*}}-martingale and the Agent’s response given WW is then a.a^{*}.

It remains to show that there actually exists a unique solution to (Y,Z,K)(Y,Z,K) in 𝕊𝔾2×𝔾2×𝔾2\mathbb{S}^{2}_{\mathbb{G}}\times\mathbb{H}^{2}_{\mathbb{G}}\times\mathbb{H}^{2}_{\mathbb{G}} to :

Yt=WtTZs((1Ns)+Ns1D=1)𝑑BstTKs(1Ns)𝑑MstTf(s,Zs,Ks)𝑑s,Y_{t}=W-\int_{t}^{T}Z_{s}((1-N_{s})+N_{s}\textbf{1}_{D=1})dB_{s}-\int_{t}^{T}K_{s}(1-N_{s})dM_{s}-\int_{t}^{T}f(s,Z_{s},K_{s})ds,

where

f(s,Zs,Ks)\displaystyle f(s,Z_{s},K_{s}) =(λKs+λγA(eγAKs1))(1Ns)\displaystyle=\left(\lambda K_{s}+\frac{\lambda}{\gamma_{A}}(e^{-\gamma_{A}K_{s}}-1)\right)(1-N_{s})
+12γAZs2((1Ns)+Ns1D=1)+infa{κ(as)asZs(1Ns)θasZs1τs,D=1}.\displaystyle+\frac{1}{2}\gamma_{A}Z_{s}^{2}((1-N_{s})+N_{s}\textbf{1}_{D=1})+\inf_{a\in\mathcal{B}}\left\{\kappa(a_{s})-a_{s}Z_{s}(1-N_{s})-\theta a_{s}Z_{s}\textbf{1}_{\tau\leq s,D=1}\right\}.

To to this, first note for any ss in [0,T][0,T] fixed, and for any t[s,T]t\in[s,T] there exists a unique pair (Yi,Zi)𝕊𝔾2×𝔾2(Y^{i},Z^{i})\in\mathbb{S}^{2}_{\mathbb{G}}\times\mathbb{H}^{2}_{\mathbb{G}} solution to the BSDE :

Yti(s)=WT1,1(s)1D=1tTf1(Zsi(s))𝑑stTZsi(s)𝑑Bs,Y^{i}_{t}(s)=W_{T}^{1,1}(s)\textbf{1}_{D=1}-\int_{t}^{T}f^{1}(Z_{s}^{i}(s))ds-\int_{t}^{T}Z_{s}^{i}(s)dB_{s}, (6.1)

where f1(z)=12γAz2+infa(κ(a)θaz)f^{1}(z)=\frac{1}{2}\gamma_{A}z^{2}+\inf_{a\in\mathcal{B}}(\kappa(a)-\theta az) and where the notation (Yi(s),Zi(s))(Y^{i}(s),Z^{i}(s)) is used to emphasize the dependency in ss of the terminal condition and its effect on the solution. This existence result simply follows from the fact that for each ss, (6.1) is now simply a Brownian BSDE that fits into the classical quadratic setting of Briand and Hu. We may then set :

W~=Yτi(τ)1τT1D=1+WTτ1,0(τ)1τT1D=0+W01T<τ,\tilde{W}=Y_{\tau}^{i}(\tau)\textbf{1}_{\tau\leq T}\textbf{1}_{D=1}+W^{1,0}_{T\wedge\tau}(\tau)\textbf{1}_{\tau\leq T}\textbf{1}_{D=0}+W^{0}\textbf{1}_{T<\tau},

which is a 𝔾Tτ\mathbb{G}_{T\wedge\tau} measurable random-variable. We set :

f2(z,k)=12γAz2+λk+λγA(eγAk1)+infa(κ(a)aZ).f^{2}(z,k)=\frac{1}{2}\gamma_{A}z^{2}+\lambda k+\frac{\lambda}{\gamma_{A}}(e^{-\gamma_{A}k}-1)+\inf_{a\in\mathcal{B}}\left(\kappa(a)-aZ\right).

This fits right into the setting of the recent work [15] on a default BSDE for Principal Agent problems. In particular, there exists a unique triplet (Y~,Z~,K~)(\tilde{Y},\tilde{Z},\tilde{K}) in 𝕊𝔾2×𝔾2×𝔾2\mathbb{S}^{2}_{\mathbb{G}}\times\mathbb{H}^{2}_{\mathbb{G}}\times\mathbb{H}^{2}_{\mathbb{G}} such that :

Y~t=W~tτTτZ~s𝑑BstτTτK~s𝑑MstτTτf2(Z~s,K~s)𝑑s.\tilde{Y}_{t}=\tilde{W}-\int_{t\wedge\tau}^{T\wedge\tau}\tilde{Z}_{s}dB_{s}-\int_{t\wedge\tau}^{T\wedge\tau}\tilde{K}_{s}dM_{s}-\int_{t\wedge\tau}^{T\wedge\tau}f^{2}(\tilde{Z}_{s},\tilde{K}_{s})ds.

Finally, setting :

  • Yt=Y~t(1Nt)+Yti(τ)Nt1D=1Y_{t}=\tilde{Y}_{t}(1-N_{t})+Y^{i}_{t}(\tau)N_{t}\textbf{1}_{D=1}

  • Zt=Z~t(1Nt)+Zti(τ)Nt1D=1Z_{t}=\tilde{Z}_{t}(1-N_{t})+Z^{i}_{t}(\tau)N_{t}\textbf{1}_{D=1}

  • Kt=K~t(1Nt)K_{t}=\tilde{K}_{t}(1-N_{t})

and noting that :

f(s,z,k)=f1(z)Ns1D=1+f2(z,k)(1Ns),f(s,z,k)=f^{1}(z)N_{s}\textbf{1}_{D=1}+f^{2}(z,k)(1-N_{s}),

we obtain that (Y,Z,K)(Y,Z,K) is a solution to :

Yt=WtTZs((1Ns)+Ns1D=1)𝑑BstTKs(1Ns)𝑑MstTf(s,Zs,Ks)𝑑s.\displaystyle Y_{t}=W-\int_{t}^{T}Z_{s}((1-N_{s})+N_{s}\textbf{1}_{D=1})dB_{s}-\int_{t}^{T}K_{s}(1-N_{s})dM_{s}-\int_{t}^{T}f(s,Z_{s},K_{s})ds. (6.2)

Finally, uniqueness holds through a classical reasoning, noting that up to a change in probability, YY is a local martingale that has continuous paths on a certain left-hand neighbourhood of TT.

References

  • [1] A. Aksamit and M. Jeanblanc. Enlargement of Filtration with Finance in View. SpringerBriefs in Quantitative Finance. Springer International Publishing, 2017.
  • [2] M. Bertrand and S. Mullainathan. Are CEOs Rewarded for Luck? The Ones Without Principals Are. The Quarterly Journal of Economics, 116(3):901–932, Aug. 2001.
  • [3] B. Biais, T. Mariotti, J.-C. Rochet, and S. Villeneuve. Large Risks, Limited Liability, and Dynamic Moral Hazard. Econometrica, 78(1):73–118, 2010.
  • [4] A. Capponi and C. Frei. Dynamic Contracting: Accidents Lead to Nonlinear Contracts. SSRN Scholarly Paper ID 2642261, Social Science Research Network, Rochester, NY, Aug. 2015.
  • [5] J. Cvitanić, X. Wan, and J. Zhang. Optimal Compensation with Hidden Action and Lump-Sum Payment in a Continuous-Time Model. Applied Mathematics and Optimization, 59(1):99–146, Feb. 2009.
  • [6] J. Cvitanić, D. Possamaï, and N. Touzi. Dynamic programming approach to principal–agent problems. Finance and Stochastics, 22(1):1–37, Jan. 2018.
  • [7] R. Dumitrescu, M. Grigorova, M.-C. Quenez, and A. Sulem. BSDEs with Default Jump. Computation and Combinatorics in Dynamics, Stochastics and Control, Abel Symposia, pages 233–263, Cham, 2018. Springer International Publishing.
  • [8] M. Jeanblanc, T. Mastrolia, D. Possamaï, and A. Réveillac. Utility maximization with random horizon: a BSDE approach. International Journal of Theoretical and Applied Finance, 18(07):1550045, Oct. 2015.
  • [9] M. Jeanblanc and M. Rutkowski. Modelling of default risk: Mathematical tools, 2000.
  • [10] T. Jeulin. Semi-martingales et grossissement d’une filtration. Lecture Notes in Mathematics. Springer-Verlag, Berlin Heidelberg, 1980.
  • [11] F. Hoffmann and S. Pfeil. Reward for Luck in a Dynamic Agency Model. The Review of Financial Studies, 23(9):3329–3345, 2010.
  • [12] B. Holmstrom and P. Milgrom. Aggregation and Linearity in the Provision of Intertemporal Incentives. Econometrica, 55(2):303–328, 1987.
  • [13] Y. Hu, P. Imkeller, and M. Müller. Utility maximization in incomplete markets. Annals of Applied Probability, 15(3):1691–1712, Aug. 2005.
  • [14] I. Kharroubi, T. Lim, and A. Ngoupeyou. Mean-Variance Hedging on Uncertain Time Horizon in a Market with a Jump. Applied Mathematics & Optimization, 68(3):413–444, Dec. 2013.
  • [15] J. Martin. A BSDE with default jump and unbounded terminal value arising in a Principal-Agent context. Working Paper, Jan. 2021.
  • [16] B. Øksendal and A. Sulem-Bialobroda. Applied Stochastic Control of Jump Diffusions. Universitext. Springer-Verlag, Berlin Heidelberg, 2nd edition, 2007.
  • [17] H. Pham. Stochastic control under progressive enlargement of filtrations and applications to multiple defaults risk management. Stochastic Processes and their Applications, 120(9):1795–1820, Aug. 2010.
  • [18] R. Rouge and N. E. Karoui. Pricing Via Utility Maximization and Entropy. Mathematical Finance, 10(2):259–276, 2000.
  • [19] Y. Sannikov. A Continuous- Time Version of the Principal: Agent Problem. The Review of Economic Studies, 75(3):957–984, 2008.
  • [20] H. Schattler and J. Sung. The First-Order Approach to the Continuous-Time Principal-Agent Problem with Exponential Utility. Journal of Economic Theory, 61(2):331–371, 1993.
  • [21] H. Schättler and J. Sung. On optimal sharing rules in discrete-and continuous-time principal-agent problems with exponential utility. Journal of Economic Dynamics and Control, 21(2):551–574, Feb. 1997.
  • [22] D. Zwillinger. Handbook of Differential Equations - 3rd Edition. Elsevier, 1997.