This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Gradually Reinforced Sample-Average-Approximation Differentiable Homotopy Method for a System of Stochastic Equations

Peixuan Li111School of Economics and Management, Southeast University. Email: lipeixuan@seu.edu.cn     Chuangyin Dang222The Corresponding Author. Department of Systems Engineering, City University of Hong Kong. Email: mecdang@cityu.edu.hk    Yang Zhan333School of Management and Engineering, Nanjing University. Email: zhanyang@nju.edu.cn
Abstract

This paper intends to apply the sample-average-approximation (SAA) scheme to solve a system of stochastic equations (SSE), which has many applications in a variety of fields. The SAA is an effective paradigm to address risks and uncertainty in stochastic models from the perspective of Monte Carlo principle. Nonetheless, a numerical conflict arises from the sample size of SAA when one has to make a tradeoff between the accuracy of solutions and the computational cost. To alleviate this issue, we incorporate a gradually reinforced SAA scheme into a differentiable homotopy method and develop a gradually reinforced sample-average-approximation (GRSAA) differentiable homotopy method in this paper. By introducing a series of continuously differentiable functions of the homotopy parameter tt ranging between zero and one, we establish a differentiable homotopy system, which is able to gradually increase the sample size of SAA as tt descends from one to zero. The set of solutions to the homotopy system contains an everywhere smooth path, which starts from an arbitrary point and ends at a solution to the SAA with any desired accuracy. The GRSAA differentiable homotopy method serves as a bridge to link the gradually reinforced SAA scheme and a differentiable homotopy method and retains the nice property of global convergence the homotopy method possesses while greatly reducing the computational cost for attaining a desired solution to the original SSE. Several numerical experiments further confirm the effectiveness and efficiency of the proposed method.

Keywords: System of Stochastic Equations, Gradually Reinforced Sample Average Approximation, Differentiable Homotopy Method

1 Introduction

This paper is concerned with the problem of finding an xnx\in\mathbb{R}^{n} satisfying

(SSE):𝔼ξ[f(x,ξ)]=0,\displaystyle\begin{array}[]{ll}(SSE):&\mathbb{E}_{\xi}[f(x,\xi)]=0,\end{array} (2)

where ξ\xi is a stochastic parameter in a probability space Ω\Omega, f(,ξ):nnf(\cdot,\xi):\mathbb{R}^{n}\rightarrow\mathbb{R}^{n} is a continuously differentiable mapping and 𝔼[]\mathbb{E}[\cdot] denotes the expected value over Ω\Omega. As an effective paradigm for addressing risks and uncertainty, the SSE (2) can be regarded as a natural extension of a system of equations (SE) to a stochastic environment and has been extensively studied and applied in a number of areas including economics, game theory, management science and engineering. An excellent review on these applications can be found in [43]. Furthermore, the SSE is closely associated with several other stochastic problems such as stochastic variational inequalities (SVI) [6, 7, 17, 29, 42], stochastic nonlinear complementarity problems (SNCP) [5, 18, 31] and stochastic optimization (SO) [12, 27, 30, 39].

To solve the deterministic SE, several well-known methods have been proposed in the literature, which include Newton methods, homotopy or path-following methods and their variants. However, the convergence of Newton methods very much depends on the starting points. The homotopy methods are a class of powerful methods for solving SEs and possess the desired property of global convergence, which play an inevitable and lasting role in various fields; see, for instances, [11, 16, 28, 53]. The homotopy methods can be classified as simplicial homotopy methods and differentiable homotopy methods. The simplicial homotopy methods, dating back to the seminal paper by [14, 44], are powerful mechanisms for solving equations with highly nonlinearities or non-smooth structures, but they cannot benefit much from the differentiability of problems and can be extremely time-consuming especially when the problem sizes are large. The differentiable homotopy methods, introduced in [33], perfectly overcome the deficiency of the simplicial homotopy methods and are capable of following a smooth path.

Over the past few decades, the differentiable homotopy methods have been substantively investigated in the literature and applied to various areas such as general equilibrium theory and game theory for solving the problems that can be reformulated as systems of differentiable equations. Two direct proofs were given in [20] to show the feasibility of linear tracing procedure, which was made differentiable for computing Nash equilibria in [22]. A stochastic version of linear tracing procedure was established in [24] for the computation and selection of stationary equilibria or stationary Markov perfect equilibria in stochastic games. Later, a more efficient homotopy method based on the idea of interior-point method was developed in [11]. An all-solution differentiable homotopy method was proposed in [32] to find all equilibria for static and dynamic games with continuous strategies. A convex quadratic penalty differentiable homotopy method was developed in [8] to compute perfect equilibria for large scale noncooperative games. A differentiable homotopy method was described in [45] to solve general equilibrium models with incomplete asset markets and its reliability was evidenced by numerous numerical experiments. A generically convergent differentiable homotopy method was constituted in [21] to compute an equilibrium for a finance version of the general equilibrium problem considered in [45]. A proximal block coordinate homotopy framework was presented in [50] to solve large scale sparse least squares problems through numerically following a piecewise smooth path. A polyhedral homotopy-baed method was designed in [36] to find solutions to generalized Nash equilibrium problems. More differentiable homotopy methods and their applications can be found in the literature such as [23, 35, 37, 51, 54] and the references therein.

Unfortunately, all the existing methods fail to directly solve a stochastic system since the underlying mapping 𝔼ξ[f(x,ξ)]\mathbb{E}_{\xi}[f(x,\xi)] of an SSE is in a form of an expected value with respect to a certain stochastic parameter vector ξ\xi and needs to be evaluated first. As demonstrated in [30], the evaluation of 𝔼ξ[f(x,ξ)]\mathbb{E}_{\xi}[f(x,\xi)] is generally a tough job, because the distribution of the stochastic parameter is usually unknown and can only be simulated with historical data. Even if the distribution is given, computing multidimensional integrals is very costly. To deal with these difficulties, two classes of methods have been proposed in the literature: stochastic approximation (SA) methods and simulation based approaches. It was proved in [41] that the SA method is almost surly convergent to a solution of an SSE, only when ff satisfies certain conditions and the samples and stepsize are suitably chosen. Notwithstanding, this method is very sensitive to the choice of the stepsize at each iteration and sometimes performs poorly in practice. A modified SA method with a better performance was proposed in [40]. This better performance, however, occurs just for a special class of convex stochastic optimization and saddle point problems.

The simulation based approaches are fashionable tools to address the uncertainty. Among them, the sample-average-approximation (SAA) is one of the most popular representatives. The basic idea of the SAA is rather simple: randomly generating samples for the stochastic parameter ξ\xi and approximating the “true” underlying mapping by the average of several deterministic mappings corresponding to these samples. As demonstrated in [26, 34, 46], an appropriate incorporation of the SAA into a numerical algorithm can lead to a reasonable performance in solving general stochastic problems, and an exponential convergence rate of the SAA for the SVI and SNCP has been established under some mild conditions. Inspired by this success, this paper intends to incorporate the SAA into a differentiable homotopy method for a solution to the SSE (2). To apply the SAA in the existing methods, one needs to make a tradeoff between a cheap coarse estimate and an expensive finer estimate. A common practice is to use a gradually reinforced SAA scheme, that is, initially selecting a small sample size and then gradually increasing the sample size to approximate the target expected value. This allows a rapid progress at early stages and reduces the overall compuational cost for finding a desired solution. However, this practice may fail to converge due to the changes in sample size [4].

It is well known that a differentiable homotopy method attains global convergence through a continuous deformation process. This naturally raises the question: can we devise a differentiable homotopy method, which is able to progressively increase the sample size of SAA during the deformation process? The question seems quite challenging and were not considered in the existing work due to the intuitive collision between the discreteness of samples and the continuous deformation of the homotopy methods. To overcome this hurdle, we introduce a sequence of continuously differentiable functions of the homotopy parameter tt ranging between zero and one and incorporate with these functions a gradually reinforced sample-average-approximation (GRSAA) scheme into a differentiable homotopy method. As a result of this incorporation, we establish a differentiable homotopy system and reap a GRSAA differentiable homotopy method. As tt descends from one to zero, the homotopy system gradually increases the sample size and eventually reaches the desired SAA (4) at t=0t=0. The solution set of the homotopy system contains an everywhere smooth path, which starts from an arbitrary point and ends at a solution to the desired SAA or a satisfactory approximate solution to the SSE (2). The GRSAA differentiable homotopy method serves as a bridge to link a gradually reinforced SAA scheme and a differentiable homotopy method and retains the inherent property of global convergence of a standard differentiable homotopy method. To evince the benefit of differentiability, we also present a GRSAA simplicial homotopy method. To make numerical comparisons with a standard differentiable homotopy method and the GRSAA simplicial homotopy method, we have implemented the GRSAA differentiable homotopy method to solve several important applications of the SSE such as the SVI and market equilibrium problems. Numerical results further verify that two main features of the GRSAA differentiable homotopy method, the gradual reinforcement in sample size and differentiability, can significantly enhance the effectiveness and efficiency.

The rest of the paper is organized as follows. In Section 2, we first give a brief review about the gradually reinforced SAA scheme and differentiable homotopy methods and then develop a gradually reinforced sample-average-approximation (GRSAA) differentiable homotopy method to solve the SAA system for the SSE (2). The convergence properties of the proposed method are discussed in Section 3. For numerical comparisons, Section 4 describes a GRSAA simplicial homotopy method, which belongs to a type of non-differentiable homotopy method. In Section 5, we employ the GRSAA differentiable homotopy method to solve some numerical examples and compare the performance of the GRSAA differentiable homotopy method with that of the GRSAA simplicial homotopy method and a standard differentiable homotopy method. We also apply the GRSAA differentiable homotopy method to solve several large-scale SSEs. All the numerical results are reported in Section 5. The paper is concluded in Section 6.

2 A Gradually Reinforced SAA Differentiable Homotopy Method

2.1 Background of the SAA and Homotopy Methods

The SAA is closely associated with the sample-path optimization (SPO) method, which is an effective paradigm for solving the problems that arise in the study of complex stochastic systems. The basic concept of the SPO method is to design some deterministic SEs with the underlying mappings being a sequence of computable functions, which has the underlying mapping of the original SSE as its limit. Some convergence conditions of the SPO method were provided in [19]. The sample-average-approximation (SAA) is generated by formulating the underlying mappings in the SPO into some sample average mappings. More specifically, suppose that ξ1,ξ2,,ξN\xi_{1},\xi_{2},\ldots,\xi_{N} are independently and identically distributed samples of ξ\xi, where NN is a positive integer. According to the Monte Carlo principle, the expected value 𝔼ξ[f(x,ξ)]\mathbb{E}_{\xi}[f(x,\xi)] can be approximated by the deterministic mapping, 1Ni=1Nf(x,ξi)\dfrac{1}{N}\sum\limits_{i=1}^{N}f(x,\xi_{i}), and the well-known strong Law of Large Numbers assures that for any ϵ>0\epsilon>0, there exists a sufficiently large number N0N_{0} such that when NN0N\geq N_{0},

P{|1Ni=1Nf(x,ξi)𝔼ξ[f(x,ξ)]|<ϵ}=1.\begin{array}[]{l}P\{|\dfrac{1}{N}\sum\limits_{i=1}^{N}f(x,\xi_{i})-\mathbb{E}_{\xi}[f(x,\xi)]|<\epsilon\}=1.\end{array} (3)

Let xNnx_{N}\in\mathbb{R}^{n} be a solution to the following SAA with very large value of NN,

1Ni=1Nf(x,ξi)=0.\displaystyle\dfrac{1}{N}\sum\limits_{i=1}^{N}f(x,\xi_{i})=0. (4)

Substituting xNx_{N} into (3), we have P{|𝔼ξf(xN,ξ)|<ϵ}=1P\{|\mathbb{E}_{\xi}f(x_{N},\xi)|<\epsilon\}=1, which implies that xNx_{N} provides a satisfactory approximate solution to the SSE (2) when NN is sufficiently large. Furthermore, when NN goes to infinity, xNx_{N} provides an accurate solution to the SSE (2) with probability one. However, the rate at which 1Ni=1Nf(x,ξi)\dfrac{1}{N}\sum\limits_{i=1}^{N}f(x,\xi_{i}) converges to 𝔼[f(x,ξ)]\mathbb{E}[f(x,\xi)] is O(N12)O(N^{-\frac{1}{2}}), that is, NN should be increased by 100 times in order to improve the accuracy of an estimate by one digit [34, 47]. Therefore, to have a highly accurate final solution, one has to choose a very large NN.

As stated in Section 1, a differentiable homotopy method provides a globally convergent solution approach to a deterministic SE. Such a method starts from the unique solution of a trivial problem, follows a smooth path in the solution set of a homotopy system and ends at a solution to the targeted problem. One straightforward approach to solving the SAA (4) is to apply a standard differentiable homotopy method and obtain a homotopy system,

(1t)1Ni=1Nf(x,ξi)+t(xx0)=0,(1-t)\dfrac{1}{N}\sum\limits_{i=1}^{N}f(x,\xi_{i})+t(x-x^{0})=0,

where x0x^{0} is any given point and tt is the homotopy parameter ranging between zero and one. Clearly, the sample size remains to be NN in the above system for any value of t[0,1)t\in[0,1) and the approach has no improvement to the classic SAA paradigm. Therefore, one would like to incorporate a differentiable homotopy method with a gradually reinforced SAA scheme so that the sample size can be increased progressively as tt descends from one to zero. For example, we can divide the interval [0,1][0,1] into NN subintervals, [0,1/N)[0,1/N), [1/N,2/N)[1/N,2/N), ,\dots, [11/N,1][1-1/N,1]. Let kk be an integer smaller than NN. At t=1k/Nt=1-k/N, only the first kk samples are used, and we are interested in the problem i=1kf(x,ξi)/k=0\sum\limits_{i=1}^{k}f(x,\xi_{i})/k=0. As tt decreases from 1k/N1-k/N to 1(k+1)/N1-(k+1)/N, ξk+1\xi_{k+1} enters the picture, and the homotopy system varies from i=1kf(x,ξi)/k=0\sum\limits_{i=1}^{k}f(x,\xi_{i})/k=0 to i=1k+1f(x,ξi)/(k+1)=0\sum\limits_{i=1}^{k+1}f(x,\xi_{i})/(k+1)=0. In this way, we do not need to use a large number of samples for every value of t[0,1)t\in[0,1).

However, there is a natural conflict between the SAA and the homotopy process. The homotopy process is a continuous deformation process, while the sample set for the SAA is discrete. As the homotopy parameter tt varies, the samples enter the homotopy system one by one, or more generally group by group. In the above example, it is natural to construct a series of homotopies for tt in different subintervals [(k1)/N,k/N][(k-1)/N,k/N], k=1,2,,Nk=1,2,\dots,N, and switch homotopies at each connection point k/Nk/N, which leads to a piecewise smooth path. The idea of switching homotopies was used in the literature for computing market equilibria in incomplete markets [3]. Since NN is very large, switching homotopies will frequently occur, which makes the computation very costly. In the next subsection, we show that this difficulty can be overcome with a sequence of continuously differentiable functions of tt.

2.2 A Gradually Reinforced SAA Differentiable Homotopy System

In this subsection, we incorporate a gradually reinforced SAA scheme into a differentiable homotopy method and develop a gradually reinforced sample-average-approximation (GRSAA) differentiable homotopy method. In this paper, the homotopy parameter will descend from one to zero. With the proposed method, the sample size in the homotopy system is getting larger and larger as the homotopy parameter tt is decreasing and becomes identical to that of the desired SAA as t=0t=0. The formulation of the GRSAA differentiable homotopy system in our method consists of four steps.

  • Construction of a sequence of sample-average mappings.

    Let LL be a positive integer with LNL\leq N and ={1,2,,L}\mathcal{L}=\{1,2,\dots,L\}. We partition the set of NN samples into LL subsets. For \ell\in\cal{L}, let qq_{\ell} be the number of samples in the first \ell subsets. Clearly, 0<q1<q2<<qL=N0<q_{1}<q_{2}<\dots<q_{L}=N. For each \ell, we define

    f(x)=1qi=1qf(x,ξi).f^{\ell}(x)=\frac{1}{q_{\ell}}\sum_{i=1}^{q_{\ell}}f(x,\xi_{i}).

    Let f0(x)=0nf^{0}(x)=0\in\mathbb{R}^{n} for all xnx\in\mathbb{R}^{n}. Then f(x)f^{\ell}(x), {0}\ell\in\mathcal{L}\cup\{0\}, form a sequence of sample-average mappings.

  • Formulation of a continuous mapping deforming from f1(x)f^{\ell-1}(x) to f(x)f^{\ell}(x) for each \ell\in\mathcal{L}.

    Let tt_{\ell}, {0}\ell\in\mathcal{L}\cup\{0\}, be a sequence with 1=t0>t1>>tL=01=t_{0}>t_{1}>\dots>t_{L}=0. Let d(x,t)d^{\ell}(x,t) be a mapping on n×[t,t1]\mathbb{R}^{n}\times[t_{\ell},t_{\ell-1}] that deforms from f1(x)f^{\ell-1}(x) to f(x)f^{\ell}(x) as tt decreases from t1t_{\ell-1} to tt_{\ell}. It is convenient to define

    d(x,t)=(1θ(t))f1(x)+θ(t)f(x),d^{\ell}(x,t)=(1-\theta_{\ell}(t))f^{\ell-1}(x)+\theta_{\ell}(t)f^{\ell}(x),

    where θ:[t,t1][0,1]\theta_{\ell}:[t_{\ell},t_{\ell-1}]\to[0,1] is a continuous function with θ(t1)=0\theta_{\ell}(t_{\ell-1})=0 and θ(t)=1\theta_{\ell}(t_{\ell})=1.

  • Composition of a continuously differentiable mapping deforming from f0(x)f^{0}(x) to fL(x)f^{L}(x).

    It is straightforward to define a mapping d:n×[0,1]nd:\mathbb{R}^{n}\times[0,1]\to\mathbb{R}^{n} as d(x,t)=d(x,t)d(x,t)=d^{\ell}(x,t) when t[t,t1]t\in[t_{\ell},t_{\ell-1}], \ell\in\mathcal{L}. Clearly, d(x,t)d(x,t) is a continuous mapping on n×[0,1]\mathbb{R}^{n}\times[0,1], which deforms from f0(x)f^{0}(x) to fL(x)f^{L}(x) as tt descends from t0=1t_{0}=1 to tL=0t_{L}=0. To ensure continuous differentiability of d(x,t)d(x,t) on n×[0,1]\mathbb{R}^{n}\times[0,1], the function θ(t)\theta_{\ell}(t) should be continuously differentiable on [t,t1][t_{\ell},t_{\ell-1}] with ddtθ(t)=0\frac{\mathrm{d}}{\mathrm{d}t}\theta_{\ell}(t_{\ell})=0 and ddtθ(t1)=0\frac{\mathrm{d}}{\mathrm{d}t}\theta_{\ell}(t_{\ell-1})=0. One potential candidate of θ(t)\theta_{\ell}(t) is given by

    θ(t)=sin2(tt1tt1π2).\theta_{\ell}(t)=\sin^{2}\left(\frac{t-t_{\ell-1}}{t_{\ell}-t_{\ell-1}}\frac{\pi}{2}\right). (5)

    In our development, we will adopt the sequence of functions θ(t)\theta_{\ell}(t) for all \ell\in\mathcal{L}.

  • Establishment of a GRSAA differentiable homotopy system.

    In order to establish a GRSAA differentiable homotopy system with a unique known solution on the level of t=1t=1, we incorporate with the homotopy parameter tt an affinely linear function into the homotopy mapping d(x,t)d(x,t) and arrive at a homotopy system,

    h(x,t)=(1t)d(x,t)+t(xx0)=0,\begin{array}[]{l}h(x,t)=(1-t)d(x,t)+t(x-x^{0})=0,\end{array} (6)

    where x0nx^{0}\in\mathbb{R}^{n} is an arbitrarily given point. Clearly, at t=t0=1t=t_{0}=1, the homotopy system (6) has a unique solution x0x^{0}. As tt is decreasing, more and more samples are entering into h(x,t)h(x,t). As tt goes to tL=0t_{L}=0, the system (6) approaches the desired SAA (4).

Lemma 1.

d(x,t)d(x,t) is continuously differentiable on n×[0,1]\mathbb{R}^{n}\times[0,1].

Proof.

Clearly, d(x,t)d(x,t) is continuously differentiable on n×(t,t1)\mathbb{R}^{n}\times\cup_{\ell\in{\mathcal{L}}}(t_{\ell},t_{\ell-1}). Next, we show that d(x,t)d(x,t) is continuously differentiable at the connection points (x,t)(x,t_{\ell}), \{L}\ell\in\mathcal{L}\backslash\{L\}. For each \ell\in\mathcal{L}, we have θ(t1)=0\theta_{\ell}(t_{\ell-1})=0, θ(t)=1\theta_{\ell}(t_{\ell})=1 and

ddtθ(t)=πtt1sin(tt1tt1π2)cos(tt1tt1π2)=π2(tt1)sin(tt1tt1π).\begin{array}[]{ll}\frac{\mathrm{d}}{\mathrm{d}t}\theta_{\ell}(t)&=\frac{\pi}{t_{\ell}-t_{\ell-1}}\sin\left(\frac{t-t_{\ell-1}}{t_{\ell}-t_{\ell-1}}\frac{\pi}{2}\right)\cos\left(\frac{t-t_{\ell-1}}{t_{\ell}-t_{\ell-1}}\frac{\pi}{2}\right)\\ \vskip 6.0pt\cr&=\frac{\pi}{2(t_{\ell}-t_{\ell-1})}\sin\left(\frac{t-t_{\ell-1}}{t_{\ell}-t_{\ell-1}}\pi\right).\end{array}

The derivatives at the connection points are given by

ddtθ(t1)=0andddtθ(t)=0.\frac{\mathrm{d}}{\mathrm{d}t}\theta_{\ell}(t_{\ell-1})=0\quad\text{and}\quad\frac{\mathrm{d}}{\mathrm{d}t}\theta_{\ell}(t_{\ell})=0.

Then, for each \{L}\ell\in\mathcal{L}\backslash\{L\},

limtttddtd(x,t)=[f+1(x)f(x)]limtttddtθ+1(t)=[f+1(x)f(x)]ddtθ+1(t)=0\lim\limits_{t\rightarrow t_{t_{\ell}}^{-}}\dfrac{\mathrm{d}}{\mathrm{d}t}d(x,t)=[f^{\ell+1}(x)-f^{\ell}(x)]\lim\limits_{t\rightarrow t_{t_{\ell}}^{-}}\dfrac{\mathrm{d}}{\mathrm{d}t}\theta_{\ell+1}(t)=[f^{\ell+1}(x)-f^{\ell}(x)]\frac{\mathrm{d}}{\mathrm{d}t}\theta_{\ell+1}(t_{\ell})=0

and

limttt+ddtd(x,t)=[f(x)f1(x)]limttt+ddtθ(t)=[f(x)f1(x)]ddtθ(t)=0.\lim\limits_{t\rightarrow t_{t_{\ell}}^{+}}\dfrac{\mathrm{d}}{\mathrm{d}t}d(x,t)=[f^{\ell}(x)-f^{\ell-1}(x)]\lim\limits_{t\rightarrow t_{t_{\ell}}^{+}}\dfrac{\mathrm{d}}{\mathrm{d}t}\theta_{\ell}(t)=[f^{\ell}(x)-f^{\ell-1}(x)]\frac{\mathrm{d}}{\mathrm{d}t}\theta_{\ell}(t_{\ell})=0.

Thus, d(x,t)d(x,t) is continuously differentiable at each connection point. This completes the proof of the lemma. ∎

From Lemma 1, we attain the following corollary immediately.

Corollary 1.

h(x,t)h(x,t) is a continuously differentiable mapping from n×[0,1]\mathbb{R}^{n}\times[0,1] to n\mathbb{R}^{n}.

Corollary 1 shows that the homotopy system (6) is a continuously differentiable system, which successfully settles the conflict between the discreteness of the sample set and the continuity requirement of homotopy deformation. With h(x,t)h(x,t), we will be able to establish the existence of a smooth path, and switching homotopies is no longer needed. Figure 1 illustrates how the GRSAA differentiable homotopy method with h(x,t)=0h(x,t)=0 works. In Figure 1, for t[0,1]t\in[0,1], the points on the path are solutions to the homotopy system (6). As a result of the continuity of h(x,t)h(x,t), every limiting point of the path, xx^{*}, satisfies fL(x)=0f^{L}(x^{*})=0. We remark that a replacement of t=1/(1+τ0)t_{\ell}=1/(1+\tau_{0}\ell) in the above development yields a GRSAA differentiable homotopy method converging to an accurate solution to (2) with probability one.

Refer to caption
Figure 1: A GRSAA Differentiable Homotopy Method

3 Development of a Smooth Path

3.1 Convergence Properties

We prove in this subsection that the GRSAA differentiable homotopy method with h(x,t)=0h(x,t)=0 is globally convergent under some mild conditions. Let us denote by int(W){\rm int}(W) the interior of the set WW. To ensure the existence of a smooth path that starts from (x0,1)(x^{0},1) and ends at a solution on the level of t=0t=0, the path is required to be trapped in a non-empty compact set. To meet this requirement, we make the following assumption about ff.

Assumption 1.

There exist compact convex sets 𝕏n\mathbb{X}\subset\mathbb{R}^{n} with nonempty interior such that, for any realization of ξ\xi, (xx0)f(x,ξ)>0(x-x^{0})^{\top}f(x,\xi)>0 for all xint(𝕏)x\notin{\rm int}(\mathbb{X}) and x0int(𝕏)x^{0}\in{\rm int}(\mathbb{X}).

Assumption 1 shows a global convergence condition in fixed point problems and holds in various problem-classes including VIs, MCPs, nonlinear programming problems (NLPs), and minimax problems (MMPs) under some mild conditions; see [2, 38, 48]. With this assumption, the solutions to the original problem, f(x,ξ)=0f(x,\xi)=0 for any realization of ξ\xi, are restricted in a compact convex set. Let x0x^{0} be an arbitrary interior point of 𝕏\mathbb{X}. Then, for any xint(𝕏)x\not\in\text{int}(\mathbb{X}), (xx0)d(x,t)>0(x-x^{0})^{\top}d(x,t)>0 and

(xx0)h(x,t)=(1t)(xx0)d(x,t)+txx02>0.\begin{array}[]{l}(x-x^{0})^{\top}h(x,t)=(1-t)(x-x^{0})^{\top}d(x,t)+t\|x-x^{0}\|^{2}>0.\end{array} (7)

We denote the set of solutions to the system (6) by

H1={(x,t)n×[0,1]|h(x,t)=0}.H^{-1}=\{(x,t)\in\mathbb{R}^{n}\times[0,1]\,|\,h(x,t)=0\}.

Under Assumption 1, we prove in the following that H1H^{-1} contains a connected component intersecting both n×{0}\mathbb{R}^{n}\times\{0\} and n×{1}\mathbb{R}^{n}\times\{1\}. For any given (x,t)𝕏×[0,1](x,t)\in\mathbb{X}\times[0,1], we constitute an unconstrained strictly convex optimization problem,

minynyh(x,t)+12yx2.\begin{array}[]{ll}\min\limits_{y\in\mathbb{R}^{n}}&y^{\top}h(x,t)+\dfrac{1}{2}\|y-x\|^{2}.\end{array} (8)

An application of the sufficient and necessary optimality condition to the problem (8) yields the system,

h(x,t)+(yx)=0.\displaystyle h(x,t)+(y-x)=0. (9)

Let φ(x,t)\varphi(x,t) denote the unique solution to the problem (8). Clearly, φ(x,t)\varphi(x,t) is a non-empty compact convex set. It follows from the system (9) that φ(x,t)=xh(x,t)\varphi(x,t)=x-h(x,t). Thus, φ(x,t)\varphi(x,t) is a continuous mapping on n×[0,1]\mathbb{R}^{n}\times[0,1]. Furthermore, we derive from (7) that for any given tt,

(x0x)(φ(x,t)x)>0,for allxint(𝕏).\displaystyle(x^{0}-x)^{\top}(\varphi(x,t)-x)>0,\quad\text{for all}\,\,x\not\in\text{int}(\mathbb{X}). (10)

These results together lead us to the following theorem.

Theorem 1.

Suppose that Assumption 1 holds. Then, for any given t[0,1]t\in[0,1], φ(x,t)\varphi(x,t) has a fixed point in int(𝕏){\rm int}(\mathbb{X}).

Proof.

Let tt be any given value in [0,1][0,1]. For any given x𝕏x\in\mathbb{X}, we denote by ψ(x)\psi(x) the unique solution to the strictly convex quadratic optimization problem,

miny𝕏yφ(x,t)22.\min\limits_{y\in\mathbb{X}}\|y-\varphi(x,t)\|^{2}_{2}.

Clearly, ψ(x)\psi(x) is a continuous mapping from 𝕏\mathbb{X} to 𝕏\mathbb{X}. Since 𝕏\mathbb{X} is a convex compact set, the well-known Brouwer’s fixed point theorem asserts that ψ(x)\psi(x) has a fixed point in 𝕏\mathbb{X}. Let x(t)𝕏x^{*}(t)\in\mathbb{X} be a fixed point of ψ(x)\psi(x). Then x(t)x^{*}(t) is the fixed point of φ(x,t)\varphi(x,t). Next we prove this assertion by contradicton.

Suppose that x(t)φ(x(t),t)x^{*}(t)\neq\varphi(x^{*}(t),t). Then we must have x(t)𝕏x^{*}(t)\in\partial\mathbb{X} and φ(x(t),t)𝕏\varphi(x^{*}(t),t)\notin\mathbb{X}. Let x~(t)=(1ρ)x(t)+ρx0\widetilde{x}^{*}(t)=(1-\rho)x^{*}(t)+\rho x^{0} for ρ(0,1]\rho\in(0,1]. x~(t)int(𝕏)\widetilde{x}^{*}(t)\in\rm int(\mathbb{X}) since 𝕏\mathbb{X} is convex. It follows that

φ(x(t),t)x~(t))22=φ(x(t),t)x(t)+x(t)(1ρ)x(t)ρx022=φ(x(t),t)x(t)ρ(x0x(t))22=φ(x(t),t)x(t)222ρ(x0x(t))(φ(x(t),t)x(t))+ρ2x(t)x02.\begin{array}[]{rl}&\|\varphi(x^{*}(t),t)-\widetilde{x}^{*}(t))\|^{2}_{2}\\ \vskip 6.0pt\cr=&\|\varphi(x^{*}(t),t)-x^{*}(t)+x^{*}(t)-(1-\rho)x^{*}(t)-\rho x^{0}\|^{2}_{2}\\ \vskip 6.0pt\cr=&\|\varphi(x^{*}(t),t)-x^{*}(t)-\rho(x^{0}-x^{*}(t))\|^{2}_{2}\\ \vskip 6.0pt\cr=&\|\varphi(x^{*}(t),t)-x^{*}(t)\|^{2}_{2}-2\rho(x^{0}-x^{*}(t))^{\top}(\varphi(x^{*}(t),t)-x^{*}(t))+\rho^{2}\|x^{*}(t)-x^{0}\|^{2}.\end{array}

Recall that (x0x(t))(φ(x(t),t)x(t))>0.(x^{0}-x^{*}(t))^{\top}(\varphi(x^{*}(t),t)-x^{*}(t))>0. Thus, when ρ>0\rho>0 is sufficiently small, one must have

φ(x(t),t)x~(t))22<φ(x(t),t)x(t)22.\|\varphi(x^{*}(t),t)-\widetilde{x}^{*}(t))\|^{2}_{2}<\|\varphi(x^{*}(t),t)-x^{*}(t)\|^{2}_{2}.

Therefore, ψ(φ(x(t),t))x(t)\psi(\varphi(x^{*}(t),t))\neq x^{*}(t). Hence,

(x0x(t))(φ(x(t),t)x(t))0.(x^{0}-x^{*}(t))^{\top}(\varphi(x^{*}(t),t)-x^{*}(t))\leq 0.

A contradiction occurs. This completes the proof. ∎

Theorem 1 says that for any given t[0,1]t\in[0,1], there exists an x𝕏x\in\mathbb{X} such that φ(x,t)=x\varphi(x,t)=x. This conclusion shows that the homotopy system (6) is precisely the same as the system (9) and the set of solutions to the homotopy system can be rewritten as

H1={(x,t)𝕏×[0,1]|x=φ(x,t)}.H^{-1}=\{(x,t)\in\mathbb{X}\times[0,1]\,|\,x=\varphi(x,t)\}.

For our further development, we need the following fixed point theorem.

Theorem 2 (Browder’s Fixed Point Theorem).

Let SS be a non-empty, compact and convex subset of n\mathbb{R}^{n} and let f:S×[0,1]Sf:S\times[0,1]\rightarrow S be a continuous correspondence. Then, the set G={(s,t)S×[0,1]|s=f(s,t)}G=\{(s,t)\in S\times[0,1]\,|\,s=f(s,t)\} contains a connected subset GcG^{\rm c} such that (S×{1})Gc(S\times\{1\})\bigcap G^{\rm c}\neq\emptyset and (S×{0})Gc(S\times\{0\})\bigcap G^{c}\neq\emptyset.

As a corollary of Theorem 2, we come to the following result.

Corollary 2.

H1H^{-1} contains a connected subset Hc1H_{c}^{-1} such that 𝕏×{0}Hc1\mathbb{X}\times\{0\}\cap H_{c}^{-1}\neq\emptyset and 𝕏×{1}Hc1\mathbb{X}\times\{1\}\cap H_{c}^{-1}\neq\emptyset.

Corollary 2 assures the global convergence of the GRSAA differentiable homotopy method.

3.2 A GRSAA Differentiable Homotopy Path

To develop an efficient GRSAA differentiable homotopy method for computing a solution to the SAA (4), we need to construct an everywhere smooth path that starts from (x0,1)(x^{0},1) and ends at a solution to the SAA (4) on the level of t=0t=0. A common sufficient condition for the existence of such a smooth path is that zero is a regular value of a homotopy system. The next lemma shows that zero is a regular value of h(x,1)h(x,1) at the starting point (x0,1)(x^{0},1).

Lemma 2.

At t=1t=1, zero is a regular value of h(x,1)h(x,1) on 𝕏\mathbb{X}.

Proof.

Taking derivatives of h(x,1)h(x,1) with respect to xx, we obtain that the Jacobian matrix of h(x,1)h(x,1) is an identity matrix. Thus zero is a regular value of h(x,1)h(x,1). This completes the proof. ∎

We prove in the following theorem that, under the condition that zero is a regular value of h(x,t)=0h(x,t)=0, the set H1H^{-1} contains an everywhere smooth path leading to a solution to (4) as tt goes to zero.

Theorem 3.

Suppose that zero is a regular value of h(x,t)=0h(x,t)=0 on n×(0,1)\mathbb{R}^{n}\times(0,1). Then there exists an everywhere smooth path in H1H^{-1}, which starts from the unique solution (x0,1)(x^{0},1) on the level of t=1t=1 and ends at a solution to (4) on the level of t=0t=0.

Proof.

Corollary 2 tells us that H1H^{-1} has a connected component that intersects both n×{1}\mathbb{R}^{n}\times\{1\} and n×{0}\mathbb{R}^{n}\times\{0\}. Since the system (6) has a unique solution (x0,1)(x^{0},1) at t=1t=1, the connected component in H1H^{-1} is unique. The condition of Theorem 3 together with the well-known implicit function theorem ensures that the connected component forms a smooth path, which starts from (x0,1)(x^{0},1) and ends at a solution to (4) on the level of t=0t=0. ∎

This theorem relies on the condition that zero is a regular value of h(x,t)h(x,t) on n×(0,1)\mathbb{R}^{n}\times(0,1). To get rid of this condition, a general approach is to subtract from h(x,t)h(x,t) a perturbation term of t(1t)αt(1-t)\alpha, where αn\alpha\in\mathbb{R}^{n} and α\|\alpha\| is sufficiently small. Subtracting from h(x,t)h(x,t) the perturbation term, we get h(x,t;α)=h(x,t)t(1t)α=0h(x,t;\alpha)=h(x,t)-t(1-t)\alpha=0. For any fixed α\alpha, let hα(x,t)=h(x,t)t(1t)αh_{\alpha}(x,t)=h(x,t)-t(1-t)\alpha and Hα1={(x,t)n×[0,1]|hα(x,t)=0}H^{-1}_{\alpha}=\{(x,t)\in\mathbb{R}^{n}\times[0,1]\,|\,h_{\alpha}(x,t)=0\}. As t=0t=0 and t=1t=1, the perturbation term disappears and h(x,t)=hα(x,t)h(x,t)=h_{\alpha}(x,t). Clearly, as α\|\alpha\| goes to zero, the distance between H1H^{-1} and Hα1H^{-1}_{\alpha} approaches zero. Therefore, as α\|\alpha\| is sufficiently small, Hα1H^{-1}_{\alpha} also contains a connected component intersecting both 𝕏×{0}\mathbb{X}\times\{0\} and 𝕏×{1}\mathbb{X}\times\{1\}. These results together lead to the following theorem.

Theorem 4.

For generic αn\alpha\in\mathbb{R}^{n} with sufficiently small α\|\alpha\|, Hα1H^{-1}_{\alpha} contains a unique smooth path starting from (x0,1)(x^{0},1).

Proof.

As a mapping of (x,t,α)(x,t,\alpha) on n×(0,1)×n\mathbb{R}^{n}\times(0,1)\times\mathbb{R}^{n}, the Jacobian matrix of h(x,t;α)h(x,t;\alpha) is an n×(2n+1)n\times(2n+1) matrix given by [Dx,th(x,t)t(1t)In][D_{x,t}h(x,t)\;-t(1-t)I_{n}], where InI_{n} is an n×nn\times n identity matrix. Therefore, the Jacobian matrix of h(x,t;α)h(x,t;\alpha) has full row rank on n×(0,1)×n\mathbb{R}^{n}\times(0,1)\times\mathbb{R}^{n}. This together with Lemma 2 asserts that zero is a regular value of h(x,t;α)h(x,t;\alpha) on n×(0,1]×n\mathbb{R}^{n}\times(0,1]\times\mathbb{R}^{n}. We know from the well-known transversality theorem [15] that, for almost all αn\alpha\in\mathbb{R}^{n}, zero is also a regular value of hα(x,t)h_{\alpha}(x,t). Therefore, when α\|\alpha\| is sufficiently small, we derive from the well-known implicit function theorem that the connected component in Hα1H^{-1}_{\alpha} intersecting 𝕏×{1}\mathbb{X}\times\{1\} forms a unique smooth path starting from (x0,1)(x^{0},1). This completes the proof. ∎

Theorem 4 ensures that one can follow the smooth path in Hα1H^{-1}_{\alpha} starting from (x0,1)(x^{0},1) to find a solution to the SAA (4). In our numerical implementation of the proposed method, we always set α=0n\alpha=0\in\mathbb{R}^{n}.

4 A Gradually Reinforced SAA Simplicial Homotopy Method

To further evince the advantage of the GRSAA differentiable homotopy method, we describe in this section a gradually reinforced SAA (GRSAA) simplicial homotopy method. As an underlying triangulation of the method, we make use of the D2D_{2}-triangulation of (0,1]×n(0,1]\times\mathbb{R}^{n} with continuous refinement of grid size, whose definition and pivot rules are given in [10]. Given τ0>0\tau_{0}>0, let t=1/(1+τ0)t_{\ell}=1/(1+\tau_{0}\ell), =0,1,\ell=0,1,\ldots. Then, tt_{\ell}, =0,1,\ell=0,1,\ldots, form a descent sequence with t0=1t_{0}=1 and limt=0\lim\limits_{\ell\rightarrow\infty}t_{\ell}=0. Given τ1>0\tau_{1}>0, we define q=τ1q_{\ell}=\tau_{1}\ell, =1,2,\ell=1,2,\ldots. Therefore, as \ell\rightarrow\infty, qq_{\ell}\rightarrow\infty. Given the initial grid size ϖ0>0\varpi_{0}>0 and r{1/j|j=1,2,}r_{\ell}\in\{1/j\;|\;j=1,2,\ldots\}, let ϖ+1=rϖ\varpi_{\ell+1}=r_{\ell}\varpi_{\ell}, =0,1,\ell=0,1,\ldots, which is the gird size of a simplex of the D2D_{2}-triangulation in {t}×n\{t_{\ell}\}\times\mathbb{R}^{n}. Let x0𝕏x^{0}\in\mathbb{X} be a given point. We still utilize the continuously differentiable function θ(t)\theta_{\ell}(t) defined in Section 2 to form a simplicial homotopy system,

g(t,x):=(1t)d¯(t,x)+t(xx0)=0,\begin{array}[]{l}g(t,x):=(1-t)\bar{d}(t,x)+t(x-x^{0})=0,\end{array} (11)

where d¯(t,x)=θ(t)1qi=1qf(x,ξi)+(1θ(t))1q+1i=1q+1f(x,ξi)\bar{d}(t,x)=\theta(t)\dfrac{1}{q_{\ell}}\sum\limits_{i=1}^{q_{\ell}}f(x,\xi_{i})+(1-\theta(t))\dfrac{1}{q_{\ell+1}}\sum\limits_{i=1}^{q_{\ell+1}}f(x,\xi_{i}) for t+1ttt_{\ell+1}\leq t\leq t_{\ell}. Clearly, when t=1t=1, the system (11) has a unique solution x0x^{0}. As follows, we exploit the homotopy mapping g(t,x)g(t,x) to attain a GRSAA simplicial homotopy method.

Let vc=(0,x0ϖ0n+1e)v^{c}=(0,x^{0}-\frac{\varpi_{0}}{n+1}e) with e=(1,1,,1)ne=(1,1,\ldots,1)^{\top}\in\mathbb{R}^{n}. Let D2D_{2} be the collection of all simplices of the D2D_{2}-triangulation of (0,1]×n(0,1]\times\mathbb{R}^{n} after the translation of vcv^{c}, where the translation ensures that (1,x0)(1,x^{0}) is in the interior of a unique simplex in {1}×n\{1\}\times\mathbb{R}^{n}. We denote a qq-dimensional simplex in D2D_{2} by σ=v0,v1,,vq\sigma=\langle v^{0},v^{1},\ldots,v^{q}\rangle, where viv^{i} is a vertex of σ\sigma for i=0,1,,qi=0,1,\ldots,q.

Definition 1.

σ=v0,v1,,vnD2\sigma=\langle v^{0},v^{1},\ldots,v^{n}\rangle\in D_{2} is a complete simplex if the system,

l=0nζl(1g(vl))=(10) and ζ0,\sum\limits_{l=0}^{n}\zeta_{l}\left(\begin{array}[]{c}1\\ g(v^{l})\end{array}\right)=\left(\begin{array}[]{c}1\\ 0\end{array}\right)\text{ and }\zeta\geq 0, (12)

has a solution.

Assumption 2 (Nondegeneracy Assumption).

The system (12) has a unique solution with ζ>0\zeta>0.444Note that this assumption can be eliminated if the lexicographic pivoting rule in Eaves [13] and Todd [48] is applied in the linear programming step of the following method.

Following a similar argument to that in [8], we know that there is a unique complete simplex contained in {1}×n\{1\}\times\mathbb{R}^{n}. Given these notations, a simplicial homotopy method for computing an approximate solution to the SSE can be stated as follows.

Step 0:

Let δ0\delta_{0} be a given sufficiently small positive number. Let τ0=v0,v1,,vn\tau_{0}=\langle v^{0},v^{1},\ldots,v^{n}\rangle be the unique complete simplex in {1}×n\{1\}\times\mathbb{R}^{n} with v0=(1,x0)v^{0}=(1,x^{0}) and σ0\sigma_{0} be the unique (n+1)(n+1)-dimensional simplex in [t1,1]×n[t_{1},1]\times\mathbb{R}^{n} with τ0\tau_{0} as its facet. Let v+v^{+} be the vertex of σ0\sigma_{0} opposite to τ0\tau_{0}. Let =0\ell=0 and go to Step 1.

Step 1:

Perform a linear programming step to bring (1g(v+))\left(\begin{array}[]{c}1\\ g(v^{+})\end{array}\right) into the system, l=0nζl(1g(vl))=(10).\sum\limits_{l=0}^{n}\zeta_{l}\left(\begin{array}[]{c}1\\ g(v^{l})\end{array}\right)=\left(\begin{array}[]{c}1\\ 0\end{array}\right). Suppose that (1g(vl))\left(\begin{array}[]{c}1\\ g(v^{l})\end{array}\right) leaves the system. Let τ+1\tau_{\ell+1} be the facet of σ\sigma_{\ell} opposite to vlv^{l} and vl=v+v^{l}=v^{+}. Go to Step 2.

Step 2:

If τ+1{t+1}×n\tau_{\ell+1}\subset\{t_{\ell+1}\}\times\mathbb{R}^{n} and t+1δ0t_{\ell+1}\leq\delta_{0}, the method terminates. Otherwise, let σ+1\sigma_{\ell+1} be the unique simplex that shares together with σ\sigma_{\ell} a common facet τ+1\tau_{\ell+1}. Let v+v^{+} be the vertex of σ+1\sigma_{\ell+1} opposite to τ+1\tau_{\ell+1} and =+1\ell=\ell+1. Go to Step 1.

As follows, we show that the simplicial path generated by the GRSAA simplicial homotopy method leads to a solution to the SSE (2). Let g1(0)={(t,x)[0,1]×n|g(t,x)=0}.g^{-1}(0)=\{(t,x)\in[0,1]\times\mathbb{R}^{n}\;|\;g(t,x)=0\}. Then it follows from Assumption 1 that g1(0)g^{-1}(0) is a compact set. As a result of Definition 1 and the continuity of ff, it is easy to verify that all the complete simplices are contained in a bounded set of [0,1]×n[0,1]\times\mathbb{R}^{n}. Under the nondegeneracy assumption, following the same argument as that in Todd [48], one can derive that all the simplices generated by the GRSAA simplicial homotopy method are different, that is, no cycle will occur. Let τ\tau_{\ell}, =1,2,\ell=1,2,\ldots, be the complete simplices generated by the method. Let t¯\bar{t}_{\ell} be the smallest value in (0,1](0,1] such that τ(0,t¯]×n\tau_{\ell}\subset(0,\bar{t}_{\ell}]\times\mathbb{R}^{n}. Since all τ\tau_{\ell}, =1,2,\ell=1,2,\ldots, are contained in a bounded set, we must have t¯0\bar{t}_{\ell}\rightarrow 0 as \ell\rightarrow\infty.

Let σ={t}×x0,x1,,xn\sigma=\{t\}\times\langle x^{0},x^{1},\ldots,x^{n}\rangle be an nn-dimensional simplex in {t}×n\{t\}\times\mathbb{R}^{n}. For y=l=0nζl(yl)y=\sum\limits_{l=0}^{n}\zeta_{l}(y^{l}) with ζl0\zeta_{l}\geq 0 and l=0nζl=1\sum\limits_{l=0}^{n}\zeta_{l}=1, a piecewise linear approximation of g(t,y)g(t,y) is given by g¯(t,y)=l=0nζlg(t,yl).\bar{g}(t,y)=\sum\limits_{l=0}^{n}\zeta_{l}g(t,y^{l}).

Lemma 3.

For t(0,1]t\in(0,1], let σ={t}×y0,y1,,yn\sigma^{*}=\{t\}\times\langle y^{0},y^{1},\ldots,y^{n}\rangle be a complete simplex in {t}×n\{t\}\times\mathbb{R}^{n} with ylny^{l}\in\mathbb{R}^{n}. Let ζ\zeta^{*} be the corresponding unique solution of the system (12). Then, (ya(t))=l=0nζlyln(y^{a}(t))=\sum\limits_{l=0}^{n}\zeta^{*}_{l}y^{l}\in\mathbb{R}^{n} is a zero point of g¯(t,)\bar{g}(t,\cdot).

Proof.

It follows from the system (12) that l=0nζlg(t,yl)=0\sum\limits_{l=0}^{n}\zeta^{*}_{l}g(t,y^{l})=0. Thus, g¯(t,ya(t))=l=0nζlg(t,yl)=0\bar{g}(t,y^{a}(t))=\sum\limits_{l=0}^{n}\zeta^{*}_{l}g(t,y^{l})=0. Therefore, ya(t)ny^{a}(t)\in\mathbb{R}^{n} is a zero point of g¯(t,)\bar{g}(t,\cdot). ∎

Since g(t,y)g(t,y) is uniformly continuous on [0,1]×n[0,1]\times\mathbb{R}^{n}, there is a constant ρ0>0\rho_{0}>0 such that g(t,y)g(t^,y^)ρ0(t,y)(t^,y^)\|g(t,y)-g(\hat{t},\hat{y})\|\leq\rho_{0}\|(t,y)-(\hat{t},\hat{y})\| for any (t,y)(t,y) and (t^,y^)(\hat{t},\hat{y}) in [0,1]×n[0,1]\times\mathbb{R}^{n}.

Lemma 4.

For t(0,1]t\in(0,1], let ya(t)ny^{a}(t)\in\mathbb{R}^{n} be a zero point of g¯(t,)\bar{g}(t,\cdot). Then, g(t,ya(t))ρ0ϖ(t)\|g(t,y^{a}(t))\|\leq\rho_{0}\varpi(t), where ϖ(t)\varpi(t) is the gird size of the triangulation restricted on {t}×n\{t\}\times\mathbb{R}^{n}.

Proof.

Assume that (t,ya(t))σ={t}×y0,y1,,yn(t,y^{a}(t))\in\sigma^{*}=\{t\}\times\langle y^{0},y^{1},\ldots,y^{n}\rangle. Then,

g(t,ya(t))=g(t,ya(t))g¯(t,ya(t))=l=0nζl(g(t,ya(t))g(t,yl))l=0nζlg(t,ya(t))g(t,yl)l=0nζlρ0ya(t)ylρ0ϖ(t).\begin{array}[]{rl}&\|g(t,y^{a}(t))\|=\|g(t,y^{a}(t))-\bar{g}(t,y^{a}(t))\|=\|\sum\limits_{l=0}^{n}\zeta^{*}_{l}(g(t,y^{a}(t))-g(t,y^{l}))\|\\ \leq&\sum\limits_{l=0}^{n}\zeta^{*}_{l}\|g(t,y^{a}(t))-g(t,y^{l})\|\leq\sum\limits_{l=0}^{n}\zeta^{*}_{l}\rho_{0}\|y^{a}(t)-y^{l}\|\leq\rho_{0}\varpi(t).\end{array}

The proof is completed. ∎

This lemma implies that every limit point of the simplicial path yields a solution to the SSE (2) with probability one as tt goes to zero.

5 Numerical Performance

We apply in this section the GRSAA differentiable homotopy method to solve several applications of the SSE (2), which include a stochastic market equilibrium problem and a stochastic variational inequality problem. To numerically trace the smooth path, we employ a standard predictor-corrector method.555Interested readers can refer to [1, 15] for more details about the predictor-corrector method. Moreover, we have made numerical comparisons of the GRSAA differentiable homotopy method with the GRSAA simplicial homotopy method and a standard differentiable homotopy method. All these methods are coded in MatLab(R2019a). The computation has been carried out on a 2.00 GHz Windows PC with CORE i7. The numerical results further confirm the effectiveness and efficiency of the GRSAA differentiable homotopy method.

5.1 A Stochastic Market Equilibrium Problem

This subsection is concerned with a stochastic market equilibrium problem. Suppose that there are three goods and two firms in an economy. The consumers in the economy can be represented by one agent, who has a constant elasticity of substitution (CES) utility function with a stochastic substitution parameter, u(x,y,z)=(2xξ+3yξ+zξ)1/ξu(x,y,z)=(2x^{\xi}+3y^{\xi}+z^{\xi})^{1/\xi}, where xx, yy and zz are the amounts of the three goods being consumed. The initial endowment is given by w=(1,1,1)w=(1,1,1)^{\top}. In the economy, the agent wants to maximize her utility by determining an optimal consumption plan. A direct application of the sufficient and necessary optimality conditions to the convex optimization problem for the agent yields the excess demand of the agent at the market price p=(px,py,pz)p=(p_{x},p_{y},p_{z})^{\top},

f(p,ξ)=pw(p1g1,p2g2,p3g3),\begin{array}[]{l}f(p,\xi)=p^{\top}w\left(\dfrac{p_{1}}{g_{1}},\dfrac{p_{2}}{g_{2}},\dfrac{p_{3}}{g_{3}}\right)^{\top},\end{array}

where g1=p1ξ+231ξ1p2ξ+21ξ1p3ξg_{1}=p_{1}^{\xi}+\frac{2}{3}^{\frac{1}{\xi-1}}p_{2}^{\xi}+2^{\frac{1}{\xi-1}}p_{3}^{\xi}, g2=321ξ1p1ξ+p2ξ+31ξ1p3ξg_{2}=\frac{3}{2}^{\frac{1}{\xi-1}}p_{1}^{\xi}+p_{2}^{\xi}+3^{\frac{1}{\xi-1}}p_{3}^{\xi}, and g3=121ξ1p1ξ+131ξ1p2ξ+p3ξg_{3}=\frac{1}{2}^{\frac{1}{\xi-1}}p_{1}^{\xi}+\frac{1}{3}^{\frac{1}{\xi-1}}p_{2}^{\xi}+p_{3}^{\xi} with p1=px1ξ1p_{1}=p_{x}^{\frac{1}{\xi-1}}, p2=py1ξ1p_{2}=p_{y}^{\frac{1}{\xi-1}}, and p3=pz1ξ1p_{3}=p_{z}^{\frac{1}{\xi-1}}. The production technology of firms are described by the following matrix,

A=[321117727119].A=\begin{bmatrix}-\dfrac{3}{2}&1&1\\ -1&-\dfrac{77}{27}&\dfrac{11}{9}\end{bmatrix}.

Since no firm makes a positive profit in an equilibrium, we have the constraint Ap0Ap\leq 0. Assume that px+py+pz1p_{x}+p_{y}+p_{z}\leq 1.666This constraint ensures that the feasible set is a compact polytope. Then the feasible set reads as P={p+3:Ap0,ep1}P=\{p\in\mathbb{R}^{3}_{+}\,:\,Ap\leq 0,\,e^{\top}p\leq 1\}, where e=(1,1,1)e=(1,1,1)^{\top}. We say an equilibrium is reached if and only if there exist vectors z6z\in\mathbb{R}^{6} and s6s\in\mathbb{R}^{6} together with pp satisfying that 𝔼ξf(p,ξ)Bz=0\mathbb{E}_{\xi}f(p,\xi)-B^{\top}z=0, Bp+sb=0Bp+s-b=0, and zs=0zs=0, where B=[A;In;e]6×3B=[A;\;-I_{n};\;e^{\top}]\in\mathbb{R}^{6\times 3} and b=(0,0,0,0,0,1)b=(0,0,0,0,0,1)^{\top} [49, 52].

The corresponding GRSAA differentiable homotopy method for this problem is as follows. We set the sample size N=104N=10^{4}. After randomly generating a batch of samples ξ1,ξ2,,ξN\xi_{1},\xi_{2},\ldots,\xi_{N} of the stochastic variable ξ\xi from the uniform distribution on [1,1][-1,1], one can approximate the expected value 𝔼ξf(p,ξ)\mathbb{E}_{\xi}f(p,\xi) by an SAA scheme, 1Ni=1Nf(p,ξi)\dfrac{1}{N}\sum\limits_{i=1}^{N}f(p,\xi_{i}). Let t0=1t_{0}=1, tL=0t_{L}=0 and tt_{\ell}, =1,2,,L1\ell=1,2,\ldots,L-1, be randomly generated from (0,1)(0,1) with t<t1t_{\ell}<t_{\ell-1}. We make up the following unconstrained convex optimization problem,

maxx3(1t)x(d(p,t)+tκ0i=16log(biBix))t2xp02,\begin{array}[]{ll}\max\limits_{x\in\mathbb{R}^{3}}&(1-t)x^{\top}(d(p,t)+t^{\kappa_{0}}\sum\limits_{i=1}^{6}\text{log}(b_{i}-B_{i}x))-\dfrac{t}{2}\|x-p^{0}\|^{2},\\ \end{array} (13)

where κ02\kappa_{0}\geq 2, p0p^{0} is any given interior point in PP and d(p,t)=d(p,t)d(p,t)=d^{\ell}(p,t) for t[t,t1]t\in[t_{\ell},t_{\ell-1}] with d(p,t)d^{\ell}(p,t) as defined in Section 2. An application of the optimality conditions to the problem (13) together with a fixed point argument leads to the following GRSAA differentiable homotopy system,

(1t)(d(p,t)Bz(y))t(pp0)=0,Bp+s(y)b=0,\begin{array}[]{l}(1-t)(d(p,t)-B^{\top}z(y))-t(p-p^{0})=0,\\ \vskip 6.0pt\cr Bp+s(y)-b=0,\end{array} (14)

where z(y)=(y2+4ty2)κ0z(y)=\left(\dfrac{\sqrt{y^{2}+4t}-y}{2}\right)^{\kappa_{0}} and s(y)=(y2+4t+y2)κ0s(y)=\left(\dfrac{\sqrt{y^{2}+4t}+y}{2}\right)^{\kappa_{0}} with y6y\in\mathbb{R}^{6} being a new variable.777A well-chosen transformation of variables can significantly reduce the number of variables and constraints so that numerical efficiency can be greatly improved. This technique has been frequently used in the literature such as [8, 25]. Therefore, there exists a smooth path contained in the set of solutions to the system (14), which starts from a unique solution on the level of t=1t=1 and ends at an approximate equilibrium for the original market equilibrium problem on the level of t=0t=0. By applying a predictor-corrector method to trace the smooth path, we eventually reach a solution p=(0.40,0.45,0.15)p=(0.40,0.45,0.15) in 12 iterations and 0.6096 seconds. Figure 2 shows the distances from the current point to the desired solution of the original problem at each iteration for the GRSAA differentiable homotopy method. The changes of different variables in iterations are illustrated in Figure 3.

Refer to caption
Figure 2: Numerical Results for the Market Equilibrium Problem
Refer to caption
Figure 3: Changes of Variables in Iterations

5.2 A Stochastic System of Equations

A simplicial homotopy method was proposed in [9] to compute a solution to a deterministic system of equations, xi5sin(ij=1nxj)=0x_{i}-5\sin(i\sum\limits_{j=1}^{n}x_{j})=0, i=1,2,,ni=1,2,\ldots,n. This subsection focuses on a stochastic version of this problem, 𝔼ξ[f(x,ξ)]=0\mathbb{E}_{\xi}[f(x,\xi)]=0, where the mapping f:nnf\,:\,\mathbb{R}^{n}\rightarrow\mathbb{R}^{n} is given by

fi(x,ξ)=xi5sin(ij=1nxj+ξ),\begin{array}[]{l}f_{i}(x,\xi)=x_{i}-5\sin(i\sum\limits_{j=1}^{n}x_{j}+\xi),\end{array} (15)

i=1,2,,ni=1,2,\ldots,n. We employ the GRSAA differentiable homotopy method to solve the SSE (15) under different n{3,4,5,6,7,8,9,10}n\in\{3,4,5,6,7,8,9,10\}. For numerical comparisons, we also apply the GRSAA simplicial homotopy method to solve the same problems. In the implementation of the GRSAA simplicial homotopy method, we set the initial grid size of the D2D_{2}-triangulation ω¯0=1\bar{\omega}_{0}=1, the factor to refine the grid size r=0.5r_{\ell}=0.5 for all \ell, t0=1t_{0}=1, t1=0.5t_{1}=0.5, t=1/(1+7000)t_{\ell}=1/(1+7000\ell) for =2,3,\ell=2,3,\ldots. Additionally, we choose the sample size q=500q_{\ell}=500\ell for =1,2,\ell=1,2,\ldots. Clearly, t0t_{\ell}\rightarrow 0 and qq_{\ell}\rightarrow\infty as \ell\rightarrow\infty. Moreover, to make the comparisons more convincing, tt_{\ell}, =0,1,\ell=0,1,\ldots, for the GRSAA differentiable homotopy method are consistent with those for the GRSAA simplicial homotopy method and both methods start from the same randomly generated starting point. Each case with different nn is tested for 20 times and the average computational costs are reported in Table 1, where “ITER” is the average number of iterations, “TIME” is the average computational time (in seconds), “GRSAA-DH” and “GRSAA-SH” represent respectively the GRSAA differentiable homotopy method and GRSAA simplicial homotopy method, and “RATIO” stands for the ratio of the numerical results of GRSAA-DH to GRSAA-SH.888Note that if the computational time exceeds 1000 and the number of iterations is larger than 10710^{7}, we record the value as “INF”.

Table 1: Average Computational Cost of Two Methods
nn GRSAA-DH GRSAA-SH RATIO (%)
ITER TIME ITER TIME ITER TIME
3 78 2.72 1181 2.58 6.60 105.42
4 96 3.76 3707 11.30 2.59 33.27
5 98 4.71 5738 12.88 1.71 36.56
6 105 5.49 14797 32.65 0.71 16.81
7 108 6.54 34011 59.83 0.32 10.93
8 123 8.53 78253 99.11 0.16 8.61
9 129 9.45 202596 198.30 0.06 4.76
10 272 16.33 INF INF - -

From the columns of Table 1, we find that both the average number of iterations (ITER) and the average computational time (TIME) become greater and greater with the increasing of nn for the two methods. This result coincides with our intuition, that is, the larger the problem is, the harder it is for the methods to solve. From the rows of Table 1, we observe that the GRSAA differentiable homotopy method significantly outperforms the GRSAA simplicial homotopy method both in the number of iterations and computational time when n>3n>3. The advantage of the GRSAA-DH method over the GRSAA-SH method becomes more remarkable when the number of variables nn is higher, which implies that the GRSAA-DH method is relatively less sensitive to nn.

Figure 4 presents the changes of the maximal, average and minimal computational time in nn among the 20 tests for the two methods. Figure 5 illustrates the changes of the maximal, average and minimal number of iterations in nn for the two methods.

Refer to caption
Refer to caption
Figure 4: Computational Time for GRSAA-DH and GRSAA-SH
Refer to caption
Refer to caption
Figure 5: Number of Iterations for GRSAA-DH and GRSAA-SH

Regarding the GRSAA-DH method, one can observe from Figure 4 and Figure 5 that as n9n\leq 9, the difference between the maximal computational time and the minimal computational time is fairly small. As n=10n=10, there are some extreme values that enlarge the gap between the maximal value and the minimal value, but the number of such extreme cases are quite few, since the average computational time is just about 12 seconds, which is near the minimal computational time of 10 seconds. This indicates that the computational time for most tests is between 10 and 12 seconds. However, for the GRSAA-SH method, when n>6n>6, one can easily see that the gaps among the maximal, average and minimal computational time are much larger. Similar phenomena can be found for the number of iterations. These numerical results show that the GRSAA-DH method performs much more stable than the GRSAA-SH method.

To further demonstrate the advantage of the GRSAA-DH method, we have used the method to solve several large-scale SSEs, for which the GRSAA-SH method fails to find a reasonable approximate solution after the computational time is over 5000 seconds. We have run the GRSAA-DH method on each test 10 times. The numerical results are reported in Table 2 and Figure 6, which once again certify that the GRSAA differentiable homotopy method is able to effectively and efficiently solve larger-scale problems.

Table 2: Computational Time of GRSAA-DH for Large-Scale Cases
      nn       MAX       MIN       AVER
      15       55.99       19.65       36.389
      16       476.85       52.27       219.771
      17       774.83       102.06       416.047
      18       1367.73       110.96       559.361
      19       1786.92       244.85       729.847
      20       1247.82       464.09       821.583
      21       1571.14       325.18       985.259
      22       1811.67       380.60       1230.041
      23       2133.52       597.04       1317.084
      24       2214.6       911.96       1396.252
      25       2826.06       1214.31       1479.845
Refer to caption
Figure 6: Computational Time for GRSAA-DH for Different Cases

5.3 A Stochastic Variational Inequality

This subsection intends to solve a stochastic variational inequality (SVI) problem,

(yx)𝔼ξ[f(x,ξ)]0,y𝕐,\begin{array}[]{ll}(y-x)^{\top}\mathbb{E}_{\xi}[f(x,\xi)]\leq 0,&\forall\,y\in\mathbb{Y},\end{array} (16)

where f:nnf\,:\,\mathbb{R}^{n}\rightarrow\mathbb{R}^{n} is a continuously differentiable mapping with fi(x,ξ)=xiecos(ij=1nxj+ξ)f_{i}(x,\xi)=x_{i}-e^{\text{cos}(i\sum_{j=1}^{n}x_{j}+\xi)} for i=1,2,,ni=1,2,\ldots,n, and 𝕐={y=(y1,,yn)n|10yi10,i=1,2,,n}\mathbb{Y}=\{y=(y_{1},\ldots,y_{n})^{\top}\in\mathbb{R}^{n}\,|\,-10\leq y_{i}\leq 10,\,\forall\,i=1,2,\ldots,n\}. To get a GRSAA differentiable homotopy system for the SVI problem, we constitute the optimization problem,

maxyny𝔼[f(x,ξ)]s.t.Cyb,\begin{array}[]{ll}\max\limits_{y\in\mathbb{R}^{n}}&y^{\top}\mathbb{E}[f(x,\xi)]\\ \text{s.t.}&Cy\leq b,\end{array}

where C=[InIn]2n×nC=[I_{n}\;-I_{n}]^{\top}\in\mathbb{R}^{2n\times n} and b=10(e,e)2nb=10(e^{\top},-e^{\top})^{\top}\in\mathbb{R}^{2n} with Inn×nI_{n}\in\mathbb{R}^{n\times n} being an identity matrix and e=(1,1,,1)ne=(1,1,\ldots,1)^{\top}\in\mathbb{R}^{n}. Applying the optimality conditions to this optimization problem, we derive with a fixed point argument an equivalent problem to the SVI problem (16),

𝔼[f(x,ξ)]Cλ=0,Cxb+z=0,λz=0,λ0,z0.\begin{array}[]{l}\mathbb{E}[f(x,\xi)]-C^{\top}\lambda=0,\\ \vskip 6.0pt\cr Cx-b+z=0,\\ \vskip 6.0pt\cr\lambda z=0,\quad\lambda\geq 0,\quad z\geq 0.\end{array} (17)

We have generated numerous samples for the stochastic variable ξ\xi and applied the SAA, 1Ni=1Nf(x,ξi)\dfrac{1}{N}\sum\limits_{i=1}^{N}f(x,\xi_{i}), to estimate the expected value 𝔼[f(x,ξ)]\mathbb{E}[f(x,\xi)]. The corresponding GRSAA differentiable homotopy system to the SSE (17) is as follows:

(1t)(d(x,t)Cλ)t(xx0)=0,Cxb+z=0,λz=tκ0,\begin{array}[]{l}\mathbb{(}1-t)(d(x,t)-C^{\top}\lambda)-t(x-x^{0})=0,\\ \vskip 6.0pt\cr Cx-b+z=0,\\ \vskip 6.0pt\cr\lambda z=t^{\kappa_{0}},\end{array} (18)

where x0x^{0} is an interior point in the compact convex set 𝕐\mathbb{Y} and d(x,t)d(x,t) is defined as in Section 2. After a transformation of variables, that is, λi(w)=(wi2+4t+wi2)κ0\lambda_{i}(w)=\left(\dfrac{\sqrt{w_{i}^{2}+4t}+w_{i}}{2}\right)^{\kappa_{0}} and zi(w)=(wi2+4twi2)κ0\quad z_{i}(w)=\left(\dfrac{\sqrt{w_{i}^{2}+4t}-w_{i}}{2}\right)^{\kappa_{0}} for i=1,2,,2ni=1,2,\ldots,2n, the third group of equations of the system (18) vanishes and the system (18) becomes an equivalent form,

(1t)(d(x,t)Cλ(w))t(xx0)=0,Cxb+z(w)=0.\begin{array}[]{l}\mathbb{(}1-t)(d(x,t)-C^{\top}\lambda(w))-t(x-x^{0})=0,\\ \vskip 6.0pt\cr Cx-b+z(w)=0.\end{array} (19)

Clearly, the efficiency of the proposed method depends on the number of variables nn, the sample size NN, and the number of divisions LL. Next, we study the impact of these factors on the performance of the GRSAA differentiable homotopy method.

  • First, the GRSAA-DH method has been used to solve the system (19) under different pairs of (N,n)(N,n), where N{104,105,106}N\in\{10^{4},10^{5},10^{6}\} and n{1,2,3}n\in\{1,2,3\}. The number of divisions for the method is fixed to be L=N/2L=N/2. To make a numerical comparison, we have also employed a standard differentiable homotopy method, which can be considered as a special case of the GRSAA-DH method without a gradual reinforcement process, that is, L=1L=1, to solve the same problems. Each case has been run 10 times and the average computational time is reported in Table 3.

    Table 3: Numerical Results for the Two Methods
    GRSAA-DH Stardard Homotopy
    nn N=104N=10^{4}
    1 1.03 1.18
    2 3.94 5.64
    3 6.57 10.98
    nn N=105N=10^{5}
    1 8.61 10.95
    2 35.61 49.32
    3 61.91 100.35
    nn N=106N=10^{6}
    1 80.20 95.55
    2 358.47 474.51
    3 619.35 1052.27

    One can see from Table 3 that the computational costs of the two methods become larger and larger with the growing of sample size and the number of variables. For each fixed nn, the computational time of the GRSAA-DH method increases approximately linearly with NN. For example, when n=2n=2 and NN changes from 10410^{4} to 10510^{5}, the computational time of the GRSAA-DH method increases from 3.94 to 35.61. Comparing the last two columns of Table 3, one can observe that the GRSAA-DH method is more competitive than the standard differentiable homotopy method, which verifies that the gradual reinforcement process indeed makes a significant difference.

  • Second, we want to explore how the computational time is influenced by the number of divisions LL and determine an appropriate value of LL for the GRSAA differentiable homotopy method. In theory, LL can be chosen as any value that is not larger than the sample size NN. Nonetheless, it is intuitive that a too large or too small value of LL may cause a low numerical efficiency, which has been partially acknowledged in Table 3, since the standard differentiable homotopy method is same as the GRSAA-DH method with L=1L=1. In order to achieve a better efficiency, it is necessary to find a suitable value of LL when implementing the GRSAA-DH method. However, finding an optimal value of LL is an NP-hard problem in theory and can only be realized through numerical experiments. We have implemented the GRSAA-DH method with different choices of LL to solve the system (19) and plotted the changes of the computational time in various values of LL in Figures 7, 8 and 9. In these experiments, N{104,105,106}N\in\{10^{4},10^{5},10^{6}\} and n{1,2,3}n\in\{1,2,3\}.

Refer to caption
Refer to caption
Refer to caption
Figure 7: n=1n=1 and N=104,105,106N=10^{4},10^{5},10^{6}
Refer to caption
Refer to caption
Refer to caption
Figure 8: n=2n=2 and N=104,105,106N=10^{4},10^{5},10^{6}
Refer to caption
Refer to caption
Refer to caption
Figure 9: n=3n=3 and N=104,105,106N=10^{4},10^{5},10^{6}

From Figures 7, 8 and 9, we find that the computational time presents a highly analogous trend for different cases especially with the relatively high number of variables and a large batch of samples. More specifically, the computational time always decreases as L=1L=1 and reaches a local minimum about L=0.55NL=0.55N. Notice that the local minimum also has a sufficiently competitive performance over the entire possible choices of LL. Hence, when applying the GRSAA-DH method, one can roughly determine the “best” number of divisions according to L=0.55NL=0.55N, which surly performs much better than the standard differentiable homotopy method.

6 Conclusions

This paper has exploited the sample-average-approximation (SAA) scheme to approximate a stochastic system of equations (SSE) and developed a gradually reinforced sample-average-approximation (GRSAA) differentiable homotopy method to solve the SSE with very large sample size. By introducing a sequence of continuously differentiable functions of the homotopy parameter ranging between zero and one, we have established a continuously differentiable homotopy system, which is able to gradually increase the sample size as the homotopy parameter decreases. The solution set of the homotopy system contains an everywhere smooth path, which starts from an arbitrarily given point and ends at a satisfactory approximate solution to the original SSE.

The GRSAA differentiable homotopy method provides an effective link between a standard differentiable homotopy method and a gradually reinforced SAA scheme. The proposed method is able to greatly reduce the computational cost for a solution to the SSE with large sample size and retain the inherent property of global convergence with a standard homotopy method. To make numerical comparisons, we have carried out extensive numerical tests. The numerical results further confirm that two main features of the GRSAA differentiable homotopy method, gradual reinforcement in sample size and differentiability, can significantly enhance numerical effectiveness and efficiency.

References

  • [1] Allgower, E.L., Georg, K.: Numerical continuation methods: an introduction, vol. 13. Springer Science & Business Media (2012)
  • [2] Awoniyi, S.A., Todd, M.J.: An efficient simplicial algorithm for computing a zero of a convex union of smooth functions. Mathematical Programming 25(1), 83–108 (1983)
  • [3] Brown, D.J., Demarzo, P.M., Eaves, B.C.: Computing equilibria when asset markets are incomplete. Econometrica 64(1), 1–27 (1996)
  • [4] Byrd, R.H., Chin, G.M., Nocedal, J., Wu, Y.: Sample size selection in optimization methods for machine learning. Mathematical Programming 134(1), 127–155 (2012)
  • [5] Chen, L., Liu, Y., Yang, X., Zhang, J.: Stochastic approximation methods for the two-stage stochastic linear complementarity problem. SIAM Journal on Optimization 32(3), 2129–2155 (2022)
  • [6] Chen, X., Shen, J.: Dynamic stochastic variational inequalities and convergence of discrete approximation. SIAM Journal on Optimization 32(4), 2909–2937 (2022)
  • [7] Chen, X., Wets, R.J.B., Zhang, Y.: Stochastic variational inequalities: residual minimization smoothing sample average approximations. SIAM Journal on Optimization 22(2), 649–673 (2012)
  • [8] Chen, Y., Dang, C.: A differentiable homotopy method to compute perfect equilibria. Mathematical Programming pp. 1–33 (2019)
  • [9] Dang, C.: The D1{D}_{1}-triangulation of rn for simplicial algorithms for computing solutions of nonlinear equations. Mathematics of Operations Research 16(1), 148–161 (1991)
  • [10] Dang, C.: The D2{D}_{2}-triangulation for simplicial homotopy algorithms for computing solutions of nonlinear equations. Mathematical Programming 59(1-3), 307–324 (1993)
  • [11] Dang, C., Herings, P.J.J., Li, P.: An interior-point differentiable path-following method to compute stationary equilibria in stochastic games. INFORMS Journal on Computing 34(3), 1403–1418 (2022)
  • [12] Duchi, J., Hazan, E., Singer, Y.: Adaptive subgradient methods for online learning and stochastic optimization. Journal of Machine Learning Research 12(7) (2011)
  • [13] Eaves, B.C.: The linear complementarity problem. Management Science 17(9), 612–634 (1971)
  • [14] Eaves, B.C.: Homotopies for computation of fixed points. Mathematical Programming 3(1), 1–22 (1972)
  • [15] Eaves, B.C., Schmedders, K.: General equilibrium models and homotopy methods. Journal of Economic Dynamics and Control 23(9), 1249–1279 (1999)
  • [16] Eibelshäuser, S., Klockmann, V., Poensgen, D., von Schenk, A.: The logarithmic stochastic tracing procedure: A homotopy method to compute stationary equilibria of stochastic games. INFORMS Journal on Computing 35(6), 1511–1526 (2023)
  • [17] Facchinei, F., Pang, J.S.: Finite-dimensional variational inequalities and complementarity problems. Springer Science & Business Media (2007)
  • [18] Ferris, M.C., Mangasarian, O.L., Pang, J.S.: Complementarity: Applications, algorithms and extensions, vol. 50. Springer Science & Business Media (2013)
  • [19] Gürkan, G., Ozge, A.Y., Robinson, S.M.: Sample-path solution of stochastic variational inequalities. Mathematical Programming 84(2), 313–333 (1999)
  • [20] Herings, P.J.J.: Two simple proofs of the feasibility of the linear tracing procedure. Economic Theory 15(2), 485–490 (2000)
  • [21] Herings, P.J.J., Kubler, F.: Computing equilibria in finance economies. Mathematics of operations research 27(4), 637–646 (2002)
  • [22] Herings, P.J.J., Peeters, R.: A differentiable homotopy to compute Nash equilibria of nn-person games. Economic Theory 18(1), 159–185 (2001)
  • [23] Herings, P.J.J., Peeters, R.: Homotopy methods to compute equilibria in game theory. Economic Theory 42(1), 119–156 (2010)
  • [24] Herings, P.J.J., Peeters, R.J.: Stationary equilibria in stochastic games: Structure, selection, and computation. Journal of Economic Theory 118, 32–60 (2004)
  • [25] Herings, P.J.J., Schmedders, K.: Computing equilibria in finance economies with incomplete markets and transaction costs. Economic Theory 27(3), 493–512 (2006)
  • [26] Hu, J., Homem-de Mello, T., Mehrotra, S.: Sample average approximation of stochastic dominance constrained programs. Mathematical Programming 133(1-2), 171–201 (2012)
  • [27] Hu, Y., Chen, X., He, N.: Sample complexity of sample average approximation for conditional stochastic optimization. SIAM Journal on Optimization 30(3), 2103–2133 (2020)
  • [28] Huang, Z.H., Li, Y.F., Miao, X.: Finding the least element of a nonnegative solution set of a class of polynomial inequalities. SIAM Journal on Matrix Analysis and Applications 44(2), 530–558 (2023)
  • [29] Iusem, A.N., Jofré, A., Thompson, P.: Incremental constraint projection methods for monotone stochastic variational inequalities. Mathematics of Operations Research 44(1), 236–263 (2019)
  • [30] Jiang, H., Xu, H.: Stochastic approximation approaches to the stochastic variational inequality problem. IEEE Transactions on Automatic Control 53(6), 1462–1475 (2008)
  • [31] Jiang, J., Sun, H., Zhou, B.: Convergence analysis of sample average approximation for a class of stochastic nonlinear complementarity problems: from two-stage to multistage. Numerical Algorithms 89(1), 167–194 (2022)
  • [32] Judd, K.L., Renner, P., Schmedders, K.: Finding all pure-strategy equilibria in games with continuous strategies. Quantitative Economics 3(2), 289–331 (2012)
  • [33] Kellogg, R.B., Li, T.Y., Yorke, J.: A constructive proof of the brouwer fixed-point theorem and computational results. SIAM Journal on Numerical Analysis 13(4), 473–483 (1976)
  • [34] Kleywegt, A.J., Shapiro, A., Homem-de Mello, T.: The sample average approximation method for stochastic discrete optimization. SIAM Journal on Optimization 12(2), 479–502 (2002)
  • [35] Kubler, F., Schmedders, K.: Computing equilibria in stochastic finance economies. Computational Economics 15(1-2), 145–172 (2000)
  • [36] Lee, K., Tang, X.: On the polyhedral homotopy method for solving generalized nash equilibrium problems of polynomials. Journal of Scientific Computing 95(1), 13 (2023)
  • [37] Li, P., Dang, C.: An arbitrary starting tracing procedure for computing subgame perfect equilibria. Journal of Optimization Theory and Applications 186(2), 667–687 (2020)
  • [38] MERRILL, O.H.: Applications and extensions of an algorithm that computes fixed points ofcertain upper semi-continuous point to set mappings. Ph.D. thesis, University of Michigan (1972)
  • [39] Na, S., Anitescu, M., Kolar, M.: Inequality constrained stochastic nonlinear optimization via active-set sequential quadratic programming. Mathematical Programming pp. 1–75 (2023)
  • [40] Nemirovski, A., Juditsky, A., Lan, G., Shapiro, A.: Robust stochastic approximation approach to stochastic programming. SIAM Journal on Optimization 19(4), 1574–1609 (2009)
  • [41] Pflug, G.C.: Optimization of Stochastic Models. Springer US (1996)
  • [42] Rockafellar, R.T., Sun, J.: Solving monotone stochastic variational inequalities and complementarity problems by progressive hedging. Mathematical Programming 174(1-2), 453–471 (2019)
  • [43] Rockafellar, R.T., Wets, R.J.B.: Variational analysis, vol. 317. Springer Science & Business Media (2009)
  • [44] Scarf, H.: The approximation of fixed points of a continuous mapping. SIAM Journal on Applied Mathematics 15(5), 1328–1343 (1967)
  • [45] Schmedders, K.: Computing equilibria in the general equilibrium model with incomplete asset markets. Journal of Economic Dynamics and Control 22(8-9), 1375–1401 (1998)
  • [46] Shapiro, A., Homem-de Mello, T.: A simulation-based approach to two-stage stochastic programming with recourse. Mathematical Programming 81(3), 301–325 (1998)
  • [47] Shapiro, A., Homem-de Mello, T.: On the rate of convergence of optimal solutions of Monte Carlo approximations of stochastic programs. SIAM Journal on Optimization 11(1), 70–86 (2000)
  • [48] Todd, M.J.: The computation of fixed points and applications. Lecture Notes in Economics and Mathematical Systems, Springer, Berlin 124 (1976)
  • [49] Van Den Elzen, A., Van Der Laan, G., Talman, D.: An adjustment process for an economy with linear production technologies. Mathematics of operations research 19(2), 341–351 (1994)
  • [50] Wang, G., Wei, X., Yu, B., Xu, L.: An efficient proximal block coordinate homotopy method for large-scale sparse least squares problems. SIAM Journal on Scientific Computing 42(1), A395–A423 (2020)
  • [51] Zhan, Y., Dang, C.: A smooth homotopy method for incomplete markets. Mathematical Programming 190(1-2), 585–613 (2021)
  • [52] Zhan, Y., Li, P., Dang, C.: A differentiable path-following algorithm for computing perfect stationary points. Computational Optimization and Applications 76, 571–588 (2020)
  • [53] Zhang, J., Xiao, Q., Li, L.: Solution space exploration of low-thrust minimum-time trajectory optimization by combining two homotopies. Automatica 148, 110798 (2023)
  • [54] Zhou, Z., Yu, B.: A smoothing homotopy method for variational inequality problems on polyhedral convex sets. Journal of Global Optimization 58(1), 151–168 (2014)