This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Solvability of Multistage Pseudomonotone Stochastic Variational Inequalities111This work was supported by the National Natural Science Foundation of China (Grant No. 11771244, 12171271).

Xingbang Cui cxb18@mails.tsinghua.edu.cn Department of Mathematical Sciences, Tsinghua University, Beijing 100084, China Jie Sun jie.sun@curtin.edu.au School of Business, National University of Singapore, Singapore 119245 School of EECMS, Curtin University, Perth WA6845, Australia Liping Zhang lipingzhang@mail.tsinghua.edu.cn
Abstract

This paper focuses on the solvability of multistage pseudomonotone stochastic variational inequalities (SVIs). On one hand, some known solvability results of pseudomonotone deterministic variational inequalities cannot be directly extended to pseudomonotone SVIs, so we construct the isomorphism between both and then establish theoretical results on the existence, convexity, boundedness and compactness of the solution set for pseudomonotone SVIs via such an isomorphism. On the other hand, the progressive hedging algorithm (PHA) is an important and effective method for solving monotone SVIs, but it cannot be directly used to solve nonmonotone SVIs. We propose some sufficient conditions on the elicitability of pseudomonotone SVIs, which opens the door for applying Rockafellar’s elicited PHA to solve pseudomonotone SVIs. Numerical results on solving a pseudomonotone two-stage stochastic market optimization problem and randomly generated two-stage pseudomonotone linear complementarity problems are presented to show the efficiency of the elicited PHA for solving pseudomonotone SVIs.

keywords:
Pseudomonotonicity , multistage stochastic optimization , stochastic variational inequalities , progressive hedging algorithm , elicited monotonicity
MSC:
[2020] 90C15 , 90C33 , 90C30 , 65K15

1 Introduction

Recently, Rockafellar and Wets [24] developed the multistage stochastic variational inequality (SVI) model, which may incorperate recourse decisions and stagewise disclosure of information. The model provides a unified framework for describing the optimal conditions of multistage stochastic optimization problems and stochastic equilibrium problems. Rockafellar and Sun [26, 27] proposed progressive hedging algorithms (PHAs) for solving monotone multistage SVI problems and monotone Lagrangian multistage SVI problems. In particular, their algorithms converge linearly in the linear-quadratic setting. Besides, Chen et al [2] formulated the two-stage SVI as a two-stage stochastic program with recourse by the expected residual minimization procedure, and solved this stochastic program via the sample average approximation (SAA). Pang et al [20] systematically studied the theory and the best response method for a two-stage nonlinear stochastic game model. Some related work to SVIs, multistage stochastic equilibrium problems and PHAs can be found in, e.g., [4, 5, 8, 9, 10, 11, 12, 13, 15, 16, 17, 18, 23, 34, 36, 37].

As indicated in [14, 26], the aforementioned literatures on PHAs are mostly limited to the monotone mappings since the PHA is essentially a variety of the proximal point algorithm (PPA) for monotone operators [29]. In practice, however, nonmonotone SVIs such as pseudomonotone SVIs arise frequently [14] in various applications such like the competitive exchange economy problem, the stochastic fractional problem and the stochastic product pricing problem. The investigations about solution methods for single-stage pseudomonotone SVIs can be found in [10, 14, 35]. A natural question is what about the solvability of the multistage pseudomonotone SVI.

In this paper, we investigate the solvability of the multistage pseudomonotone SVI including solution basic theory and solution method. On one hand, some remarkable theoretical results on pseudomonotone deterministic VIs can be found in [7]. However, these known results cannot be directly extended to the multistage pseudomonotone SVI. So, we design an isomorphism between the pseudomonotone SVI and the pseudomonotone deterministic VI. Via such an isomorphism, we obtain some properties of its solution set including the nonemptiness, compactness, boundedness and convexity. On the other hand, the PHA is a very effective method for solving monotone SVIs, but it cannot be directly used to solve the multistage pseudomonotone SVI. Fortunately, Rockafellar recently introduced the notion of elicitable monotonicity (see definition in Section 2) [26] in an attempt to extend the PPA and its varieties from monotone mappings to certain nonmonotone mappings. Motivated by his work, we devote to investigate the elicitability of the multistage pseudomonotone SVIs for the purpose of applying the PHA to solve the multistage pseudomonotone SVI, which opens the door for applying Rockafellar’s elicited PHA [26] to solve pseudomonotone SVIs. As long as the monotonicity of the multistage pseudomonotone SVI can be elicited, the elicited PHA in [33, 38], which is a specialization of Rockafellar’s progressive decoupling algorithm [26], can be applied to solve the multistage pseudomonotone SVI. Our numerical results indicate that such PHA works pretty well for linear pseudomonotone complementarity problems of ordinary size (i.e., several hundreds of variables and scenarios).

The main contribution of this paper is summarized as follows.

  • 1.

    The isomorphism between the pseudomonotone SVI and the pseudomonotone deterministic VI is presented, which stands as a stepping stone in studying the solvability of pseudomonotone SVIs.

  • 2.

    Some properties of the solution set of pseudomonotone SVIs, such as the existence, compactness, boundedness and convexity of solutions, are established.

  • 3.

    Some criteria to identify the elicitability of the pseudomonotone SVIs are provided, which ensure Rockafellar’s elicited PHA can be applied to solve the pseudomonotone SVIs.

The rest of this paper is organized as follows. We present the formulation and some basic notions of the multistage SVI in Section 2. The isomorphism between the pseudomonotone SVI and the pseudomonotone deterministic VI is introduced in Section 3. We derive some results of the solution sets for the pseudomonotone multistage SVI in Section 4. Some criteria for the elicitability of a pseudomonotone SVI are deduced in Section 5. We demonstrate the effectiveness of the elicited PHA by various numerical experiments in Section 6. The paper is concluded in Section 7.

Notations: For any positive integer mm, m\mathbb{R}^{m} denotes the mm-dimensional Euclidean space and m\mathcal{L}_{m} denotes the the mm-dimensional Hilbert space, with +m={xm:x0}\mathbb{R}^{m}_{+}=\{x\in\mathbb{R}^{m}:~{}x\geq 0\}. Let m×m\mathbb{R}^{m\times m} denote the set of real matrices of mm rows and mm columns. For Qm×mQ\in\mathbb{R}^{m\times m}, tr(Q)(Q) denotes the trace of QQ. Given sets S,US,U, intSS denotes the interior of SS, bdSS denotes the boundary of SS, ri(S)(S) denotes the relative interior of SS, and S\US\backslash U is the set {x:xS,xU}\{x:x\in S,\ x\notin U\}. We use S={d:x+τdS,xS,τ0}S_{\infty}=\{d:x+\tau d\in S,~{}\forall x\in S,\,\tau\geq 0\} to designate the recession cone of SS and use S={d:v,d0,vS}S^{*}=\{d:\left<v,d\right>\geq 0,~{}\forall v\in S\} to designate the dual cone of SS. We also use int(S)(S_{\infty})^{*} to stand for int((S))((S_{\infty})^{*}).

2 Preliminaries

2.1 Formulation of the multistage SVI

Consider an N-stage sequence

x1,ξ1,x2,ξ2,,xN,ξN,x_{1},\xi_{1},x_{2},\xi_{2},\ldots,x_{N},\xi_{N},

where xknkx_{k}\in\mathbb{R}^{n_{k}} is the decision vector at the kk-th stage and ξkΞk\xi_{k}\in\Xi_{k} is a random vector with Ξk\Xi_{k} being its support and becoming known only after xkx_{k} is determined. Let ξ=(ξ1,ξ2,,ξN)\xi=(\xi_{1},\xi_{2},\ldots,\xi_{N}) be the random vector defined on the finite sample space Ξ=Ξ1×Ξ2××ΞN\Xi=\Xi_{1}\times\Xi_{2}\times\cdots\times\Xi_{N}, where each realization of ξ\xi has a probability p(ξ)>0p(\xi)>0, and these probabilities add up to 11.

Throughout this paper, define n=k=1Nnkn=\sum\limits_{k=1}^{N}n_{k} and let n\mathcal{L}_{n} denote the Hilbert space consisting of the mapping from Ξ\Xi to n\mathbb{R}^{n}

x():ξx(ξ)=(x1(ξ)T,x2(ξ)T,,xN(ξ)T)T,x(\cdot):\xi\mapsto x(\xi)=(x_{1}(\xi)^{T},x_{2}(\xi)^{T},\ldots,x_{N}(\xi)^{T})^{T},

where xk(ξ)Tx_{k}(\xi)^{T} denotes the transpose of xk(ξ)x_{k}(\xi) for k=1,2,,Nk=1,2,\ldots,N. The inner product of n\mathcal{L}_{n} is defined as

x(),w()=ξΞp(ξ)k=1Nxk(ξ),wk(ξ),x(),w()n,\langle x(\cdot),w(\cdot)\rangle=\sum\limits_{\xi\in\Xi}p(\xi)\sum\limits_{k=1}^{N}\left<x_{k}(\xi),w_{k}(\xi)\right>,\quad x(\cdot),w(\cdot)\in\mathcal{L}_{n}, (1)

where xk(ξ),wk(ξ)\left<x_{k}(\xi),w_{k}(\xi)\right> is the Euclidean inner product of xk(ξ)x_{k}(\xi) and wk(ξ)w_{k}(\xi). Further, by restricting mapping x()nx(\cdot)\in\mathcal{L}_{n} to the following closed subspace

𝒩n={x():xk(ξ)does not depend onξk,,ξN},\mathcal{N}_{n}=\{x(\cdot):x_{k}(\xi)\ \text{does not depend on}\ \xi_{k},\ldots,\xi_{N}\},

we introduce the nonanticipativity constraint on n\mathcal{L}_{n}. That is, x()𝒩nx(\cdot)\in\mathcal{N}_{n} means that xk(ξ)x_{k}(\xi) will be influenced by ξ1,,ξk1\xi_{1},\ldots,\xi_{k-1}, but not ξk,,ξN\xi_{k},\ldots,\xi_{N}. The orthogonal complement of 𝒩n\mathcal{N}_{n} is denoted by

n={w()n:Eξ|ξ1,,ξk1wk(ξ)=0,k=1,,N},\mathcal{M}_{n}=\{w(\cdot)\in\mathcal{L}_{n}:E_{\xi|\xi_{1},\ldots,\xi_{k-1}}w_{k}(\xi)=0,\ k=1,\ldots,N\}, (2)

where Eξ|ξ1,,ξk1wk(ξ)E_{\xi|\xi_{1},\ldots,\xi_{k-1}}w_{k}(\xi) is the conditional expectation given (ξ1,,ξk1)(\xi_{1},\ldots,\xi_{k-1}). Moreover, w()w(\cdot) corresponds to the nonanticipativity multiplier in [24], which is understood in stochastic programming as furnishing the shadow price of information. As we will see later, w()w(\cdot) enables decomposition into a separate problem for each secnario ξ\xi. In addition, the closed convex set 𝒞n\mathcal{C}\subset\mathcal{L}_{n} is denoted by

𝒞={x():x(ξ)C(ξ),ξΞ},\mathcal{C}=\{x(\cdot):x(\xi)\in C(\xi),\ \forall\xi\in\Xi\}, (3)

where C(ξ)C(\xi) is a nonempty closed convex set in n\mathbb{R}^{n}. The mapping :nn\mathcal{F}:\mathcal{L}_{n}\rightarrow\mathcal{L}_{n} is defined as

(x()):ξF(x(ξ),ξ),ξΞ,\mathcal{F}(x(\cdot)):\xi\mapsto F(x(\xi),\xi),\ \forall\xi\in\Xi,

where F(,ξ)F(\cdot,\xi) is a function from n\mathbb{R}^{n} to n\mathbb{R}^{n}. The NN-stage SVI in basic form can be expressed as to find x()x(\cdot) such that

(x())N𝒞𝒩n(x()),-\mathcal{F}(x(\cdot))\in N_{\mathcal{C}\cap\mathcal{N}_{n}}(x(\cdot)), (4)

where N𝒞𝒩n(x())N_{\mathcal{C}\cap\mathcal{N}_{n}}(x(\cdot)) is the normal cone of 𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n} at x()x(\cdot). In other words,

N𝒞𝒩n(x())={v():v(),y()x()0,y()𝒞𝒩n}.N_{\mathcal{C}\cap\mathcal{N}_{n}}(x(\cdot))=\{v(\cdot):\left<v(\cdot),y(\cdot)-x(\cdot)\right>\leq 0,\ \forall y(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n}\}.

For convenience, we use SVI(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) to denote (4) and use SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) to denote the corresponding solution set. The extensive form of SVI (4) can be formulated as to find x()𝒩nx(\cdot)\in{\cal N}_{n} and w()nw(\cdot)\in{\cal M}_{n} such that

F(x(ξ),ξ)w(ξ)NC(ξ)(x(ξ)),ξΞ.-F(x(\xi),\xi)-w(\xi)\in N_{C(\xi)}(x(\xi)),\quad\forall\xi\in\Xi. (5)

The following theorem given in [24] exhibits the relationship between (4) and (5).

Theorem 1

If x()x(\cdot) solves (5), then x()x(\cdot) solves (4). Conversely, assume that x()x(\cdot) solves (4), then x()x(\cdot) is sure to solve (5) if there exists some x^()𝒩n\hat{x}(\cdot)\in\mathcal{N}_{n} such that x^(ξ)riC(ξ)\hat{x}(\xi)\in riC(\xi) for all ξΞ\xi\in\Xi (This is called the constraint qualification).

2.2 Pseudomonotonicity

We recall the concept of pseudomonotone mapping and its properties.

Definition 1

(a)(a) A mapping F:KmmF:K\subset\mathbb{R}^{m}\to\mathbb{R}^{m} is said to be pseudomonotone if for all x,yKx,y\in K,

yx,F(x)0yx,F(y)0.\left<y-x,F(x)\right>\geq 0\Rightarrow\left<y-x,F(y)\right>\geq 0.

(b)(b) Let 𝒦={x()|x(ξ)K(ξ),ξΞ}\mathcal{K}=\{x(\cdot)|x(\xi)\in K(\xi),\ \forall\xi\in\Xi\} be a closed convex set in n\mathcal{L}_{n}, where K(ξ)K(\xi) is a closed convex set in n\mathbb{R}^{n} for all ξΞ\xi\in\Xi. Then a mapping :𝒦nn\mathcal{F}:\mathcal{K}\subset\mathcal{L}_{n}\to\mathcal{L}_{n} is said to be pseudomonotone on 𝒦\mathcal{K} if for all x(),y()𝒦x(\cdot),y(\cdot)\in\mathcal{K},

y()x(),(x())0y()x(),(y())0.\left<y(\cdot)-x(\cdot),\mathcal{F}(x(\cdot))\right>\geq 0\Rightarrow\left<y(\cdot)-x(\cdot),\mathcal{F}(y(\cdot))\right>\geq 0.
Remark 1

Obviously, a monotone mapping FF must be pseudomonotone while the converse is not necessarily true. The counterexample is F(x)=1/xF(x)=1/x and K=[1,2]K=[1,2].

Some properties of pseudomonotone functions were given in [1].

Lemma 1

Let FF be a continuously differentiable function defined on an open convex set SmS\subset\mathbb{R}^{m}. Assume that FF is pseudomonotone on SS, then the following statements hold.

(a)(a) Let DF(x)DF(x) denote the Jacobian of FF at xx.

xS,um,u,F(x)=0u,DF(x)u0.x\in S,u\in\mathbb{R}^{m},\left<u,F(x)\right>=0\Rightarrow\left<u,DF(x)u\right>\geq 0. (6)

Further, if F(x)0F(x)\neq 0 for all xSx\in S, then the statement (6) is also sufficient.

(b)(b) Given xSx\in S, the Jacobian DF(x)DF(x) has at most one negative eigenvalue, where an eigenvalue with multiplicity k(k>1)k(k>1) is counted as kk eigenvalues.

In order to give some criteria for elicitable monotonicity, we recall some basic results about matrices which were given in [31].

Lemma 2

Let A,Bm×mA,B\in\mathbb{R}^{m\times m} be symmetric matrices. Then AB=BAAB=BA if and only if there exists orthogonal matrix QQ such that Q1AQQ^{-1}AQ and Q1BQQ^{-1}BQ are diagonal matrices, where Q1Q^{-1} is the inverse of QQ.

Let ΛA=Q1AQ\Lambda_{A}=Q^{-1}AQ and ΛB=Q1BQ\Lambda_{B}=Q^{-1}BQ. From Lemma 2, if matrices A,Bm×mA,B\in\mathbb{R}^{m\times m} satisfies AB=BAAB=BA, we employ λi(A)\lambda_{i}(A) (or λi(B)\lambda_{i}(B)) to describe the entry located at the ii-th row and ii-th column of ΛA\Lambda_{A} (or ΛB\Lambda_{B}) with respect to QQ, as is the eigenvalue of AA (or BB) obviously.

For Am×mA\in\mathbb{R}^{m\times m}, we denote the ii-th largest eigenvalue of AA by λi(A)\lambda_{i}^{\downarrow}(A). The following lemma is about the comparison of eigenvalues.

Lemma 3

Let A,Bm×mA,B\in\mathbb{R}^{m\times m} be symmetric matrices. Then

λj(A)λi(B)+λji+n(AB)\lambda_{j}^{\downarrow}(A)\geq\lambda_{i}^{\downarrow}(B)+\lambda_{j-i+n}^{\downarrow}(A-B) (7)

for 1jim1\leq j\leq i\leq m.

We then introduce the diagonally dominant and strictly diagonally dominant matrices [31].

Definition 2

A matrix Am×mA\in\mathbb{R}^{m\times m} is diagonally dominant if

|Aii|ji|Aij|i=1,,m.|A_{ii}|\geq\sum\limits_{j\neq i}|A_{ij}|\quad\forall i=1,\ldots,m.

It is strictly diagonally dominant if

|Aii|>ji|Aij|i=1,,m.|A_{ii}|>\sum\limits_{j\neq i}|A_{ij}|\quad\forall i=1,\ldots,m.
Lemma 4

Let Am×mA\in\mathbb{R}^{m\times m} be strictly diagonally dominant. If AA is symmetric and Aii>0A_{ii}>0 for i=1,,mi=1,\ldots,m, then AA is positive definite.

3 Isomorphism between n\mathcal{L}_{n} and Euclidean space

In this section, we build up an isomorphism between n\mathcal{L}_{n} and the Euclidean space, which is the basis of this paper.

Let n¯=n×J\bar{n}=n\times J, where JJ denotes the cardinality of space Ξ\Xi defined by {ξ1,,ξJ}\{\xi^{1},\ldots,\xi^{J}\}. For 1iJ1\leq i\leq J, let xpi=p(ξi)x(ξi)x_{p_{i}}=\sqrt{p(\xi^{i})}x(\xi^{i}) and group these xpix_{p_{i}} in a vector xx, i.e., x=(xp1T,xp2T,,xpJT)Tx=(x_{p_{1}}^{T},x_{p_{2}}^{T},\ldots,x_{p_{J}}^{T})^{T}. Define a mapping ϕ:nn¯\phi:\mathcal{L}_{n}\to\mathbb{R}^{\bar{n}} as

ϕ:x()x.\phi:\quad x(\cdot)\mapsto x. (8)

Clearly, ϕ\phi is the isomorphism between n\mathcal{L}_{n} and n¯\mathbb{R}^{\bar{n}}.

Lemma 5

Let ϕ\phi be the mapping defined by (8). Then ϕ\phi is a linear isometric isomorphism between n\mathcal{L}_{n} and n¯\mathbb{R}^{\bar{n}}.

Before giving some operational formulas, we introduce some related notations. The norm of x()x(\cdot) in n\mathcal{L}_{n} induced by the inner product (1) can be defined as x()=x(),x()\|x(\cdot)\|=\sqrt{\left<x(\cdot),x(\cdot)\right>}. For the ease of statement, we use notation \|\cdot\| to denote both the norm in n\mathcal{L}_{n} and the Euclidean norm in n¯\mathbb{R}^{\bar{n}}. In addition, define B(x0,ϵ)={x|xx0ϵ}B(x_{0},\epsilon)=\{x|\|x-x_{0}\|\leq\epsilon\}.

Proposition 1

Let A,BnA,B\subset\mathcal{L}_{n} and ϕ\phi be defined by (8). Then, we have the following conclusions.

  • (a)(a)

    ϕ(AB)=ϕ(A)ϕ(B)\phi(A\cup B)=\phi(A)\cup\phi(B), ϕ(AB)=ϕ(A)ϕ(B)\phi(A\cap B)=\phi(A)\cap\phi(B);

  • (b)(b)

    int(ϕ(A))=ϕ(intA)\textrm{int}(\phi(A))=\phi(\textrm{int}A), bd(ϕ(A))=ϕ(bdA)\textrm{bd}(\phi(A))=\phi(\textrm{bd}A);

  • (c)(c)

    ϕ(A)=ϕ(A)\phi(A_{\infty})=\phi(A)_{\infty}, ϕ(A)=ϕ(A)\phi(A)^{*}=\phi(A^{*});

  • (d)(d)

    ϕ(aA+bB)=aϕ(A)+bϕ(B)\phi(aA+bB)=a\phi(A)+b\phi(B), a,bR\forall a,b\in R;

  • (e)(e)

    ϕ(A)\phi(A) is convex (or compact) if and only if AA is convex (or compact);

  • (f)(f)

    ϕ(Pn(x()))=Pϕ(n)(ϕ(x()))\phi(P_{\mathcal{M}_{n}}(x(\cdot)))=P_{\phi(\mathcal{M}_{n})}(\phi(x(\cdot))), where PnP_{\mathcal{M}_{n}} (or Pϕ(n)P_{\phi(\mathcal{M}_{n})}) is the projection operator onto subspace n\mathcal{M}_{n} defined by (2) (or ϕ(n)\phi(\mathcal{M}_{n})).

  • (g)(g)

    ϕ(N𝒞(x()))=Nϕ(𝒞)(ϕ(x()))\phi(N_{\mathcal{C}}(x(\cdot)))=N_{\phi(\mathcal{C})}(\phi(x(\cdot))), where 𝒞\mathcal{C} is defined by (3).

Proof.  We only need to prove (b)(b), (c)(c) and (f)(f), inasmuch as the other conclusions are obvious.

To prove (b), it suffices to prove ϕ(B(x(),ϵ))=B(ϕ(x()),ϵ)\phi(B(x(\cdot),\epsilon))=B(\phi(x(\cdot)),\epsilon). Given ϕ(y())\phi(y(\cdot)) with y()x()ϵ\|y(\cdot)-x(\cdot)\|\leq\epsilon, we have ϕ(y())ϕ(x())ϵ\|\phi(y(\cdot))-\phi(x(\cdot))\|\leq\epsilon since x()=ϕ(x())\|x(\cdot)\|=\|\phi(x(\cdot))\| for any x()nx(\cdot)\in\mathcal{L}_{n}. Thus, ϕ(B(x(),ϵ))B(ϕ(x()),ϵ)\phi(B(x(\cdot),\epsilon))\subset B(\phi(x(\cdot)),\epsilon). Similarly, we obtain B(ϕ(x()),ϵ)ϕ(B(x(),ϵ))B(\phi(x(\cdot)),\epsilon)\subset\phi(B(x(\cdot),\epsilon)). Hence, (b) holds.

We now prove (c). Given d()Ad(\cdot)\in A_{\infty}, we have x()+λd()Ax(\cdot)+\lambda d(\cdot)\in A for any λ0\lambda\geq 0 and x()Ax(\cdot)\in A. It follows that ϕ(x())+λϕ(d())ϕ(A)\phi(x(\cdot))+\lambda\phi(d(\cdot))\in\phi(A) for any λ0\lambda\geq 0 and x()Ax(\cdot)\in A, which indicates ϕ(d())ϕ(A)\phi(d(\cdot))\in\phi(A)_{\infty}. So we get ϕ(A)ϕ(A)\phi(A_{\infty})\subset\phi(A)_{\infty}. The converse inclusion can be obtained similarly.

To prove (f). Let y()=Pn(x())y(\cdot)=P_{\mathcal{M}_{n}}(x(\cdot)). According to the projection theorem in [21], ϕ(y())=Pϕ(n)(ϕ(x()))\phi(y(\cdot))=P_{\phi(\mathcal{M}_{n})}(\phi(x(\cdot))) if and only if ϕ(y())ϕ(x()),ϕ(z())ϕ(y())0\left<\phi(y(\cdot))-\phi(x(\cdot)),\phi(z(\cdot))-\phi(y(\cdot))\right>\geq 0 for all z()nz(\cdot)\in\mathcal{M}_{n}. Further, based on the definition of mapping ϕ\phi and Lemma 5, we have

ϕ(y())ϕ(x()),ϕ(z())ϕ(y())=ϕ(y()x()),ϕ(z()y())=1iJp(ξi)(y(ξi)x(ξi)),p(ξi)(z(ξi)y(ξi))=1iJp(ξi)y(ξi)x(ξi),z(ξi)y(ξi)=y()x(),z()y().\begin{array}[]{rcl}&&\left<\phi(y(\cdot))-\phi(x(\cdot)),\phi(z(\cdot))-\phi(y(\cdot))\right>=\left<\phi(y(\cdot)-x(\cdot)),\phi(z(\cdot)-y(\cdot))\right>\\ &=&\sum\limits_{1\leq i\leq J}{\left<\sqrt{p(\xi^{i})}(y(\xi^{i})-x(\xi^{i})),\sqrt{p(\xi^{i})}(z(\xi^{i})-y(\xi^{i}))\right>}\\ &=&\sum\limits_{1\leq i\leq J}{p(\xi^{i})\left<y(\xi^{i})-x(\xi^{i}),z(\xi^{i})-y(\xi^{i})\right>}=\left<y(\cdot)-x(\cdot),z(\cdot)-y(\cdot)\right>.\end{array} (9)

Actually, the last expression in (9) is nonnegative as long as we see that y()=Pn(x())y(\cdot)=P_{\mathcal{M}_{n}}(x(\cdot)) and employ the projection theorem again. Thus we obtain the conclusion that ϕ(Pn(x()))=Pϕ(n)(ϕ(x()))\phi(P_{\mathcal{M}_{n}}(x(\cdot)))=P_{\phi(\mathcal{M}_{n})}(\phi(x(\cdot))). \qed

The mapping ϕ\phi also provides an alternative description of SVI (4). Given a mapping (x())\mathcal{F}(x(\cdot)) from n\mathcal{L}_{n} to n\mathcal{L}_{n}, define

F^(x)=(p(ξ1)F(1p(ξ1)xp1,ξ1)T,,p(ξJ)F(1p(ξJ)xpJ,ξJ)T)T,\hat{F}(x)=(\sqrt{p(\xi^{1})}F(\frac{1}{\sqrt{p(\xi^{1})}}x_{p_{1}},\xi^{1})^{T},\ldots,\sqrt{p(\xi^{J})}F(\frac{1}{\sqrt{p(\xi^{J})}}x_{p_{J}},\xi^{J})^{T})^{T}, (10)

where x=(xp1T,xp2T,,xpJT)Tx=(x_{p_{1}}^{T},x_{p_{2}}^{T},\ldots,x_{p_{J}}^{T})^{T}. Obviously, ϕ((x()))=F^(ϕ(x()))\phi(\mathcal{F}(x(\cdot)))=\hat{F}(\phi(x(\cdot))). Then, SVI (4) can be transformed into

F^(x)Nϕ(𝒞𝒩n)(x).-\hat{F}(x)\in N_{\phi(\mathcal{C}\cap\mathcal{N}_{n})}(x). (11)

We use VI(ϕ(𝒞𝒩n)\phi(\mathcal{C}\cap\mathcal{N}_{n}),F^\hat{F}) to denote the problem (11)(\ref{eq3}), and its solution set is denoted as SOL(ϕ(𝒞𝒩n)\phi(\mathcal{C}\cap\mathcal{N}_{n}),F^\hat{F}).

Lemma 6

Let ϕ\phi be defined by (8). Then ϕ\phi establishes a one-to-one correspondence between SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) and SOL(ϕ(𝒞𝒩n)\phi(\mathcal{C}\cap\mathcal{N}_{n}),F^\hat{F}). In other words, if x()x(\cdot) is a solution of (4), then ϕ(x())\phi(x(\cdot)) is a solution of (11), and vice versa.

Proof.  Assume that x()x(\cdot) is the solution for (4). Let x=ϕ(x())x=\phi(x(\cdot)). It suffices to show that F^(x),yx0\left<\hat{F}(x),y-x\right>\geq 0 for all y=ϕ(y())ϕ(𝒞𝒩n)y=\phi(y(\cdot))\in\phi(\mathcal{C}\cap\mathcal{N}_{n}). Actually,

F^(x),yx=\displaystyle\left<\hat{F}(x),y-x\right>= 1iJp(ξi)F(1p(ξi)xpi,ξi),ypixpi\displaystyle\sum\limits_{1\leq i\leq J}{\left<\sqrt{p(\xi^{i})}F(\frac{1}{\sqrt{p(\xi^{i})}}x_{p_{i}},\xi^{i}),y_{p_{i}}-x_{p_{i}}\right>}
=\displaystyle= 1iJp(ξi)F(x(ξi),ξi),p(ξi)y(ξi)p(ξi)x(ξi)\displaystyle\sum\limits_{1\leq i\leq J}{\left<\sqrt{p(\xi^{i})}F(x(\xi^{i}),\xi^{i}),\sqrt{p(\xi^{i})}y(\xi^{i})-\sqrt{p(\xi^{i})}x(\xi^{i})\right>}
=\displaystyle= (x()),y()x()\displaystyle\left<\mathcal{F}(x(\cdot)),y(\cdot)-x(\cdot)\right>
\displaystyle\geq 0.\displaystyle 0.

To prove the converse conclusion, we only need to reverse the above deductions. \qed

4 Properties of solution sets of pseudomonotone SVIs

In this section, we generalize the solution theory of pseudomonotone determined VIs given in [7] to multistage pseudomonotone SVIs. We first introduce the results from [7] in the following theorems.

Theorem 2

Let KmK\subset\mathbb{R}^{m} be closed convex and F:KmF:K\rightarrow\mathbb{R}^{m} be continuous. Assume that F(x)F(x) is pseudomonotone on KK. Then the following three statements (a),(b)(a),(b) and (c)(c) are equivalent. Moreover, if there exists some x^K\hat{x}\in K such that the set

L={xK:F(x),xx^0}L_{\leq}=\{x\in K:\left<F(x),x-\hat{x}\right>\leq 0\}

is bounded, then SOL(KK,FF) is nonempty and compact.

  • (a)(a)

    There exists x^K\hat{x}\in K such that the set

    L<={xK:F(x),xx^<0}L_{<}=\{x\in K:\left<F(x),x-\hat{x}\right><0\}

    is bounded (possibly empty).

  • (b)(b)

    There exist a bounded open set Ω\Omega and some x^KΩ\hat{x}\in K\cap\Omega such that

    F(x),xx^0,xKbdΩ.\left<F(x),x-\hat{x}\right>\geq 0,\quad\forall x\in K\cap bd\Omega.
  • (c)(c)

    VI(KK,FF) has a solution.

Theorem 3

Let KmK\subset\mathbb{R}^{m} be closed convex and F:KmF:K\rightarrow\mathbb{R}^{m} be continuous. If FF is pseudomonotone on KK, then the following statements hold.

  • (a)(a)

    The solution set SOL(KK,FF) is convex.

  • (b)(b)

    If there exists x^K\hat{x}\in K such that F(x^)int(K)F(\hat{x})\in{\rm int}(K_{\infty})^{*}, then SOL(KK,FF) is nonempty, convex and compact.

Theorem 4

Let KmK\subset\mathbb{R}^{m} be closed convex and F:KmF:K\rightarrow\mathbb{R}^{m} be continuous. Assume that FF is pseudomonotone on KK. Then the set SOL(KK,FF) is nonempty and bounded if and only if

K[(F(K))]={0}.K_{\infty}\cap[-(F(K)^{*})]=\{0\}.

Theorems 2, 3 and 4 do not pertain to the pseudomonotne SVI (4)(\ref{eq1}) due to the fact that they are based on the Euclidean space. Nonetheless, this extension follows the isomorphim introduced in Section 3, which are presented below.

Theorem 5

Let (x()):nn\mathcal{F}(x(\cdot)):\mathcal{L}_{n}\rightarrow\mathcal{L}_{n} be continuous. Assume that (x())\mathcal{F}(x(\cdot)) is pseudomonotone on 𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n}. Then the following three statements (a),(b)(a),(b) and (c)(c) are equivalent. Moreover, if there exists some x^()𝒞𝒩n\hat{x}(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n} such that the set

L={x()𝒞𝒩n:(x()),x()x^()0}L_{\leq}=\{x(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n}:\left<\mathcal{F}(x(\cdot)),x(\cdot)-\hat{x}(\cdot)\right>\leq 0\}

is bounded, then SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) is nonempty and compact.

  • (a)(a)

    There exists x^()𝒞𝒩n\hat{x}(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n} such that the set

    L<={x()𝒞𝒩n:(x()),x()x^()<0}L_{<}=\{x(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n}:\left<\mathcal{F}(x(\cdot)),x(\cdot)-\hat{x}(\cdot)\right><0\}

    is bounded (possibly empty).

  • (b)(b)

    There exist a bounded open set Ω\Omega and some x^()𝒞𝒩nΩ\hat{x}(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n}\cap\Omega such that

    (x()),x()x^()0,x()𝒞𝒩nbdΩ.\left<\mathcal{F}(x(\cdot)),x(\cdot)-\hat{x}(\cdot)\right>\geq 0,\quad\forall x(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n}\cap bd\Omega.
  • (c)(c)

    SVI(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) has a solution.

Proof.  We first prove the statements (a), (b) and (c) are equivalent.

``(a)(b)"``(a)\Rightarrow(b)". Let Ω\Omega be a bounded open set containing L<{x^()}L_{<}\cup\{\hat{x}(\cdot)\}. It suffices to see that L<bdΩ=L_{<}\cap bd\Omega=\emptyset.

``(b)(c)"``(b)\Rightarrow(c)". Denote x=ϕ(x())x=\phi(x(\cdot)) with x()𝒞𝒩nΩx(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n}\cap\Omega, and x^=ϕ(x^())\hat{x}=\phi(\hat{x}(\cdot)). Following the same procedure in Lemma 6, we have

F^(x^),xx^0,xϕ(𝒞𝒩n)bd(ϕ(Ω)).\left<\hat{F}(\hat{x}),x-\hat{x}\right>\geq 0,\quad\forall x\in\phi(\mathcal{C}\cap\mathcal{N}_{n})\cap\textrm{bd}(\phi(\Omega)).

Then, SOL(ϕ(𝒞𝒩n)\phi(\mathcal{C}\cap\mathcal{N}_{n}),F^\hat{F}) is nonempty by means of Theorem 2. In terms of Lemma 6, SVI(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) has a solution.

``(c)(a)"``(c)\Rightarrow(a)". Let x^()\hat{x}(\cdot) be a solution of SVI(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}). Then,

(x^()),x()x^()0,x()𝒞𝒩n.\left<\mathcal{F}(\hat{x}(\cdot)),x(\cdot)-\hat{x}(\cdot)\right>\geq 0,\quad\forall x(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n}.

Since \mathcal{F} is pseudomonotone on 𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n}, we have (x()),x()x^()0\left<\mathcal{F}(x(\cdot)),x(\cdot)-\hat{x}(\cdot)\right>\geq 0 for any x()𝒞𝒩nx(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n}, as indicates the emptiness of L<L_{<}.

Since there exists some x^()𝒞𝒩n\hat{x}(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n} such that the set such that LL_{\leq} is bounded, (a) holds due to L<LL_{<}\subset L_{\leq}. Thus, (c) holds, which implies that SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) is nonempty. Furthermore, SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) is compact because SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) is closed and SOL(𝒞𝒩n,)L(\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F})\subset L_{\leq}. \qed

Theorem 6

Let (x()):nn\mathcal{F}(x(\cdot)):\mathcal{L}_{n}\rightarrow\mathcal{L}_{n} be continuous. Assume that (x())\mathcal{F}(x(\cdot)) is pseudomonotone on 𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n}, the following statements hold.

  • (a)(a)

    The solution set SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) is convex.

  • (b)(b)

    If there exists x^()𝒞𝒩n\hat{x}(\cdot)\in\mathcal{C}\cap\mathcal{N}_{n} such that (x^())int((𝒞𝒩n))\mathcal{F}(\hat{x}(\cdot))\in{\rm int}((\mathcal{C}\cap\mathcal{N}_{n})_{\infty})^{*}, then SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) is nonempty, convex and compact.

Proof.``(a)"``(a)" Obviously, if \mathcal{F} is pseudomonotone on 𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n}, then F^\hat{F} is also pseudomonotone on ϕ(𝒞𝒩n)\phi(\mathcal{C}\cap\mathcal{N}_{n}). According to Theorem 3, SOL(ϕ(𝒞𝒩n)\phi(\mathcal{C}\cap\mathcal{N}_{n}),F^\hat{F}) is convex. Then SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) is convex by Lemma 6 and Proposition 1.

``(b)"``(b)" If (x^())int((𝒞𝒩n))\mathcal{F}(\hat{x}(\cdot))\in{\rm int}((\mathcal{C}\cap\mathcal{N}_{n})_{\infty})^{*}, then ϕ((x^()))ϕ(int((𝒞𝒩n)))\phi(\mathcal{F}(\hat{x}(\cdot)))\in\phi({\rm int}((\mathcal{C}\cap\mathcal{N}_{n})_{\infty})^{*}). Let x^=ϕ(x^())\hat{x}=\phi(\hat{x}(\cdot)). Due to the definition of F^\hat{F} and Proposition 1, we have F^(x^)int((ϕ(𝒞𝒩n)))\hat{F}(\hat{x})\in{\rm int}((\phi(\mathcal{C}\cap\mathcal{N}_{n}))_{\infty})^{*}. Then the fact that SOL(ϕ(𝒞𝒩n)\phi(\mathcal{C}\cap\mathcal{N}_{n}),F^\hat{F}) is nonempty, convex and compact follows from 3. It follows from Proposition 1 that SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) is nonempty, convex and compact. \qed

Theorem 7

Let (x()):nn\mathcal{F}(x(\cdot)):\mathcal{L}_{n}\rightarrow\mathcal{L}_{n} be continuous. Assume that (x())\mathcal{F}(x(\cdot)) is pseudomonotone on 𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n}. Then, the set SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) is nonempty and bounded if and only if

(𝒞𝒩n)[((𝒞𝒩n))]={0}.(\mathcal{C}\cap\mathcal{N}_{n})_{\infty}\cap[-(\mathcal{F}(\mathcal{C}\cap\mathcal{N}_{n})^{*})]=\{0\}. (12)

Proof.  By Lemma 6 and Proposition 1, SOL(𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n},\mathcal{F}) is nonempty and bounded if and only if SOL(ϕ(𝒞𝒩n)\phi(\mathcal{C}\cap\mathcal{N}_{n}),F^\hat{F}) is nonempty and bounded. On the other hand, if \mathcal{F} is pseudomonotone on 𝒞𝒩n\mathcal{C}\cap\mathcal{N}_{n}, then F^\hat{F} is pseudomonotone on ϕ(𝒞𝒩n)\phi(\mathcal{C}\cap\mathcal{N}_{n}). Based on Theorem 4, SOL(ϕ(𝒞𝒩n)\phi(\mathcal{C}\cap\mathcal{N}_{n}),F^\hat{F}) is nonempty and bounded if and only if (ϕ(𝒞𝒩n))([F^(ϕ(𝒞𝒩n))])={0}{(\phi(\mathcal{C}\cap\mathcal{N}_{n}))}_{\infty}\cap([-\hat{F}(\phi(\mathcal{C}\cap\mathcal{N}_{n}))^{*}])=\{0\}, as is equivalent to (12) due to Proposition 1. Thus we complete the proof. \qed

5 Finding solutions to SVI via PHA

5.1 Description of elicited PHA

As is presented in Remark 1, the monotone mapping \mathcal{F} must be pseudomonotone, but the converse is not necessarily true. Thus the original PHA in [26] can not be applied to the pseudomonotone SVI (4) directly. Instead, we employ the elicited PHA proposed in [38] to solve the pseudomonotone SVI (4), which is motivated by the work of Rockafellar in [30].

We first introduce the concept of (global) maximal monotonicity in [25] and (global) elicited monotonicity in [30, 38].

Definition 3

Denote the graph of set-valued mapping T:HHT:H\rightrightarrows H by gphT={(x,y):yT(x)}gphT=\{(x,y):y\in T(x)\}. Then a monotone mapping T:HHT:H\rightrightarrows H is maximal monotone if no enlargement of gphTgphT is possible in H×HH\times H without destroying monotonicity, or in other words, if for every pair (x^,y^)(H×H)\gphT(\hat{x},\hat{y})\in(H\times H)\backslash gphT there exists (x~,y~)gphT(\tilde{x},\tilde{y})\in gphT with y^y~,x^x~<0\left<\hat{y}-\tilde{y},\hat{x}-\tilde{x}\right><0, where HH stands for the Hilbert space.

Definition 4

Let :nn\mathcal{F}:\mathcal{L}_{n}\to\mathcal{L}_{n}. Given closed convex set 𝒞\mathcal{C}, subspace 𝒩n\mathcal{N}_{n} and its complement n\mathcal{M}_{n}, the monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} is said to be elicited at level s>0s>0 (with respect to 𝒩n\mathcal{N}_{n})) if +N𝒞+sPn\mathcal{F}+N_{\mathcal{C}}+sP_{\mathcal{M}_{n}} is maximal monotone globally.

Despite the fact that +N𝒞\mathcal{F}+N_{\mathcal{C}} may be not maximal monotone, +N𝒞+sPn\mathcal{F}+N_{\mathcal{C}}+sP_{\mathcal{M}_{n}} may be maximal monotone for some s>0s>0. So we can apply the PPA to +N𝒞+sPn\mathcal{F}+N_{\mathcal{C}}+sP_{\mathcal{M}_{n}}, which is the core idea of the elicited PHA. We refer readers to [30, 38] for more details about the elicited PHA.

We intend to use the elicited PHA to solve the pseudomonotone multistage SVIs. The algorithmic framework and the convergence analysis of the elicited PHA in [38] are listed below for completeness.

Algorithm 1 Elicited PHA for multistage pseudomonotone SVIs
   Initialization. Given parameter r>s0r>s\geq 0, x0()𝒩nx^{0}(\cdot)\in{\cal N}_{n} and w0()nw^{0}(\cdot)\in{\cal M}_{n}. Set k=0k=0.
   Iterations.
  • Step 1 For each ξΞ\xi\in\Xi, find x^k(ξ):=\hat{x}^{k}(\xi):= the unique x(ξ)x(\xi) such that

    F(x(ξ),ξ)wk(ξ)r[x(ξ)xk(ξ)]N𝒞(ξ)(x(ξ)).-F(x(\xi),\xi)-w^{k}(\xi)-r[x(\xi)-x^{k}(\xi)]\in N_{\mathcal{C}(\xi)}(x(\xi)). (13)
  • Step 2 (primal update). xk+1()=P𝒩n(x^k())x^{k+1}(\cdot)=P_{\mathcal{N}_{n}}({\hat{x}}^{k}(\cdot)).

  • Step 3 (dual update). wk+1()=wk()+(rs)(x^k()xk+1())w^{k+1}(\cdot)=w^{k}(\cdot)+(r-s)({\hat{x}}^{k}(\cdot)-x^{k+1}(\cdot)).

  Set k:=k+1k:=k+1; repeat until a stopping criterion is met.
Theorem 8

Suppose that +N𝒞\mathcal{F}+N_{\mathcal{C}} in SVI (4) is globally elicited at level ss and the constraint qualification holds. As long as SVI (4) has a solution, the sequence {xk(),ωk()}\{x^{k}(\cdot),\omega^{k}(\cdot)\} generated by Algorithm 1 will converge to some pair (x¯(),ω¯())(\bar{x}(\cdot),\bar{\omega}(\cdot)) satisfying (5) and thus furnish x¯()\bar{x}(\cdot) as a solution to (4). In the special case that \mathcal{F} is linear and 𝒞\mathcal{C} is polyhedral, the convergence will surely be at a linear rate with respect to the norm

(x(),ω())r,s2=x()2+1r(rs)ω()2.{\|(x(\cdot),\omega(\cdot))\|}^{2}_{r,s}={\|x(\cdot)\|}^{2}+\frac{1}{r(r-s)}{\|\omega(\cdot)\|}^{2}.

5.2 Some criteria for the elicited monotonicity

In this subsection, we will provide some criteria for the elicited monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} in SVI (4). At first, we provide one useful fact about the eigenvalues of the Jacobian of pseudomonotone function.

Lemma 7

Let F:SmF:S\rightarrow\mathbb{R}^{m} be a continuously differentiable function defined on an open convex set SmS\subset\mathbb{R}^{m}. Assume that FF is pseudomonotone on SS Then the Jacobian DF(x)DF(x) has at most one negative eigenvalue if it is symmetric, where an eigenvalue with multiplicity k(k>1)k\ (k>1) is counted as kk eigenvalues.

Proof.  According to [7, Theorem 1.3.1], there is a real-valued function θ:m\theta:\mathbb{R}^{m}\rightarrow\mathbb{R} such that θ(x)=F(x)\nabla\theta(x)=F(x) for xSx\in S. By [1, Theorems 3.4.1, 5.5.2], the Hessian matrix 2θ(x)\nabla^{2}\theta(x) has at most one negative eigenvalue. So DF(x)2θ(x)DF(x)\equiv\nabla^{2}\theta(x) has at most one negative eigenvalue. \qed

Remark 2

The condition “DF(x)DF(x) is symmetric” in Lemma 7 is called the symmetry condition. Actually, the symmetry condition holds true if and only if there is a real-valued differentiable function θ(x)\theta(x) such that θ(x)=F(x)\nabla\theta(x)=F(x) for all xSx\in S, which is equivalent to the integrability condition that FF is integrable on SS. For more details, refer to [7, Theorem 1.3.1].

There are two important examples where DF(x)DF(x) is symmetric. One is the separable function, i.e., F(x)=(Fi(xi):i=1,2,,m)F(x)=(F_{i}(x_{i}):i=1,2,\ldots,m), where Fi(xi)F_{i}(x_{i}), the component function of F(x)F(x), only depends on the single variable xix_{i}. If F(x)F(x) is separable and differentiable, then Jacobian DF(x)DF(x) is diagonal for all xSx\in S. The other one is the linear function F(x)=Mx+qF(x)=Mx+q with Mm×mM\in\mathbb{R}^{m\times m} being symmetric and qmq\in\mathbb{R}^{m}. In this case, DF(x)=MDF(x)=M is symmetric for xSx\in S.

The first criterion is based on [30, Theorem 6], which is shown in the theorem below.

Theorem 9

Let HH be the Hilbert space, and NsN_{s} and MsM_{s} be the orthogonal subspaces of HH. Denote T=F+NCT=F+N_{C} for a nonempty closed convex subset CHC\subset H and a continuously differentiable mapping F:CHF:C\rightarrow H with Jacobians DF(x)DF(x). Assume that there exists α(x)\alpha(x) such that when xCx\in C

y,DF(x)yα(x)y2,yNs.\left<y,DF(x)y\right>\geq\alpha(x){\|y\|}^{2},\quad\forall\ y\in N_{s}.

Let

β(x)=12PNs(DF(x)+(DF(x))T)PMs,γ(x)=PMsDF(x)PMs.\beta(x)=\frac{1}{2}\|P_{N_{s}}(DF(x)+(DF(x))^{T})P_{M_{s}}\|,\quad\gamma(x)=\|P_{M_{s}}DF(x)P_{M_{s}}\|.

Define

e~=supxCNs{β2(x)α(x)+γ(x)}.\tilde{e}=\sup\limits_{x\in C\cap N_{s}}\left\{\frac{\beta^{2}(x)}{\alpha(x)}+\gamma(x)\right\}.

If e~<+\tilde{e}<+\infty, then the monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} is globally elicited at level s>e~s>\tilde{e} with respect to NsN_{s}.

Similarly, the isomorphism in Section 3 enables us to extend Theorem 9 to pseudomonotone SVI (4).

Theorem 10

Consider SVI (4). Let :𝒞n\mathcal{F}:\mathcal{C}\rightarrow\mathcal{L}_{n} be continuously differentiable and pseudomonotone on 𝒞\mathcal{C}. Let F^\hat{F} be defined in (10). Assume that there exists α(x)\alpha(x) such that when xϕ(𝒞)x\in\phi(\mathcal{C})

y,DF^(x)yα(x)y2,yϕ(𝒩n).\left<y,D{\hat{F}}(x)y\right>\geq\alpha(x){\|y\|}^{2},\quad\forall\ y\in\phi(\mathcal{N}_{n}). (14)

Let

β(x)=12Pϕ(𝒩n)(DF^(x)+(DF^(x))T)Pϕ(n),γ(x)=Pϕ(n)DF^(x)Pϕ(n).\beta(x)=\frac{1}{2}\|P_{\phi(\mathcal{N}_{n})}(D\hat{F}(x)+(D\hat{F}(x))^{T})P_{\phi(\mathcal{M}_{n})}\|,\quad\gamma(x)=\|P_{\phi(\mathcal{M}_{n})}D\hat{F}(x)P_{\phi(\mathcal{M}_{n})}\|.

Define

e0=supxϕ(𝒞)ϕ(𝒩n){β2(x)α(x)+γ(x)}.e_{0}=\sup\limits_{x\in\phi(\mathcal{C})\cap\phi(\mathcal{N}_{n})}\left\{\frac{\beta^{2}(x)}{\alpha(x)}+\gamma(x)\right\}.

If e0<+e_{0}<+\infty, then the monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} is globally elicited at level s>e0s>e_{0}.

Proof.  Let 𝒯s=+N𝒞+sPn\mathcal{T}_{s}=\mathcal{F}+N_{\mathcal{C}}+sP_{\mathcal{M}_{n}} and 𝒯^s=F^+Nϕ(𝒞)+sPϕ(n){\hat{\mathcal{T}}}_{s}=\hat{F}+N_{\phi(\mathcal{C})}+sP_{\phi(\mathcal{M}_{n})}. By Theorem 9 and the hypotheses, 𝒯^s{\hat{\mathcal{T}}}_{s} is globally maximal monotone for s>e0s>e_{0}.

Suppose by contradiction that there exists s>e0s^{\prime}>e_{0} such that 𝒯s\mathcal{T}_{s^{\prime}} is not maximal monotone. Then there is a pair (xu(),yu())n×n(x_{u}(\cdot),y_{u}(\cdot))\in\mathcal{L}_{n}\times\mathcal{L}_{n} with yu()𝒯s(xu())y_{u}(\cdot)\notin\mathcal{T}_{s^{{}^{\prime}}}(x_{u}(\cdot)) such that

xu()xv(),yu()yv()0,(xv(),yv())with yv()𝒯s(xv()).\left<x_{u}(\cdot)-x_{v}(\cdot),y_{u}(\cdot)-y_{v}(\cdot)\right>\geq 0,\quad\forall(x_{v}(\cdot),y_{v}(\cdot))~{}\text{with $y_{v}(\cdot)\in\mathcal{T}_{s^{\prime}}(x_{v}(\cdot))$}. (15)

Let xu=ϕ(xu())x_{u}=\phi(x_{u}(\cdot)), xv=ϕ(xv())x_{v}=\phi(x_{v}(\cdot)), yu=ϕ(yu())y_{u}=\phi(y_{u}(\cdot)) and yv=ϕ(yv())y_{v}=\phi(y_{v}(\cdot)). Obviously, yu𝒯^s(xu)y_{u}\notin{\hat{\mathcal{T}}}_{s^{\prime}}(x_{u}) and yv𝒯^s(xv)y_{v}\in{\hat{\mathcal{T}}}_{s^{\prime}}(x_{v}). Following the proof of Lemma 6, (15) yields

xuxv,yuyv0,\left<x_{u}-x_{v},y_{u}-y_{v}\right>\geq 0,

which implies that 𝒯^s{\hat{\mathcal{T}}}_{s^{\prime}} is not globally maximal monotone. This is a contradiction and hence we complete the proof. \qed

Remark 3

In the following two special cases, we can simplify the condition (14) in Theorem 10.

Case I: Denote Ker(F^(x)):={z:z,F^(x)=0}.{\rm Ker}(\hat{F}(x)):=\{z:\left<z,\hat{F}(x)\right>=0\}. Assume that FF is pseudomonotone on S:={Pϕ(𝒩n)z:zRn¯\Ker(F^(x))}S^{\prime}:=\{P_{\phi(\mathcal{N}_{n})}z:z\in R^{\bar{n}}\backslash{\rm Ker}(\hat{F}(x))\}. By Lemma 1, the condition (14) can be simplified as follows:

y,DF^(x)yα(x)y2,yS,whenxintϕ(𝒞),y,DF^(x)yα(x)y2,yϕ(𝒩n),whenxbdϕ(𝒞).\begin{array}[]{rcl}&&\left<y,D{\hat{F}}(x)y\right>\geq\alpha(x){\|y\|}^{2},\quad\forall y\in S^{\prime},\ \mbox{when}\ x\in{\rm int}\phi(\mathcal{C}),\\ &&\left<y,D{\hat{F}}(x)y\right>\geq\alpha(x){\|y\|}^{2},\quad\forall y\in\phi(\mathcal{N}_{n}),\ \mbox{when}\ x\in{\rm bd}\phi(\mathcal{C}).\end{array} (16)

Case II: Let S:={Pϕ(𝒩n)z:zRn¯\Ker(F^(x))}S^{\prime}:=\{P_{\phi(\mathcal{N}_{n})}z:z\in R^{\bar{n}}\backslash{\rm Ker}(\hat{F}(x))\}. If there exists an open convex set 𝒞0𝒞\mathcal{C}_{0}\supset\mathcal{C} such that :𝒞0n\mathcal{F}:\mathcal{C}_{0}\rightarrow\mathcal{L}_{n} is continuously differentiable and pseudomonotone on 𝒞0\mathcal{C}_{0}, then the condition (14) can be simplified as

y,DF^(x)yα(x)y2,yS,whenxϕ(𝒞).\left<y,D{\hat{F}}(x)y\right>\geq\alpha(x){\|y\|}^{2},\ \forall y\in S^{\prime},\ \mbox{when}\ x\in\phi(\mathcal{C}).

We now give a numerical example to illustrate the above criteria. To avoid the unnecessary confusion brought by the stages and the cardinality of the sample space Ξ\Xi, we assume that stage NN and the cardinality of Ξ\Xi are both 11.

Example 1

Let

F(x1,x2)=(x1,x2)T,𝒩={(l,0)T:l},𝒞={(x1,x2)T:x12,x22,x2x10}.\begin{array}[]{rcl}&&F(x_{1},x_{2})=(x_{1},-x_{2})^{T},\quad\mathcal{N}=\{(l,0)^{T}:l\in\mathbb{R}\},\\ &&\mathcal{C}=\{(x_{1},x_{2})^{T}:x_{1}\geq 2,x_{2}\geq 2,x_{2}-x_{1}\geq 0\}.\end{array}

Then SVI (4) is exhibited as

F(x1,x2)N𝒞𝒩(x1,x2).-F(x_{1},x_{2})\in N_{\mathcal{C}\cap\mathcal{N}}(x_{1},x_{2}).

Obviously, DF(x1,x2))=[1,0;0,1]DF(x_{1},x_{2}))=[1,0;0,-1]. The orthogonal complement of 𝒩\mathcal{N} is ={(0,t)T:t}\mathcal{M}=\{(0,t)^{T}:t\in\mathbb{R}\} and the corresponding projection matrix P=[0,0;0,1]P_{\mathcal{M}}=[0,0;0,1].

We now prove that F(x1,x2)F(x_{1},x_{2}) is pseudomonotone on 𝒞\mathcal{C}. Construct

𝒞0={(x1,x2)T:x1>1,x2>1,x22x1>0}.\mathcal{C}_{0}=\{(x_{1},x_{2})^{T}:x_{1}>1,x_{2}>1,x_{2}-2x_{1}>0\}.

Then 𝒞0𝒞\mathcal{C}_{0}\supset\mathcal{C} is an open convex set. For (x1,x2)T𝒞0(x_{1},x_{2})^{T}\in\mathcal{C}_{0}, we have F(x1,x2)0F(x_{1},x_{2})\neq 0 and

Ker(F(x1,x2))={(u1,u2)T:u2=x1x2u1}.{\rm Ker}(F(x_{1},x_{2}))=\{(u_{1},u_{2})^{T}:u_{2}=\frac{x_{1}}{x_{2}}u_{1}\}.

Given (u1,u2)TKer(F(x1,x2))(u_{1},u_{2})^{T}\in{\rm Ker}(F(x_{1},x_{2})), we have

(u1,u2)DF(x1,x2)(u1,u2)T=u12u22=u12(x1x2)2u12>0.(u_{1},u_{2})DF(x_{1},x_{2})(u_{1},u_{2})^{T}={u_{1}}^{2}-{u_{2}}^{2}={u_{1}}^{2}-{(\frac{x_{1}}{x_{2}})}^{2}{u_{1}}^{2}>0.

By Lemma 1, F(x1,x2)F(x_{1},x_{2}) is pseudomonotone on 𝒞\mathcal{C}. But F(x1,x2)F(x_{1},x_{2}) is not monotone since DF(x1,x2))DF(x_{1},x_{2})) has an negative eigenvalue.

Given x𝒞x\in\mathcal{C}, we have u,DF(x)uu2\left<u,DF(x)u\right>\geq{\|u\|}^{2} for u𝒩u\in\mathcal{N}. Thus α(x)=1\alpha(x)=1, β(x)=0\beta(x)=0 and γ(x)=1\gamma(x)=1 in Theorem 10. Then the monotonicity of F+N𝒞F+N_{\mathcal{C}} is globally elicited at level se0=1s\geq e_{0}=1 via Theorem 10.

Theorem 11

Consider SVI (4). Let 𝒞0𝒞\mathcal{C}_{0}\supset\mathcal{C} be an open convex set in n\mathcal{L}_{n}, and :𝒞0n\mathcal{F}:\mathcal{C}_{0}\rightarrow\mathcal{L}_{n} be continuously differentiable and pseudomonotone on 𝒞0\mathcal{C}_{0}. Assume that F^\hat{F} defined in (10) satisfies the following conditions:

(i)DF^(x) is symmetric;(ii)DF^(x)Pϕ(n)=Pϕ(n)DF^(x)\begin{array}[]{l}(i)\ \text{$D\hat{F}(x)$ is symmetric};\\ (ii)\ D\hat{F}(x)P_{\phi(\mathcal{M}_{n})}=P_{\phi(\mathcal{M}_{n})}D\hat{F}(x)\end{array} (17)

for any xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}), and

(iii)λi(Pϕ(n))>0if λi(DF^(x))<0 for some xϕ(𝒞0).(iii)\ \lambda_{i}(P_{\phi(\mathcal{M}_{n})})>0\ \text{if $\lambda_{i}(D\hat{F}(x))<0$ for some $x\in\phi(\mathcal{C}_{0})$}.

Let

e1=infxϕ(𝒞0)λn¯(DF^(x)).e_{1}=\inf\limits_{x\in\phi(\mathcal{C}_{0})}\lambda_{\bar{n}}^{\downarrow}(D\hat{F}(x)).

If e1>e_{1}>-\infty, then the monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} is globally elicited at level se1s\geq-e_{1}.

Proof.  According to Lemma 2, the eigenvalues of DF^(x)+sPϕ(n)D\hat{F}(x)+sP_{\phi(\mathcal{M}_{n})} can be denoted as λi(DF^(x))+sλi(Pϕ(n))\lambda_{i}(D\hat{F}(x))+s\lambda_{i}(P_{\phi(\mathcal{M}_{n})}) for xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}) and 1in¯1\leq i\leq\bar{n}. Inasmuch as the eigenvalues of projection matrix is 0 or 1, λi(DF^(x))+sλi(Pϕ(n))0\lambda_{i}(D\hat{F}(x))+s\lambda_{i}(P_{\phi(\mathcal{M}_{n})})\geq 0 when λi(DF^(x))0\lambda_{i}(D\hat{F}(x))\geq 0. If λi(DF^(x))<0\lambda_{i}(D\hat{F}(x))<0 for some xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}), it is not difficult to see that λi(DF^(x))+sλi(Pϕ(n))0\lambda_{i}(D\hat{F}(x))+s\lambda_{i}(P_{\phi(\mathcal{M}_{n})})\geq 0 when se1s\geq-e_{1}. By [7, Proposition 2.3.2], F^+sPϕ(n)\hat{F}+sP_{\phi(\mathcal{M}_{n})} is monotone on ϕ(𝒞0)\phi(\mathcal{C}_{0}). Thus F^+Nϕ(𝒞)+sPϕ(n)\hat{F}+N_{\phi(\mathcal{C})}+sP_{\phi(\mathcal{M}_{n})} is maximal monotone from [28, Theorem 3], which indicates the maximal monotonicity of +N𝒞+sPn\mathcal{F}+N_{\mathcal{C}}+sP_{\mathcal{M}_{n}}. \qed

Example 2

Assume that stage NN and the cardinality of Ξ\Xi are both 1. Let F(x1,x2)=(0,x2)TF(x_{1},x_{2})=(0,-x_{2})^{T}, 𝒞={(x1,x2)T:x11,x21}\mathcal{C}=\{(x_{1},x_{2})^{T}:x_{1}\geq 1,x_{2}\geq 1\}, and 𝒩={(l,0)T:l}\mathcal{N}=\{(l,0)^{T}:l\in\mathbb{R}\}. Then SVI (4) is exhibited as

F(x1,x2)N𝒞𝒩(x1,x2).-F(x_{1},x_{2})\in N_{\mathcal{C}\cap\mathcal{N}}(x_{1},x_{2}).

Obviously, DF(x1,x2))=[0,0;0,1]DF(x_{1},x_{2}))=[0,0;0,-1], the orthogonal complement of 𝒩\mathcal{N} is ={(0,t)T:tR}\mathcal{M}=\{(0,t)^{T}:t\in R\} and the corresponding projection matrix is P=[0,0;0,1]P_{\mathcal{M}}=[0,0;0,1].

Construct 𝒞0={(x1,x2)T:x1>0,x2>0}\mathcal{C}_{0}=\{(x_{1},x_{2})^{T}:x_{1}>0,x_{2}>0\}. Then 𝒞0𝒞\mathcal{C}_{0}\supset\mathcal{C} is an open convex set. It holds that F(x1,x2)F(x_{1},x_{2}) is pseudomonotone on 𝒞0\mathcal{C}_{0}. In fact, for any given (x1,x2)T,(y1,y2)T𝒞0(x_{1},x_{2})^{T},(y_{1},y_{2})^{T}\in\mathcal{C}_{0} with F(x1,x2),(y1x1,y2x2)T0\left<-F(x_{1},x_{2}),(y_{1}-x_{1},y_{2}-x_{2})^{T}\right>\geq 0, we have y2x20y_{2}-x_{2}\leq 0 due to

F(x1,x2),(y1x1,y2x2)T=x2(y2x2),x2>0.\left<-F(x_{1},x_{2}),(y_{1}-x_{1},y_{2}-x_{2})^{T}\right>=-x_{2}(y_{2}-x_{2}),\quad x_{2}>0.

This together with y20y_{2}\geq 0 implies that

F(y1,y2),(y1x1,y2x2)T=y2(y2x2)0.\left<-F(y_{1},y_{2}),(y_{1}-x_{1},y_{2}-x_{2})^{T}\right>=-y_{2}(y_{2}-x_{2})\geq 0.

Hence, F(x1,x2)F(x_{1},x_{2}) is pseudomonotone on 𝒞0\mathcal{C}_{0}. However, F(x1,x2)F(x_{1},x_{2}) is not monotone on 𝒞0\mathcal{C}_{0} because of the negative eigenvalue of DF(x1,x2)DF(x_{1},x_{2}). Nevertheless, for (x1,x2)T𝒞(x_{1},x_{2})^{T}\in\mathcal{C}, DF(x1,x2)DF(x_{1},x_{2}) is symmetric, and DF(x1,x2)P=PDF(x1,x2)DF(x_{1},x_{2})P_{\mathcal{M}}=P_{\mathcal{M}}DF(x_{1},x_{2}). Besides, λi(P)=1>0\lambda_{i}(P_{\mathcal{M}})=1>0 while λi(DF(x))<0\lambda_{i}(DF(x))<0. Thus, by Theorem 11, F+N𝒞+sPF+N_{\mathcal{C}}+sP_{\mathcal{M}} is globally maximal monotone at level se1=1s\geq e_{1}=1.

Remark 4

Note that the condition (ii)(ii) in (17) may hold true even when DF^(x)D\hat{F}(x) and Pϕ(n)P_{\phi(\mathcal{M}_{n})} are not diagonal matrices. For example, assume that stage NN and the cardinality of Ξ\Xi are both 1, C=+4C=\mathbb{R}_{+}^{4}, F^=Mx\hat{F}=Mx with

M=[10001041001401005],M=\begin{bmatrix}10&0&0&1\\ 0&4&1&0\\ 0&1&4&0\\ 1&0&0&5\end{bmatrix},

and subspace M4=PM4xM_{4}=P_{M_{4}}x with

PM4=[100000.50.5000.50.500001].P_{M_{4}}=\begin{bmatrix}1&0&0&0\\ 0&0.5&0.5&0\\ 0&0.5&0.5&0\\ 0&0&0&1\end{bmatrix}.

PM4P_{M_{4}} is a projection matrix due to PM42=PM4P_{M_{4}}^{2}=P_{M_{4}} and PM4T=PM4P_{M_{4}}^{T}=P_{M_{4}}. Since MM is positive semidefinite, F^\hat{F} is monotone on CC and thus pseudomonotone on CC. Via some simple calculations, we can see that DF^(x)PM4=PM4DF^(x)D\hat{F}(x)P_{M_{4}}=P_{M_{4}}D\hat{F}(x) for xCx\in C, as is consistent with condition (ii)(ii) in (17).

In case that it is difficult to calculate the minimum eigenvalue λn¯(DF^(x))\lambda_{\bar{n}}^{\downarrow}(D\hat{F}(x)) over ϕ(𝒞0)\phi({\cal C}_{0}), the following corollary says that we may use the spectral radius to replace the minimum eigenvalue.

Corollary 1

Consider SVI (4). Let 𝒞0𝒞\mathcal{C}_{0}\supset\mathcal{C} be an open convex set in n\mathcal{L}_{n}, and :𝒞0n\mathcal{F}:\mathcal{C}_{0}\rightarrow\mathcal{L}_{n} be continuously differentiable and pseudomonotone on 𝒞0\mathcal{C}_{0}. Assume that F^\hat{F} defined in (10) satisfies the following conditions for any xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}):

(i)DF^(x) is symmetric;(ii)tr((DF^(x)Pϕ(n))2)=tr((DF^(x))2Pϕ(n));(iii)DF^(x)Pϕ(n) is not positive semidefinite.\begin{array}[]{l}(i)\ \text{$D\hat{F}(x)$ is symmetric};\\ (ii)\ {\rm tr}((D\hat{F}(x)P_{\phi(\mathcal{M}_{n})})^{2})={\rm tr}((D\hat{F}(x))^{2}P_{\phi(\mathcal{M}_{n})});\\ (iii)\ \text{$D\hat{F}(x)P_{\phi(\mathcal{M}_{n})}$ is not positive semidefinite.}\end{array} (18)

Let e^1=supxϕ(𝒞0)ρ(DF^(x))\hat{e}_{1}=\sup\limits_{x\in\phi(\mathcal{C}_{0})}\rho(D\hat{F}(x)). If e^1<+\hat{e}_{1}<+\infty, then the monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} is globally elicited at level se^1s\geq\hat{e}_{1}.

Proof.  On account of the fact that ρ(DF^(x))|λn¯(DF^(x))|\rho(D\hat{F}(x))\geq|\lambda_{\bar{n}}^{\downarrow}(D\hat{F}(x))|, it suffices to prove that the conditions (ii) and (iii) imply the conditions in Theorem 11.

Firstly, DF^(x)D\hat{F}(x) commutes with Pϕ(n)P_{\phi(\mathcal{M}_{n})} if and only if tr((DF^(x)Pϕ(n))2)=tr((DF^(x))2Pϕ(n))tr((D\hat{F}(x)P_{\phi(\mathcal{M}_{n})})^{2})={\rm tr}((D\hat{F}(x))^{2}P_{\phi(\mathcal{M}_{n})}) [31]. This is condition (ii).

Secondly, when DF^(x)Pϕ(n)D\hat{F}(x)P_{\phi(\mathcal{M}_{n})} is not positive semidefinite, condition (iii)(iii) in Theorem 11 holds true. In fact, suppose that there exist orthogonal matrix UU such that UTDF^(x)U=ΛAU^{T}D\hat{F}(x)U=\Lambda_{A} and UTPϕ(n)U=ΛBU^{T}P_{\phi(\mathcal{M}_{n})}U=\Lambda_{B}, where ΛA\Lambda_{A} and ΛB\Lambda_{B} are diagonal matrices and the existence of UU is assured by Lemma 2. Then we have

UTDF^(x)UUTPϕ(n)U=UTDF^(x)Pϕ(n)U=ΛAΛB.U^{T}D\hat{F}(x)UU^{T}P_{\phi(\mathcal{M}_{n})}U=U^{T}D\hat{F}(x)P_{\phi(\mathcal{M}_{n})}U=\Lambda_{A}\Lambda_{B}.

Thus DF^(x)Pϕ(n)D\hat{F}(x)P_{\phi(\mathcal{M}_{n})} and ΛAΛB\Lambda_{A}\Lambda_{B} have same eigenvalues. Since DF^(x)Pϕ(n)D\hat{F}(x)P_{\phi(\mathcal{M}_{n})} is not positive semidefinite, there exists at least one negative eigenvalue for DF^(x)Pϕ(n)D\hat{F}(x)P_{\phi(\mathcal{M}_{n})}, and so is ΛAΛB\Lambda_{A}\Lambda_{B}, as indicates condition (iii)(iii) in Theorem 11. \qed

Now we give some remarks to explain that the conditions in Theorem 11 may hold.

Remark 5

Since DF^(x)D\hat{F}(x) is block diagonal for xϕ(𝒞)x\in\phi(\mathcal{C}), i.e.,

DF^(x)=[DF(1p(ξ1)xp1,ξ1)0DF(1p(ξ2)xp2,ξ2)0DF(1p(ξJ)xpJ,ξJ)],D\hat{F}(x)=\begin{bmatrix}DF(\frac{1}{\sqrt{p(\xi^{1})}}x_{p_{1}},\xi^{1})&&&0\\ &DF(\frac{1}{\sqrt{p(\xi^{2})}}x_{p_{2}},\xi^{2})&&\\ &&\ddots&\\ 0&&&DF(\frac{1}{\sqrt{p(\xi^{J})}}x_{p_{J}},\xi^{J})\end{bmatrix},

there also exist some other criteria for condition DF^(x)Pϕ(n)=Pϕ(n)DF^(x)D\hat{F}(x)P_{\phi(\mathcal{M}_{n})}=P_{\phi(\mathcal{M}_{n})}D\hat{F}(x).

For instance, provided that Pϕ(n)P_{\phi(\mathcal{M}_{n})} is block diagonal, i.e.,

Pϕ(n)=[P10P20PJ],P_{\phi(\mathcal{M}_{n})}=\begin{bmatrix}P_{1}&&&0\\ &P_{2}&&\\ &&\ddots&\\ 0&&&P_{J}\end{bmatrix},

and

DF(1p(ξi)xpi,ξi)Pi=PiDF(1p(ξi)xpi,ξi),1iJ.DF(\frac{1}{\sqrt{p(\xi^{i})}}x_{p_{i}},\xi^{i})P_{i}=P_{i}DF(\frac{1}{\sqrt{p(\xi^{i})}}x_{p_{i}},\xi^{i}),\quad 1\leq i\leq J.

Then, it is easy to see that DF^(x)Pϕ(n)=Pϕ(n)DF^(x)D\hat{F}(x)P_{\phi(\mathcal{M}_{n})}=P_{\phi(\mathcal{M}_{n})}D\hat{F}(x).

Remark 6

Since projection matrix Pϕ(n)P_{\phi(\mathcal{M}_{n})} is usually sparse, the computation cost of the principal minors (or the eigenvalues) of matrix DF^(x)Pϕ(n)D\hat{F}(x)P_{\phi(\mathcal{M}_{n})} is relatively low, especially when F^\hat{F} is linear. In this case, it is easy to test whether DF^(x)Pϕ(n)D\hat{F}(x)P_{\phi(\mathcal{M}_{n})} is positive semidefinite since the matrix is constant.

Theorem 12

Consider SVI (4). Let 𝒞0𝒞\mathcal{C}_{0}\supset\mathcal{C} be an open convex set in n\mathcal{L}_{n}, and :𝒞0n\mathcal{F}:\mathcal{C}_{0}\rightarrow\mathcal{L}_{n} be continuously differentiable and pseudomonotone on 𝒞0\mathcal{C}_{0}. Let DF^(x)D\hat{F}(x) be symmetric with F^\hat{F} defined in (10) for xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}). If e2>0e_{2}>0 satisfies the condition that the multiplicity of the minimum eigenvalue of DF^(x)+e2Pϕ(n)D\hat{F}(x)+e_{2}P_{\phi(\mathcal{M}_{n})} is strictly larger than one for all xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}), then the monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} is globally elicited at level se2s\geq e_{2}.

Proof.  Let A=DF^(x)+e2Pϕ(n)A=D\hat{F}(x)+e_{2}P_{\phi(\mathcal{M}_{n})}, B=e2Pϕ(n)B=e_{2}P_{\phi(\mathcal{M}_{n})} and j=n¯1j=\bar{n}-1, i=n¯i=\bar{n}. It follows from Lemma 3 that

λn¯1(DF^(x)+e2Pϕ(n))λn¯(e2Pϕ(n))+λn¯1(DF^(x)).\lambda_{\bar{n}-1}^{\downarrow}(D\hat{F}(x)+e_{2}P_{\phi(\mathcal{M}_{n})})\geq\lambda_{\bar{n}}^{\downarrow}(e_{2}P_{\phi(\mathcal{M}_{n})})+\lambda_{\bar{n}-1}^{\downarrow}(D\hat{F}(x)).

By Lemma 1, λn¯1(DF^(x))0\lambda_{\bar{n}-1}^{\downarrow}(D\hat{F}(x))\geq 0. Based on the hypothesis on e2e_{2}, we have

λn¯(DF^(x)+e2Pϕ(n))=λn¯1(DF^(x)+e2Pϕ(n))0.\lambda_{\bar{n}}^{\downarrow}(D\hat{F}(x)+e_{2}P_{\phi(\mathcal{M}_{n})})=\lambda_{\bar{n}-1}^{\downarrow}(D\hat{F}(x)+e_{2}P_{\phi(\mathcal{M}_{n})})\geq 0.

By [31, Corollary 4.3.15], when se2s\geq e_{2}, it holds

λn¯(DF^(x)+sPϕ(n))λn¯(DF^(x)+e2Pϕ(n)),\lambda_{\bar{n}}^{\downarrow}(D\hat{F}(x)+sP_{\phi(\mathcal{M}_{n})})\geq\lambda_{\bar{n}}^{\downarrow}(D\hat{F}(x)+e_{2}P_{\phi(\mathcal{M}_{n})}),

which indicates the monotonicity of F^+sPϕ(n)\hat{F}+sP_{\phi(\mathcal{M}_{n})} on ϕ(𝒞0)\phi(\mathcal{C}_{0}). Via the same procedure in the proof of Theorem 11, +N𝒞+sPn\mathcal{F}+N_{\mathcal{C}}+sP_{\mathcal{M}_{n}} is maximal monotone. \qed

Example 3

Assume that stage NN and the cardinality of Ξ\Xi are both 1. Let F(x1,x2,x3)=(x1,0,0)TF(x_{1},x_{2},x_{3})=(-x_{1},0,0)^{T}, 𝒩={(0,l,t)T:l,t}\mathcal{N}=\{(0,l,t)^{T}:l,t\in\mathbb{R}\}, and 𝒞={(x1,x2,x3)T:x11,x21,x31}\mathcal{C}=\{(x_{1},x_{2},x_{3})^{T}:x_{1}\geq 1,x_{2}\geq 1,x_{3}\geq 1\}. Then SVI (4) is exhibited as F(x1,x2,x3)N𝒞𝒩(x1,x2,x3)-F(x_{1},x_{2},x_{3})\in N_{\mathcal{C}\cap\mathcal{N}}(x_{1},x_{2},x_{3}).

Obviously, DF(x1,x2,x3))=[1,0,0;0,0,0;0,0,0]DF(x_{1},x_{2},x_{3}))=[-1,0,0;0,0,0;0,0,0]. In addition, the orthogonal complement of 𝒩\mathcal{N} is ={(v,0,0)T:v}\mathcal{M}=\{(v,0,0)^{T}:v\in\mathbb{R}\} and the corresponding projection matrix P=[1,0,0;0,0,0;0,0,0]P_{\mathcal{M}}=[1,0,0;0,0,0;0,0,0].

Construct 𝒞0={(x1,x2,x3)T:x1>0,x2>0,x3>0}\mathcal{C}_{0}=\{(x_{1},x_{2},x_{3})^{T}:x_{1}>0,x_{2}>0,x_{3}>0\}. Then 𝒞0𝒞\mathcal{C}_{0}\supset\mathcal{C} is an open convex set. Obviously, F(x1,x2,x3)F(x_{1},x_{2},x_{3}) is pseudomonotone on 𝒞0\mathcal{C}_{0}. In fact, for (x1,x2,x3)T,(y1,y2,y3)T𝒞0(x_{1},x_{2},x_{3})^{T},(y_{1},y_{2},y_{3})^{T}\in\mathcal{C}_{0} with

F(x1,x2,x3),(y1x1,y2x2,y3x3)T0,\left<-F(x_{1},x_{2},x_{3}),(y_{1}-x_{1},y_{2}-x_{2},y_{3}-x_{3})^{T}\right>\geq 0,

we have y1x10y_{1}-x_{1}\leq 0 since

F(x1,x2,x3),(y1x1,y2x2,y3x3)T=x1(y1x1),x1>0.\left<-F(x_{1},x_{2},x_{3}),(y_{1}-x_{1},y_{2}-x_{2},y_{3}-x_{3})^{T}\right>=-x_{1}(y_{1}-x_{1}),\quad x_{1}>0.

This implies that

F(y1,y2,y3),(y1x1,y2x2,y3x3)T=y1(y1x1)0.\left<-F(y_{1},y_{2},y_{3}),(y_{1}-x_{1},y_{2}-x_{2},y_{3}-x_{3})^{T}\right>=-y_{1}(y_{1}-x_{1})\geq 0.

Thus F(x1,x2,x3)F(x_{1},x_{2},x_{3}) is pseudomonotone on 𝒞0\mathcal{C}_{0}.

On the other hand, F(x1,x2,x3)F(x_{1},x_{2},x_{3}) is not monotone on 𝒞0\mathcal{C}_{0} as a result of the negative eigenvalue of DF(x1,x2,x3)DF(x_{1},x_{2},x_{3}). But F+N𝒞+sPF+N_{\mathcal{C}}+sP_{\mathcal{M}} is globally maximal monotone at level s1s\geq 1 from Theorem 12, inasmuch as DF(x1,x2,x3)DF(x_{1},x_{2},x_{3}) is symmetric and the multiplicity of the minimum eigenvalue of DF(x1,x2,x3)DF(x_{1},x_{2},x_{3}) is 22.

Corollary 2

Consider SVI (4). Let 𝒞0𝒞\mathcal{C}_{0}\supset\mathcal{C} be an open convex set in n\mathcal{L}_{n}, and :𝒞0n\mathcal{F}:\mathcal{C}_{0}\rightarrow\mathcal{L}_{n} be continuously differentiable and pseudomonotone on 𝒞0\mathcal{C}_{0}. Let DF^(x)D\hat{F}(x) be symmetric with F^\hat{F} defined in (10) for xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}). If e^2>0\hat{e}_{2}>0 satisfies the condition that DF^(x)+e^2Pϕ(n)D\hat{F}(x)+\hat{e}_{2}P_{\phi(\mathcal{M}_{n})} is block diagonal with k(x)k(x) pairs of same blocks for all xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}), where k(x)1k(x)\geq 1 is an integer, i.e.,

DF^(x)+e^2Pϕ(n)=[A10A20A2k(x)],D\hat{F}(x)+\hat{e}_{2}P_{\phi(\mathcal{M}_{n})}=\begin{bmatrix}A_{1}&&&0\\ &A_{2}&&\\ &&\ddots&\\ 0&&&A_{2k(x)}\end{bmatrix},

and there is partition {i1,i2,,ik(x)}\{i_{1},i_{2},\ldots,i_{k(x)}\} and {j1,j2,,jk(x)}\{j_{1},j_{2},\ldots,j_{k(x)}\} of {1,,2k(x)}\{1,\ldots,2k(x)\} such that Ail=AjlA_{i_{l}}=A_{j_{l}}, 1lk(x)1\leq l\leq k(x), then the monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} is globally elicited at level se^2s\geq\hat{e}_{2}.

Proof.  It suffices to prove that the multiplicity of every eigenvalue of DF^(x)+e^2Pϕ(n)D\hat{F}(x)+\hat{e}_{2}P_{\phi(\mathcal{M}_{n})} is strictly larger than one. Actually, the eigenvalues of DF^(x)+e^2Pϕ(n)D\hat{F}(x)+\hat{e}_{2}P_{\phi(\mathcal{M}_{n})} are the aggregation of the eigenvalues of all blocks. Since two same blocks have same eigenvalues, the multiplicity of every eigenvalue of DF^(x)+e^2Pϕ(n)D\hat{F}(x)+\hat{e}_{2}P_{\phi(\mathcal{M}_{n})} is strictly larger than one. \qed

Theorem 13

Consider SVI (4). Let 𝒞0𝒞\mathcal{C}_{0}\supset\mathcal{C} be an open convex set in n\mathcal{L}_{n}, and :𝒞0n\mathcal{F}:\mathcal{C}_{0}\rightarrow\mathcal{L}_{n} be continuously differentiable and pseudomonotone on 𝒞0\mathcal{C}_{0}. Let DF^(x)D\hat{F}(x) be symmetric with F^\hat{F} defined in (10) for xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}). If e3>0e_{3}>0 satisfies the following conditions:

(i)(DF^(x))ii+e3(Pϕ(n))ii>0;(ii)DF^(x)+e3Pϕ(n) is strictly diagonally dominant,i.e.,\begin{array}[]{l}(i)(D\hat{F}(x))_{ii}+e_{3}(P_{\phi(\mathcal{M}_{n})})_{ii}>0;\\ (ii)\ \text{$D\hat{F}(x)+e_{3}P_{\phi(\mathcal{M}_{n})}$ is strictly diagonally dominant},i.e.,\end{array} (19)
(DF^(x))ii+e3(Pϕ(n))ii>ji|(DF^(x))ij+e3(Pϕ(n))ij|(D\hat{F}(x))_{ii}+e_{3}(P_{\phi(\mathcal{M}_{n})})_{ii}>\sum\limits_{j\neq i}|(D\hat{F}(x))_{ij}+e_{3}(P_{\phi(\mathcal{M}_{n})})_{ij}|

for i=1,2,,n¯i=1,2,\ldots,\bar{n} and xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}), then the monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} is globally elicited at level se3s\geq e_{3}.

Proof.  By Lemma 4, the matrix DF^(x)+e3Pϕ(n)D\hat{F}(x)+e_{3}P_{\phi(\mathcal{M}_{n})} is positive semidefinite. Hence, by [31, Corollary 4.3.15], DF^(x)+sPϕ(n)D\hat{F}(x)+sP_{\phi(\mathcal{M}_{n})} is positive semidefinite when se3s\geq e_{3}. Thus the monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} is globally elicited at level se3s\geq e_{3}. \qed

Note that the projection matrices have nonnegative diagonal elements and are usually diagonally dominant. Define the index set

I={1in¯:Pϕ(n))ii>0}.I=\{1\leq i\leq\bar{n}:P_{\phi(\mathcal{M}_{n})})_{ii}>0\}.

Let Pϕ(n)IP_{\phi(\mathcal{M}_{n})}^{I} be the submatrix of Pϕ(n)P_{\phi(\mathcal{M}_{n})} with the iith row and the iith column removed for iIi\notin I. We have the following result.

Corollary 3

Consider SVI (4). Let 𝒞0𝒞\mathcal{C}_{0}\supset\mathcal{C} be an open convex set in n\mathcal{L}_{n}, and :𝒞0n\mathcal{F}:\mathcal{C}_{0}\rightarrow\mathcal{L}_{n} be continuously differentiable and pseudomonotone on 𝒞0\mathcal{C}_{0}. Let DF^(x)D\hat{F}(x) be symmetric with F^\hat{F} defined in (10) for xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}). Assume that

(i)Pϕ(n)I is strictly diagonally dominant;(ii)the ith row of DF^(x) and Pϕ(n) are zero for all iI and xϕ(𝒞0).\begin{array}[]{rcl}(i)&&\text{$P_{\phi(\mathcal{M}_{n})}^{I}$ is strictly diagonally dominant};\\ (ii)&&\text{the $i$th row of $D\hat{F}(x)$ and $P_{\phi(\mathcal{M}_{n})}$ are zero for all $i\notin I$ and $x\in\phi(\mathcal{C}_{0})$.}\end{array}

Define

e^3=max{supxϕ(𝒞0),iIji|(DF^(x))ij|(DF^(x))ii(Pϕ(n))iiji|(Pϕ(n))ij|,supxϕ(𝒞0),iI{(DF^(x))ii}}.\hat{e}_{3}=\max\left\{\sup\limits_{x\in\phi(\mathcal{C}_{0}),i\in I}\frac{\sum\limits_{j\neq i}|(D\hat{F}(x))_{ij}|-(D\hat{F}(x))_{ii}}{(P_{\phi(\mathcal{M}_{n})})_{ii}-\sum\limits_{j\neq i}|(P_{\phi(\mathcal{M}_{n})})_{ij}|},\sup\limits_{x\in\phi(\mathcal{C}_{0}),i\in I}\{-(D\hat{F}(x))_{ii}\}\right\}.

If e^3<+\hat{e}_{3}<+\infty, then the monotonicity of +N𝒞\mathcal{F}+N_{\mathcal{C}} is globally elicited at level s>e^3s>\hat{e}_{3}.

Proof.  If s>e^3s>\hat{e}_{3}, then DF^I(x)+sPϕ(n)ID\hat{F}^{I}(x)+sP_{\phi(\mathcal{M}_{n})}^{I} is positive semidefinite for xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}) from Lemma 4. Since the iith row and the iith column of DF^(x)D\hat{F}(x) and Pϕ(n)P_{\phi(\mathcal{M}_{n})} are zero for all iIi\notin I and xϕ(𝒞0)x\in\phi(\mathcal{C}_{0}), the matrix DF^(x)+sPϕ(n)D\hat{F}(x)+sP_{\phi(\mathcal{M}_{n})} is positive semidefinite. Hence, the desired result holds. \qed

6 Numerical experiments

In this section, we demonstrate the effectiveness of the elicited PHA in solving a two-stage pseudomonotone stochastic linear complementarity problem (SLCP). The two-stage SLCP is given as a special case of

(x())w()N𝒞(x()),-\mathcal{F}(x(\cdot))-w(\cdot)\in N_{\mathcal{C}}(x(\cdot)),

where F(x(ξ),ξ)=M(ξ)x(ξ)+q(ξ)F(x(\xi),\xi)=M(\xi)x(\xi)+q(\xi) with M(ξ)n×nM(\xi)\in\mathbb{R}^{n\times n} and q(ξ)nq(\xi)\in\mathbb{R}^{n}, C(ξ)=+nC(\xi)=\mathbb{R}^{n}_{+}, and x(ξ)=(x1(ξ)T,x2(ξ)T)Tx(\xi)=(x_{1}(\xi)^{T},x_{2}(\xi)^{T})^{T} with xi(ξ)nix_{i}(\xi)\in\mathbb{R}^{n_{i}} being the ii-th stage decision vector for i=1,2i=1,2. In this model, the nonanticipativity subspace is described as 𝒩n={x()=(x1(),x2()):x1(ξ)is same for allξΞ}\mathcal{N}_{n}=\{x(\cdot)=(x_{1}(\cdot),x_{2}(\cdot)):x_{1}(\xi)\ \text{is same for all}\ \xi\in\Xi\}, and the corresponding complement is n={w()=(w1(),w2()):Eξ(w1(ξ))=0,w2(ξ)=0for allξΞ}\mathcal{M}_{n}=\{w(\cdot)=(w_{1}(\cdot),w_{2}(\cdot)):E_{\xi}(w_{1}(\xi))=0,w_{2}(\xi)=0\ \text{for all}\ \xi\in\Xi\}. By denoting

M(ξ)=[M11(ξ)M12(ξ)M21(ξ)M22(ξ)],q(ξ)=[q1(ξ)q2(ξ)]M(\xi)=\left[\begin{array}[]{cc}M_{11}(\xi)&M_{12}(\xi)\\ M_{21}(\xi)&M_{22}(\xi)\end{array}\right],\ q(\xi)=\left[\begin{array}[]{c}q_{1}(\xi)\\ q_{2}(\xi)\end{array}\right]

for ξΞ\xi\in\Xi, where Mij(ξ)ni×njM_{ij}(\xi)\in\mathbb{R}^{n_{i}\times n_{j}}, qi(ξ)niq_{i}(\xi)\in\mathbb{R}^{n_{i}} with 1i,j21\leq i,j\leq 2, the extensive form of the two-stage SLCP is formulated as

{0x1M11(ξ)x1+M12(ξ)x2(ξ)+q1(ξ)+w1(ξ)0,0x2(ξ)M21(ξ)x1+M22(ξ)x2(ξ)+q2(ξ)0ξΞ.\left\{\begin{array}[]{l}0\leq x_{1}\perp M_{11}(\xi)x_{1}+M_{12}(\xi)x_{2}(\xi)+q_{1}(\xi)+w_{1}(\xi)\geq 0,\\ 0\leq x_{2}(\xi)\perp M_{21}(\xi)x_{1}+M_{22}(\xi)x_{2}(\xi)+q_{2}(\xi)\geq 0\quad\forall\xi\in\Xi.\end{array}\right. (20)

In the execution of Algorithm 1, we use the semismooth Newton method [22] to solve (13). It is worth mentioning that we take x^k(ξ)\hat{x}^{k}(\xi) as the starting point in the k+1k+1 iteration, which is termed the “warm start” feature of PHA in [26]. In addition, inspired by [38] and [6], we change the step size in the dual update to be

wk+1(ξ)=wk(ξ)+ρ(rs)(x^k(ξ)xk+1(ξ)),w^{k+1}(\xi)=w^{k}(\xi)+\rho(r-s)({\hat{x}}^{k}(\xi)-x^{k+1}(\xi)),

where ρ=1.618\rho=1.618, and ss is supposed to be r/2r/2.

If x(ξ)x(\xi) and w(ξ)w(\xi) solve (13), then x(ξ)x(\xi) and w(ξ)w(\xi) also satisfy

{0x1Eξ(M11(ξ)x1)+Eξ(M12(ξ)x2(ξ))+Eξ(q1(ξ))0,0x2(ξ)M21(ξ)x1+M22(ξ)x2(ξ)+q2(ξ)0ξΞ,\left\{\begin{array}[]{l}0\leq x_{1}\perp E_{\xi}(M_{11}(\xi)x_{1})+E_{\xi}(M_{12}(\xi)x_{2}(\xi))+E_{\xi}(q_{1}(\xi))\geq 0,\\ 0\leq x_{2}(\xi)\perp M_{21}(\xi)x_{1}+M_{22}(\xi)x_{2}(\xi)+q_{2}(\xi)\geq 0\quad\forall\xi\in\Xi,\end{array}\right. (21)

as is obtained by taking the expectation in the first subproblem of (20). Thus the stopping criterion can be designated as

err=max{err1,err2},err=\max\{err_{1},err_{2}\},

where

err1=x1Π0(x1(Eξ(M11(ξ)x1)+Eξ(M12(ξ)x2(ξ))+Eξ(q1(ξ))))1+x1,err_{1}=\frac{\|x_{1}-\Pi_{\geq 0}(x_{1}-(E_{\xi}(M_{11}(\xi)x_{1})+E_{\xi}(M_{12}(\xi)x_{2}(\xi))+E_{\xi}(q_{1}(\xi))))\|}{1+\|x_{1}\|},
err2=maxξ{x2(ξ)Π0(x2(ξ)(M21(ξ)x1+M22(ξ)x2(ξ)+q2(ξ)))1+x2(ξ)},err_{2}=\max\limits_{\xi}\left\{\frac{\|x_{2}(\xi)-\Pi_{\geq 0}(x_{2}(\xi)-(M_{21}(\xi)x_{1}+M_{22}(\xi)x_{2}(\xi)+q_{2}(\xi)))\|}{1+\|x_{2}(\xi)\|}\right\},

with (Π0(a))j=max{aj,0}(\Pi_{\geq 0}(a))_{j}=\max\{a_{j},0\}. Set the tolerance to be 10510^{-5} and the maximal iterations to be 20002000. In the next subsections, we apply Algorithm 1 to solve a two-stage pseudomonotone SVI from real life and some randomly generated problems.

6.1 Test on a two-stage orange market model

Consider a two-stage orange market model in [19], where the supply and demand curves are linear. Specifically, an orange firm mainly sells two kinds of products. One is the juice converted from oranges, and the other one is exactly the fresh oranges. Assume that the producer makes 11 unit of juice from 22 oranges. In the first stage, the firm has to decide the supply quantity (QSQ_{S}) with the supply price (PSP_{S}) determined as follows:

PS=3+0.0005QS.P_{S}=3+0.0005Q_{S}. (22)

In the second stage, the orange firm needs to determine the quantity of the juice (QJQ_{J}) and the quantity of fresh oranges (QFQ_{F}). Similarly, the price for orange juice (PJP_{J}) and price for fresh oranges (PFP_{F}) are related linearly to quantity QJQ_{J} and QFQ_{F}, i.e.,

[PJPF]=M[QJQF]+q,\begin{bmatrix}P_{J}\\ P_{F}\end{bmatrix}=M\begin{bmatrix}Q_{J}\\ Q_{F}\end{bmatrix}+q, (23)

where M2×2M\in\mathbb{R}^{2\times 2} and q2×1q\in\mathbb{R}^{2\times 1}. However, due to the uncertainties involving the climate, natural disaster, water resources and so on, linear relationship (23) may not be fixed. In our setting, three scenarios of uncertainties, ξ1\xi^{1}, ξ2\xi^{2}, ξ3\xi^{3}, with respective probabilities, 0.5, 0.3, 0.2, are considered. Then we denote the quantity variables QJQ_{J} and QFQ_{F} and price variables PJP_{J} and PFP_{F} for every scenario ξΞ\xi\in\Xi by QJ(ξ)Q_{J}(\xi), QF(ξ)Q_{F}(\xi) and PJ(ξ)P_{J}(\xi), PF(ξ)P_{F}(\xi) respectively, where Ξ:={ξ1,ξ2,ξ3}\Xi:=\{\xi^{1},\xi^{2},\xi^{3}\}. Define MM and qq, for ξΞ\xi\in\Xi, as

M(ξ1)=[0.0050.00020.00020.001],q(ξ2)=[7.54],\displaystyle M(\xi^{1})=\begin{bmatrix}-0.005&-0.0002\\ -0.0002&-0.001\end{bmatrix},q(\xi^{2})=\begin{bmatrix}7.5\\ 4\end{bmatrix}, (24)
M(ξ2)=[0.0040.00010.00010.0005],q(ξ2)=[73.5],\displaystyle M(\xi^{2})=\begin{bmatrix}-0.004&-0.0001\\ -0.0001&-0.0005\end{bmatrix},q(\xi^{2})=\begin{bmatrix}7\\ 3.5\end{bmatrix},
M(ξ3)=[0.0060.00030.00030.0015],q(ξ3)=[84.5].\displaystyle M(\xi^{3})=\begin{bmatrix}-0.006&-0.0003\\ -0.0003&-0.0015\end{bmatrix},q(\xi^{3})=\begin{bmatrix}8\\ 4.5\end{bmatrix}.

We build up the following two-stage optimization model:

min\displaystyle\min PSQS+Eξ[ϕ(QS,QJ(ξ),QF(ξ))]\displaystyle P_{S}Q_{S}+E_{\xi}[\phi(Q_{S},Q_{J}(\xi),Q_{F}(\xi))] (25)
s.t. QS0,\displaystyle Q_{S}\geq 0,

where ϕ(QS,QJ(ξ),QF(ξ))\phi(Q_{S},Q_{J}(\xi),Q_{F}(\xi)) is the optimal value of the second-stage optimization:

min\displaystyle\min PJ(ξ)QJ(ξ)PF(ξ)QF(ξ)\displaystyle-P_{J}(\xi)Q_{J}(\xi)-P_{F}(\xi)Q_{F}(\xi) (26)
s.t. 2QJ(ξ)+QF(ξ)QS.\displaystyle 2Q_{J}(\xi)+Q_{F}(\xi)\leq Q_{S}.

Substituting (22)-(24) into (25)-(26), we get

min\displaystyle\min 0.0005QS2+3QS+Eξ[ϕ(QS,QJ(ξ),QF(ξ))]\displaystyle 0.0005Q_{S}^{2}+3Q_{S}+E_{\xi}[\phi(Q_{S},Q_{J}(\xi),Q_{F}(\xi))] (27)
s.t. QS0,\displaystyle Q_{S}\geq 0,

where ϕ(QS,QJ(ξ),QF(ξ))\phi(Q_{S},Q_{J}(\xi),Q_{F}(\xi)) is the optimal value of the second-stage optimization:

min\displaystyle\min [QJ(ξ)QF(ξ)]M(ξ)[QJ(ξ)QF(ξ)]q(ξ)T[QJ(ξ)QF(ξ)]\displaystyle-\begin{bmatrix}Q_{J}(\xi)&Q_{F}(\xi)\end{bmatrix}M(\xi)\begin{bmatrix}Q_{J}(\xi)\\ Q_{F}(\xi)\end{bmatrix}-q(\xi)^{T}\begin{bmatrix}Q_{J}(\xi)\\ Q_{F}(\xi)\end{bmatrix} (28)
s.t. QS[21][QJ(ξ)QF(ξ)]0.\displaystyle Q_{S}-\begin{bmatrix}2&1\end{bmatrix}\begin{bmatrix}Q_{J}(\xi)\\ Q_{F}(\xi)\end{bmatrix}\geq 0.

The necessary optimality condition of (27)-(28) is that there exist ω()\omega(\cdot) and dual vector η()\eta(\cdot) such that, for ξΞ\xi\in\Xi, the following condition holds [38]:

0[QSQJ(ξ)QF(ξ)η(ξ)][0.00100102M11(ξ)2M12(ξ)202M21(ξ)2M22(ξ)11210][QSQJ(ξ)QF(ξ)η(ξ)]+[3q1(ξ)q2(ξ)0]+[ω(ξ)000]0,\displaystyle 0\leq\begin{bmatrix}Q_{S}\\ Q_{J}(\xi)\\ Q_{F}(\xi)\\ \eta(\xi)\end{bmatrix}\perp\begin{bmatrix}0.001&0&0&-1\\ 0&2M_{11}(\xi)&2M_{12}(\xi)&2\\ 0&2M_{21}(\xi)&2M_{22}(\xi)&1\\ 1&-2&-1&0\end{bmatrix}\begin{bmatrix}Q_{S}\\ Q_{J}(\xi)\\ Q_{F}(\xi)\\ \eta(\xi)\end{bmatrix}+\begin{bmatrix}3\\ q_{1}(\xi)\\ q_{2}(\xi)\\ 0\end{bmatrix}+\begin{bmatrix}\omega(\xi)\\ 0\\ 0\\ 0\end{bmatrix}\geq 0, (29)

which is exactly a two-stage SLCP with x1(ξ)=QSx_{1}(\xi)=Q_{S}, x2(ξ)=[QJ(ξ),QF(ξ),η(ξ)]Tx_{2}(\xi)=[Q_{J}(\xi),Q_{F}(\xi),\eta(\xi)]^{T}, C(ξ)=+4C(\xi)=\mathbb{R}_{+}^{4}, and

F(x(ξ),ξ)=[0.00100102M11(ξ)2M12(ξ)202M21(ξ)2M22(ξ)11210]x(ξ)+[3q1(ξ)q2(ξ)0].F(x(\xi),\xi)=\begin{aligned} \begin{bmatrix}0.001&0&0&-1\\ 0&2M_{11}(\xi)&2M_{12}(\xi)&2\\ 0&2M_{21}(\xi)&2M_{22}(\xi)&1\\ 1&-2&-1&0\end{bmatrix}x(\xi)+\begin{bmatrix}3\\ q_{1}(\xi)\\ q_{2}(\xi)\\ 0\end{bmatrix}.\end{aligned}

Obviously, (x())\mathcal{F}(x(\cdot)) is pseudomonotone, and its monotonicity can be globally elicited at any level s0s\geq 0 via Theorem 10. We use Algorithm 1 to solve (29). The algorithm ends up with the solution: QS=393Q_{S}=393, QJ(ξ1)=56Q_{J}(\xi^{1})=56, QF(ξ1)=281Q_{F}(\xi^{1})=281, QJ(ξ2)=64Q_{J}(\xi^{2})=64, QF(ξ2)=265Q_{F}(\xi^{2})=265, QJ(ξ3)=52Q_{J}(\xi^{3})=52, QF(ξ3)=288Q_{F}(\xi^{3})=288. The corresponding price solution is PS=4.96P_{S}=4.96, PJ(ξ1)=7.16P_{J}(\xi^{1})=7.16, PF(ξ1)=3.71P_{F}(\xi^{1})=3.71, PJ(ξ2)=6.72P_{J}(\xi^{2})=6.72, PF(ξ2)=3.36P_{F}(\xi^{2})=3.36, PJ(ξ3)=7.60P_{J}(\xi^{3})=7.60, QF(ξ3)=4.05Q_{F}(\xi^{3})=4.05.

6.2 Test on randomly generated problems

We implement Algorithm 1 to solve the randomly generated numerical SLCPs. Assume that the space Ξ\Xi has JJ scenarios denoted as ξ1,ξ2,,ξJ\xi^{1},\xi^{2},\ldots,\xi^{J}. Then the two-stage pseudomonotone SLCPs are generated randomly based on Corollary 6.6.2 in [1].

  • 1.

    Set M(ξ1)=abT+baTM(\xi^{1})=ab^{T}+ba^{T}, q(ξ1)=b0a+a0b+cq(\xi^{1})=b_{0}a+a_{0}b+c, and c=αa+βbc=\alpha a+\beta b, where a,bna,b\in\mathbb{R}^{n} with a0,b0a\geq 0,b\leq 0 are linearly independent, a0,b0a_{0},b_{0}\in\mathbb{R} are randomly generated, and α,β\alpha,\beta is arbitrarily selected as long as α<b0\alpha<-b_{0} and β>a0\beta>-a_{0}.

  • 2.

    Set s=3n/4s=\lceil 3n/4\rceil. Randomly generate number ai(ξk)>0a_{i}(\xi^{k})>0 and vector vi(ξk)nv_{i}(\xi^{k})\in\mathbb{R}^{n} for i=1,2,,si=1,2,\ldots,s and k=2,,Jk=2,\ldots,J. Let M(ξk)=i=1sai(ξk)vi(ξk)vi(ξk)TM(\xi^{k})=\sum\limits^{s}_{i=1}a_{i}(\xi^{k})v_{i}(\\ \xi^{k})v_{i}(\xi^{k})^{T}. Randomly generate q(ξk)nq(\xi^{k})\in\mathbb{R}^{n} for k=2,,Jk=2,\ldots,J.

  • 3.

    Randomly generate the probabilities p(ξk)>0p(\xi^{k})>0 for k=1,2,,Jk=1,2,\ldots,J.

We test three groups of two-stage pseudomonotone SLCPs listed as follows:

  • G1:

    The dimensions of the problems (dimdim for short) are set to be [40,20][40,20]. The number of the scenarios (snsn for short) in sample space Ξ\Xi is increased from 5050 to 400400. 10 numerical examples are randomly generated for every setting of the problems. The numerical results including the average convergence iteration number (avg-iter for short) and average convergence iteration time (avg-time for short) are presented in Table 1 and Fig. 1.

  • G2:

    The number of the scenarios in sample space Ξ\Xi is set to be 5050. The dimensions of the problems are increased from [50,50][50,50] to [400,400][400,400]. 10 numerical examples are randomly generated for every setting of the problems. The numerical results including the average convergence iteration number and average convergence iteration time are presented in Table 2 and Fig. 2.

  • G3:

    Since the choice of rr is crucial in terms of the effectiveness of Algorithm 1 [26], we set the same r=1r=1 and r=n1+n2r=\sqrt{n_{1}+n_{2}} as [26], and randomly generate 1010 numerical examples respectively for each setting of the former two groups of problems.

Remark 7

Note that F(x(ξ),ξ)F(x(\xi),\xi) is pseudomonotone if (x())\mathcal{F}(x(\cdot)) is pseudomonotone, but the converse may be not true. Then, the set of pseudomonotone mappings (x())\mathcal{F}(x(\cdot)) is contained in the set of mappings (x())\mathcal{F}(x(\cdot)) with F(x(ξ),ξ)F(x(\xi),\xi) being pseudomonotone for ξΞ\xi\in\Xi. So, it is reasonable to design the above numerical experiments.

Table 1: Numerical results for the change of snsn (dimdim=[40,20])
snsn r=1r=1 r=60r=\sqrt{60}
avg-iter avg-time(s) avg-iter avg-time(s)
50 255.4 4.7 82.4 1.5
100 266.3 9.7 84.5 3.1
150 278.6 15.3 87.9 4.9
200 291.6 21.3 93.7 6.9
250 305.4 27.6 94.0 8.5
300 331.4 35.8 96.4 10.5
350 345.4 44.2 92.4 11.8
400 347.5 52.9 96.7 14.8
Refer to caption
Refer to caption
Figure 1: Convergence results when snsn increases
Table 2: Numerical results for the change of dimdim (snsn=50)
dimdim r=1r=1 r=n1+n2r=\sqrt{n_{1}+n_{2}}
avg-iter avg-time(s) avg-iter avg-time(s)
[50,50] 247.2 5.2 62.2 1.4
[100,100] 473.4 13.7 37.4 1.2
[150,150] 627.9 24.8 29.1 1.4
[200,200] 840.0 62.9 27.7 2.6
[250,250] 958.0 99.0 35.0 4.3
[300,300] 1177.8 200.0 42.4 8.2
[350,350] 1462.4 326.8 50.0 12.7
[400,400] 1575.3 536 52.3 20.0
Refer to caption
Refer to caption
Figure 2: Convergence results when dimdim increases

We first fix the dimension of the decision vector to see the performance of Algorithm 1. From Table 1 and Fig.1, we can see that when r=1r=1, the iteration number and the computation time both grow roughly at a linear rate with the increasing of the scenario number. When r=60r=\sqrt{60}, the computation time also grow roughly at a linear rate when the scenario number increases, while the growth rate of the iteration number is pretty slow. The performance of the elicited PHA with r=60r=\sqrt{60} is much better than that with r=1r=1, which match with the numerical results given in [26]. Actually, the gap concerning the iteration number and the computation time between two choices of rr is widening with the increasing of the scenario number.

On the other hand, we fix the scenario number at 5050 and observe the influence of the dimension of the decision vector on the performance of Algorithm 1. From Table 2 and Fig.2, the growth rates of the iteration number and the computation time are steady with the increase of the scenario number when r=1r=1. However, when r=n1+n2r=\sqrt{n_{1}+n_{2}}, the iteration number remains stable and is around 4040. The computation time also grows very slowly when the scenario number increases. Similarly, the gap between two choices of rr is widening with the increasing of the dimension.

In summary, the above numerical experiments show that the elicited PHA is effective for the pseudomonotone two-stage SLCPs. Even for the relatively large cases, the elicited PHA can find the solution in a reasonable amount of time. When it comes to the choice of rr, the performance of elicited PHA with r=n1+n2r=\sqrt{n_{1}+n_{2}} is much better than that with r=1r=1, at least in our setting.

Remark 8

The results in our numerical experiments is basically consistent with the results in [26, Table 2, 4], but is slightly different from the results in [33, Table 5, 6], where the numerical results show that increasing parameter rr leads to more iterations and convergence time. Nevertheless, it is worth noting that the settings of the experiments in our paper and [33] are different, where s=r/2s=r/2 in our paper while s=r1s=r-1 in [33]. Actually, parameter rr with value n1+n2\sqrt{n_{1}+n_{2}} deserves to be explored further for its possible advantages.

7 Conclusions

We studied the multistage pseudomonotone SVI. First, we established some theoretical results on the existence, convexity, boundedness and compactness of the solution set based on constructing the isomorphism between Hilbert space n\mathcal{L}_{n} and Euclidean space Rn¯R^{\bar{n}}. Second, aiming at extension of the PHA from monotone SVI problems to nonmonotone ones, we presented some sufficient conditions on the elicitability of the pseudomonotone SVIs, which opens the door for applying Rockafellar’s elicited (nonmonotone) PHA to solve pseudomonotone SVIs. Numerical results on solving a pseudomonotone two-stage stochastic market optimization problem and experimental results on solving randomly generated two-stage SLCPs showed that the efficiency of the elicited PHA for solving pseudomonotone SVIs.

References

  • [1] A. Cambini and L. Martein, Generalized Convexity and Optimization: Theory and Applications, Springer, Berlin, 2009.
  • [2] X.J. Chen, T.K. Pong, R.J.-B. Wets, Two-stage stochastic variational inequalities: an ERM-solution procedure, Math. Program., 165 (2017), pp. 71–111.
  • [3] X.J. Chen, H.L. Sun, H.F. Xu, Discrete approximation of two-stage stochastic and distributionally robust linear complementarity problems, Math. Program., 177 (2018), pp. 255–289.
  • [4] X.J. Chen and M. Fukushima, Expected residual minimization method for stochastic linear complementarity problems, Math. Oper. Res., 30 (2005), pp. 1022–1038.
  • [5] X.J. Chen, C. Zhang, M. Fukushima, Robust solution of monotone stochastic linear complementarity problems, Math. Program, 117 (2009), pp. 51–80.
  • [6] J. Eckstein and D.P. Bertsekas, On the Douglas-Rachford splitting method and the proximal point algorithm for maximal monotone operators, Math. Program., 55 (1992), pp. 293–318.
  • [7] F. Facchinei and J.-S. Pang, Finite-dimensional variational inequalities and complementarity problems, Springer, New York, 2003.
  • [8] H.T. Fang, X.J. Chen, M. Fukushima, Stochastic R0R_{0} matrix linear complementarity problem, SIAM J. Optim., 18 (2007), pp. 82–506.
  • [9] G. Gürkan, A.Y. Özge, S.M. Robinson, Sample-path solution of stochastic variational inequalities, Math. Program., 84 (1999), pp. 313–333.
  • [10] A. Iusem, A. Jofré, R.I. Oliveira, P. Thompson, Extragradient method with variance reduction for stochastic variational inequalities, SIAM J. Optim., 27 (2017), pp. 686–724.
  • [11] A. Iusem, A. Jofré, P. Thompson, Incremental constraint projection methods for monotone stochastic variational inequalities, Math. Oper. Res., 44 (2018), pp. 236–263.
  • [12] A. Iusem, A. Jofré, R.I. Oliveira, P. Thompson, Variance-based stochastic extragradient methods with line search for stochastic variational inequalies, SIAM J. Optim., 29 (2019), pp. 175–206.
  • [13] H.Y. Jiang and H.F. Xu, Stochastic approximation approaches to the stochastic variational inequality problem, IEEE Trans. Autom. Control, 53 (2008), pp. 1462–1475.
  • [14] A. Kannan and U.V. Shanbhag, Optimal stochastic extragradient schemes for pseudomonotone stochastic variational inequality problems and their variants, Comput. Optim. Appl., 74 (2019), pp. 779–820.
  • [15] A.J. King and R.T. Rockafellar, Asymptotic theory for solutions in statistical estimation and stochastic programming, Math. Oper. Res., 18 (1993), pp. 148–162.
  • [16] C. Ling, L. Qi, G. Zhou, L. Caccetta, The SC1SC^{1} property of an expected residual function arising from stochastic complementarity problems, Oper. Res. Lett., 36 (2008), pp. 456–460.
  • [17] M.J. Luo and G.H. Lin, Expected residual minimization method for stochastic variational inequality problems, J. Optim. Theory Appl., 140 (2009), pp. 103–116.
  • [18] S. Lu, Confidence regions for stohastic variational inequalities, Math. Oper. Res., 38 (2013), pp. 545–568.
  • [19] B.A. McCarl, H. Moskowitz, H. Furtan, Quadratic programming applications, Int. J. Mgmt. Sci., 5 (1977), pp. 43–55.
  • [20] J.-S. Pang, S. Sen , U. Shanbhag, Two-stage non-cooperative games with risk-averse players. Math. Program., 165 (2017), pp.  235-290.
  • [21] G.C. Philippe, Linear and Nonlinear Functional Analysis with Applications, SIAM, Philadelphia, 2013.
  • [22] L. Qi and J. Sun, A nonsmooth version of Newton’s method, Math. Program., 58 (1993), pp. 353–367.
  • [23] U. Ravat and U.V. Shanbhag, On the characterization of solution sets of smooth and nonsmooth convex stochastic Nash games, SIAM J. Optim., 21 (2011), pp. 1168–1199.
  • [24] R.T. Rockafellar and R.J.-B. Wets, Stochastic variational inequalities: single-stage to multistage, Math. Program., 165 (2017), pp. 291–330.
  • [25] R.T. Rockafellar and R.J.-B. Wets, Variational Analysis, Springer, Berlin, 1998.
  • [26] R.T. Rockafellar and J. Sun, Solving monotone stochastic variational inequalities and complementarity problems by progressive hedging, Math. Program., 174 (2019), pp. 453–471.
  • [27] R.T. Rockafellar and J. Sun, Solving Lagrangian variational inequalities with applications to stochastic programming, Math. Program., 181 (2020), pp. 435–451.
  • [28] R.T. Rockafellar, On the maximality of sums of nonlinear monotone operators, Trans. Amer. Math. Soc., 149 (1970), pp. 75–88.
  • [29] R.T. Rockafellar, Monotone operators and the proximal point algorithm, SIAM J. Control Optim., 14(5) (1976), pp. 877-898.
  • [30] R.T. Rockafellar, Progressive decoupling of linkages in optimization and variational inequalities with elicitable convexity or monotonicity, Set-Valued Var. Anal., 27 (2018), pp. 863–893.
  • [31] A.H. Roger and R.J. Charles, Matrix Analysis, Cambridge University Press, Cambridge, 2013.
  • [32] A. Shapiro, D. Dentcheva, A. Ruszczyński, Lectures on Stochastic Programming: Modeling and Theory, MPS-SIAM Series on Optimization, SIAM, Philadelphia, 2009.
  • [33] J. Sun and M. Zhang, The elicited progressive decoupling algorithm: a note on the rate of convergence and a preliminary numerical experiment on the choice of parameters, submitted to Set-Valued Var. Anal., (2021).
  • [34] H.F. Xu, Sample average approximation methods for a class of variational inequality problems, Asia-Pac. J. Oper. Res., 27 (2010), pp. 103–119.
  • [35] F. Yousefian, A. Nedić, U.V. Shanbhag, On stochastic mirror-prox algorithms for stochastic Cartesian variational inequalities: randomized block coordinate and optimal averaging schemes, Set-Valued Var. Anal., 26 (2018), pp. 789–819.
  • [36] C. Zhang and X.J. Chen, Smoothing projected gradient method and its application to stochastic linear complementarity problems, SIAM J. Optim., 20 (2009), pp. 627–649.
  • [37] M. Zhang, L.S. Hou, J. Sun, A.L. Yan, A model of multistage risk-averse stochastic optimization and its solution by scenario-based decomposition algorithms, Asia-Pac. J. Oper. Res., 37(4) (2020), 2040004 doi:10.1142/S0217595920400047.
  • [38] M. Zhang, J. Sun, H.L. Xu, Two-stage quadratic games under uncertainty and their solution by progressive hedging algorithms, SIAM J. Optim., 29 (2019), pp. 1799–1818.