This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Continuous and Discrete Systems for Quasi Variational Inequalities with Application to Game Theory

Oday Hazaimah 111E-mail: odayh982@yahoo.com. https://orcid.org/0009-0000-8984-2500. St. Louis, MO, USA
Abstract

A new class of projected dynamical systems of third order is investigated for quasi (parametric) variational inequalities in which the convex set in the classical variational inequality also depends upon the solution explicitly or implicitly. We study the stability of a continuous method of a gradient type. Some iterative implicit and explicit schemes are suggested as counterparts of the continuous case by inertial proximal methods. The convergence analysis of these proposed methods is established under sufficient mild conditions. Moreover, some applications dealing with the generalized Nash equilibrium problems are presented.

Keywords: Quasi-variational Inequalities; Projected Dynamical Systems; Discretization, Generalized Nash Equilibrium Games.

Mathematics Subject Classification: 34B15, 34B16, 65L05, 65L10, 65L11.

1 Introduction

Variational inequalities is a general mathematical framework arising naturally in many theoretical and applied fields, such as finance, engineering, mechanics and operations research [7, 8, 9, 10, 11, 12, 13, 17]. Variational inequalities involving a nonlinear term is called the mixed variational inequality while variational inequalities with a moving constraint set (i.e., the convex set in the variational inequality depends upon the solution) is called the parametric quasivariational inequality (QVI) which was first introduced by Bensoussan et al. [6] and it will be our paradigm of study in this work. Note that if the involved set does not depend upon the solution then quasi variational inequality reduces to the Stampacchia variational inequality [27]. Variational inequality can be formulated in terms of dynamical systems to study the existence and stability of the solution [9, 14, 24]. It is well known that the solution of the variational inequality exists if either the constraint set is bounded or the corresponding mapping is strongly monotone. In recent years, the study of dynamical systems associated with variational inequalities provides qualitative insights for analyzing complex dynamics and optimizing systems by using methods of resolvent operators and projection operators over a set including; the proximal point algorithm, the gradient projection algorithm, among others (see, for instance, [1, 2, 5, 18, 20, 13, 24]). The convergence of the projection methods requires that the operator must be strongly monotone and Lipschitz continuous. Dynamical systems theory goes further than finite-dimensional variational inequalities since it allows for the study of the dynamics of equilibrium problems. The equilibrium points of a dynamical system form the solution set of the corresponding variational inequality problem. Unlike the existing extensive literature on variational inequalities, there is not much theory for parametric constraint situations.

Quasi-variational inequalities are known to be very useful for the modelling and analysis of many problems of equilibrium optimization and game theory. Solving QVI (1) by means of the optimization reformulation (4) provides global optimal solutions for general nonconvex objective functions. However, it is not easy to find global minimizers on the feasible set. Unlike in the particular case, the variational inequality, strict monotonicity is given on the objective function under which any equilibrium point of a dynamical system solves the variational inequality. Therefore it would be an interesting subject of future research to develop an optimization reformulation of the QVI that possesses the following property: any equilibrium point is essentially a global optimal solution. Several techniques using the notion of dynamical systems have been studied for approximating QVI, for instance, Khan et al. [16] considered second order dynamical system associated with quasi variational inequalities by applying some forward finite difference schemes. Mijajlovic and Jacimovic used proximal method in [18], while the same authors considered continuous methods for solving quasi-variational inequalities in [19]. Mijajlovic et al. in [20] used gradient-type projection methods for quasi variational inequalities. As it is known that the variational inequality can be extended to the Nash equilibrium game, generalized Nash equilibrium problems (GNEP) can be formulated in terms of QVIs. This formulation is referred to Bensoussan [6] in which each player’s strategy set depends on the rival players strategies. The QVI has recently attracted growing attention in connection to game theory. Moreover, necessary and sufficient conditions for Nash equilibria of a game in normal strategic form cane be constructed in terms of the generated optimization problems.

This paper aims at proposing a new continuous-time method of the third order and based on this design we also derive a variety of implicit and explicit discrete-time algorithms for solving parametric quasi-variational inequalities. Inspired by the applications of third-order ODEs which are used to describe and model the motion in electrical circuits [11], we adopt this idea in the manuscript. The suggested third-order projected dynamical system technique is very similar to the one used by Hazaimah [14] for mixed variational inequalities. It is worth noted that the resolvent dynamical system was proposed for mixed variational inequalities while in this note, the projected dynamical system is proposed for quasi variational inequalities. Moreover, the coefficients of the inertial and damping terms are constants in [14] while the coefficients in our case are time-dependent. To the best of our knowledge, this work is the first to use third-order dynamical systems to model QVIs by projection operators. Thus, our aim can be summarized as: (i) analysing a continuous gradient-type method with the most applicable form of the moving constraint set. (ii) using finite difference processes to identify the class of QVI by implicit and explicit discretizations for the associated dynamical system represented in terms of projection operators, (iii) discuss the global stability for solutions of the third-order dynamical system, and finally (iv) discuss some applications in the eyes of QVIs.

2 Mathematical Preliminaries

Some mathematical foundations and significant definitions are presented in this section from monotone operators theory, dynamical systems theory, convex analysis and variational inequalities, see [17] for more details. Let \mathcal{H} be a real Hilbert space equipped with inner product ,\langle\cdot,\cdot\rangle and induced norm :=,\|\cdot\|:=\sqrt{\langle\cdot,\cdot\rangle}. Let T:T:\mathcal{H}\rightrightarrows\mathcal{H} be a set-valued map. Let Fix(T):={x:xT(x)}{\rm\textbf{Fix}\,}(T):=\{x\in\mathcal{H}:x\in T(x)\} be the set of all fixed points of the operator TT. We are interested in designing dynamical systems models to derive discrete-time schemes for finding approximate solutions to the quasi variational inequality problem which can be formulated as finding x𝒦(x)x^{*}\in\mathcal{K}(x^{*}), such that

T(x),xx0,x𝒦(x)\langle T(x^{*}),\;x-x^{*}\rangle\geq 0\ ,\ \forall\ x\in\mathcal{K}(x^{*}) (1)

where TT is a general vector field and continunous map and 𝒦:\mathcal{K}:\mathcal{H}\rightrightarrows\mathcal{H} is a dynamic constraint set-valued mapping such that 𝒦(x)\mathcal{K}(x)\subseteq\mathcal{H} is nonempty, closed and convex for all xx\in\mathcal{H}. This parametric quasi variational inequality (1) was studied by Bensoussan et al. [6]. If 𝒦(x)𝒦\mathcal{K}(x)\equiv\mathcal{K}, where 𝒦\mathcal{K} is a closed and convex set in \mathcal{H} then the parametric variational inequality (1) is equivalent to the classical variational inequality which was studied and considered by Stampacchia [27] as follows: find xx^{*}\in\mathcal{H} such that

T(x),xx0,x𝒦.\langle T(x^{*}),\;x-x^{*}\rangle\geq 0\ ,\ \forall\ x\in\mathcal{K}. (2)

In many applied situations, the moving convex set 𝒦(x)\mathcal{K}(x) has the form

𝒦(x)=k(x)+K0.\mathcal{K}(x)=k(x)+K_{0}. (3)

where K0K_{0}\subseteq\mathcal{H} is a fixed closed convex set and k(x):k(x):\mathcal{H}\to\mathcal{H} is a continuous function satisfying the Lipschitz property i.e., k(x)k(y)lxy,for some positivel>0.\|k(x)-k(y)\|\leq l\|x-y\|,\ \text{for some positive}\ l>0. Assuming xFix(T)x^{*}\in{\rm\textbf{Fix}\,}(T) is a fixed point of the operator TT converges to the solution of the associated QVI (1), then for a fixed xx^{*}\in\mathcal{H} the QVI is precisely a dynamic-constrained optimization problem

minx𝒦(x)Tx,xx.\min_{x\in\mathcal{K}(x^{*})}\langle Tx^{*},x-x^{*}\rangle. (4)

If 𝒦(x)={x:x,y0,y𝒦(x)}\mathcal{K}^{*}(x)=\{x\in\mathcal{H}:\langle x,y\rangle\geq 0,\ \forall y\in\mathcal{K}(x)\} is a polar (dual set) of a convex-valued cone 𝒦(x)\mathcal{K}(x) in \mathcal{H} then the inequality (2) is equivalent to finding x𝒦x\in\mathcal{K} such that

x𝒦(x),T(x)𝒦(x)andTx,x=0,x\in\mathcal{K}(x),\ T(x)\in\mathcal{K}^{*}(x)\ \ \text{and}\ \ \langle Tx,x\rangle=0,

which is called the quasi complementarity problem [17, 26]. If the operator TT in (2) is smooth, then the following well known result holds and can be viewed as a first order necessary optimality condition for minimizing smooth functions.

Theorem 1.

Let 𝒦\mathcal{K} be a nonempty, convex and closed subset of \mathcal{H}. Let TT be a smooth convex function. Then x𝒦x\in\mathcal{K} is the minimum of the smooth convex T(x)T(x) if and only if, x𝒦x\in\mathcal{K} satisfies

T(x),yx0,y𝒦\langle T^{\prime}(x),y-x\rangle\geq 0,\forall y\in\mathcal{K}

where TT^{\prime} is the Frechet derivative of TT at x𝒦x\in\mathcal{K}.

This theorem shows that the variational inequalities are analogous to the minimization of the convex differentiable functional subject to certain constraint which has led to study a more general framework of variational inequalities applied to nonconstrained and nonsmooth optimization problems. In the following, we state some useful definitions and properties for several kinds of monotone maps followed by well-known facts on projection operators and quasi variational inequalities.

Definition 1.

The nonlinear operator T:T:\mathcal{H}\to\mathcal{H}, is said to be:

  • (i)

    Monotone, if

    T(x)T(y),xy0,x,y.\langle T(x)-T(y),x-y\rangle\geq 0,\ \ \forall x,y\in\mathcal{H}.
  • (ii)

    Strictly monotone if the above inequality is strict for all xyx\not=y in .\mathcal{H}.

  • (iii)

    Strongly monotone if there exists a modulus μ>0\mu>0 such that

    T(x)T(y),xyμxy2,x,y.\langle T(x)-T(y),x-y\rangle\geq\mu\|x-y\|^{2},\ \ \forall x,y\in\mathcal{H}.

Notice that the implication (iii)(i)(iii)\implies(i) holds, whereas the converse need not be true generally, meaning that monotonicity is a weaker property than strongly monotonicity.

Definition 2.

The operator T:T:\mathcal{H}\to\mathcal{H} is called Lipschitz continuous or LL-Lipschitz if there exists some nonnegative L0L\geq 0, such that

TxTyLxy,x,y.\|Tx-Ty\|\leq L\|x-y\|\ ,\quad\forall x,y\in\mathcal{H}.
Lemma 2.

For a given zz\in\mathcal{H}. The necessary and sufficient characterizations of the projection are:

Π𝒦(x)(z)𝒦(x),\Pi_{\mathcal{K}(x)}(z)\in\mathcal{K}(x),

if and only if

Π𝒦(x)(z)z,wΠ𝒦(x)(z)0,w𝒦(x).\langle\Pi_{\mathcal{K}(x)}(z)-z,w-\Pi_{\mathcal{K}(x)}(z)\rangle\geq 0,\ \forall w\in\mathcal{K}(x).

or equivalently, zΠ𝒦(x)(z),wΠ𝒦(x)(z)0,w𝒦(x),\langle z-\Pi_{\mathcal{K}(x)}(z),w-\Pi_{\mathcal{K}(x)}(z)\rangle\leq 0,\ \forall w\in\mathcal{K}(x),

where Π𝒦(x)\Pi_{\mathcal{K}(x)} is called the implicit projection of \mathcal{H} onto the closed convex-valued set 𝒦(x).\mathcal{K}(x)\subset\mathcal{H}. Note that the implicit projection Π𝒦(x)\Pi_{\mathcal{K}(x)} is nonexpansive (i.e., Π𝒦(x)(u)Π𝒦(x)(v)uv\|\Pi_{\mathcal{K}(x)}(u)-\Pi_{\mathcal{K}(x)}(v)\|\leq\|u-v\| ), and satisfy the condition

Π𝒦(x)(u)Π𝒦(y)(u)δxy,x,y,u,\|\Pi_{\mathcal{K}(x)}(u)-\Pi_{\mathcal{K}(y)}(u)\|\leq\delta\|x-y\|,\ \ \forall x,y,u\in\mathcal{H}, (5)

for some constant δ>0.\delta>0. By applying Lemma 2, one can introduce the fixed point formulation of parametric variational inequalities as follows.

Proposition 3 ([15]).

Let Π𝒦(x)\Pi_{\mathcal{K}(x)} be the projection operator onto a closed convex set-valued 𝒦(x)\mathcal{K}(x)\subset\mathcal{H}. Then x𝒦(x)x\in\mathcal{K}(x) is a solution to the quasi variational inequality (1), i.e.,

Tx,yx0,y𝒦(x)\langle Tx,y-x\rangle\leq 0,\ \ \forall y\in\mathcal{K}(x)

if and only if  x=Π𝒦(x)(xλT(x)),x=\Pi_{\mathcal{K}(x)}(x-\lambda T(x)),  for some λ>0.\lambda>0.

A particular case if 𝒦(x)𝒦\mathcal{K}(x)\equiv\mathcal{K} then the implicit projection is exactly the Euclidean projection Π𝒦(x)=Π𝒦\Pi_{\mathcal{K}(x)}=\Pi_{\mathcal{K}} which is defined as

Π𝒦(u):=argminx{12xu22}.\Pi_{\mathcal{K}}(u):={\rm arg}\!\min_{x}\Big{\{}\displaystyle\frac{1}{2}\|x-u\|^{2}_{2}\Big{\}}.

Furthermore, for all x,yx,y\in\mathcal{H} and all z𝒦z\in\mathcal{K} the Π𝒦\Pi_{\mathcal{K}} is firmly nonexpansive:

Π𝒦(x)Π𝒦(y)2xy2(xΠ𝒦(x))(yΠ𝒦(y))2.\|\Pi_{\mathcal{K}}(x)-\Pi_{\mathcal{K}}(y)\|^{2}\leq\|x-y\|^{2}-\|(x-\Pi_{\mathcal{K}}(x))-(y-\Pi_{\mathcal{K}}(y))\|^{2}.

Since 𝒦(x)\mathcal{K}(x) has the form (3), we can write the implicit projection onto the dynamic convex set as

Π𝒦(x)(u)=Πk(x)+K0(u)=k(x)+ΠK0(uk(x)),u.\Pi_{\mathcal{K}(x)}(u)=\Pi_{k(x)+K_{0}}(u)=k(x)+\Pi_{K_{0}}(u-k(x)),\ \ \forall u\in\mathcal{H}. (6)

Suppose that the first term in (3) is Lipschitz continuous with l>0l>0, using (5), the Cauchy-Schwarz inequality and the nonexpansiveness of the Euclidean projection ΠK\Pi_{K}, we then have

Π𝒦(x)(u)Π𝒦(y)(u)=k(x)k(y)+ΠK0(uk(x))ΠK0(uk(y))k(x)k(y)+ΠK0(uk(x))ΠK0(uk(y)) 2k(x)k(y)2lxy,x,y.\begin{split}\|\Pi_{\mathcal{K}(x)}(u)-\Pi_{\mathcal{K}(y)}(u)\|=&\ \|k(x)-k(y)+\Pi_{K_{0}}(u-k(x))-\Pi_{K_{0}}(u-k(y))\|\\ \leq&\ \|k(x)-k(y)\|+\|\Pi_{K_{0}}(u-k(x))-\Pi_{K_{0}}(u-k(y))\|\\ \leq&\ 2\|k(x)-k(y)\|\leq 2l\|x-y\|,\ \ \ \forall x,y\in\mathcal{H}.\end{split} (7)

From Proposition (3), it follows that x𝒦(x)x\in\mathcal{K}(x) such that

x=Π𝒦(x)(xλTx)=Πk(x)+K0(xλTx)=k(x)+ΠK0(xλTxk(x)),u.x=\Pi_{\mathcal{K}(x)}(x-\lambda Tx)=\Pi_{k(x)+K_{0}}(x-\lambda Tx)=k(x)+\Pi_{K_{0}}(x-\lambda Tx-k(x)),\ \ \forall u\in\mathcal{H}.

This implies that

xk(x)=ΠK0(xλTxk(x)).x-k(x)=\Pi_{K_{0}}(x-\lambda Tx-k(x)).

By Lemma 2, this is equivalent to

Tx,g(y)g(x)0,y,\langle Tx,g(y)-g(x)\rangle\geq 0,\ \ \forall y\in\mathcal{H}, (8)

where g(y)=yg(y)=y and g(x)=xk(x).g(x)=x-k(x). Inequality (8) is called the general variational inequality, which was introduced by Noor [21], and is actually equivalent to the QVI (1).

Definition 3.

The dynamical system is said to be globally exponentially stable if any trajectory x(t)x(t) satisfies

x(t)xρx(t0)xexp(η(tt0)),tt0\|x(t)-x^{*}\|\leq\rho\|x(t_{0})-x^{*}\|exp(-\eta(t-t_{0})),\ \ \ \forall t\geq t_{0}

where ρ,η>0\rho,\eta>0 are constants and do not depend on the initial point.

If the dynamical system is stable at the equilibrium point xx^{*} in the Lyapunov sense then the dynamical system is globally asymptotically stable at that point. It is noted that globally exponentially stable means the system must be globally stable and converge fast.

For deriving the convergence of our methods, the following theorem is needed:

Theorem 4 ([28]).

Let the operator T:T:\mathcal{H}\to\mathcal{H} be a μ\mu-strongly monotone and LL-Lipschitz continuous with μ,L>0.\mu,L>0. Then

T(x)T(y)2+μLxy2(L+μ)T(x)T(y),xy,x,y\|T(x)-T(y)\|^{2}+\mu L\|x-y\|^{2}\leq(L+\mu)\langle T(x)-T(y),x-y\rangle\ ,\forall\ x,y\in\mathcal{H}

holds.

3 Main Results

In this section, we introduce, derive and analyze continuous and discrete methods based on a third order dynamical system in the continuous case, and on the central finite difference and forward/backward difference methods in the discrete case besides their rates of convergences for quasi variational inequalities. Using the fixed point formulation, a new projected dynamical system of the third order associated with quasi variational inequalities (1) is proposed and some attached suitable discretizations forms are investigated. These continuous-time dynamical systems and their discrete-time counterparts suggest some inertial-type implicit and explicit proximal methods for solving quasi variational inequalities.

3.1 Continuous gradient method

The suggested projected dynamical system designed in a continuous-time form. Consider the problem of finding a trajectory x(t)x(t)\in\mathcal{H} such that

{α(t)x˙˙˙(t)+β(t)x¨(t)+γ(t)x˙(t)+x(t)=Π𝒦(x(t))(x(t)λ(t)T(x(t))),x(t0)=x0,x˙(t0)=x1,x¨(t0)=x2,\left\{\begin{array}[]{lll}\alpha(t)\dddot{x}(t)+\beta(t)\ddot{x}(t)+\gamma(t)\dot{x}(t)+x(t)=\Pi_{\mathcal{K}(x(t))}(x(t)-\lambda(t)T(x(t))),\\ x(t_{0})=x_{0},\\ \dot{x}(t_{0})=x_{1},\\ \ddot{x}(t_{0})=x_{2},\end{array}\right. (9)

where x(t)x(t) is the state variable and the initial points x0,x1,x2x_{0},x_{1},x_{2}\in\mathcal{H}. The differential system (9) recovers several existing dynamics-type approaches and projection-based algorithms for solving several classes of variational inequalities. This model (9) is quite similar to the dynamical system used by Hazaimah [14] for mixed variational inequalities, the difference between the two approaches is by taking the coeffecients of the first three terms on the left hand side of (9) to be time scaling. Particular cases of the general system (9) are discussed next. If α(t)0,β(t)β,γ(t)1\alpha(t)\equiv 0,\beta(t)\equiv\beta,\gamma(t)\equiv 1, then (9) is equivalent to the continuous second-order dynamical system introduced by Antipin et al. [3], while the same technique was examined in the discrete case in [22] with implicit iterative methods. Extra gradient method for solving quasi variational inequalities is introduced in [2]. Some second-order iterative versions are studied in [4]. If α(t)0β(t)γ(t)\alpha(t)\equiv 0\equiv\beta(t)\equiv\gamma(t) and 𝒦(x)=𝒦\mathcal{K}(x)=\mathcal{K}, then the system (9) is reduced to the classical gradient projection for smooth constrained optimization problems and projection-like methods for solving variational inequalities.

The projected dynamical system (9) can be rewritten, utilizing the mapping form (6), as

α(t)x˙˙˙(t)+β(t)x¨(t)+x˙(t)+x(t)=k(x(t))+ΠK0(x(t)λ(t)T(x(t))k(x(t))).\alpha(t)\dddot{x}(t)+\beta(t)\ddot{x}(t)+\dot{x}(t)+x(t)=k(x(t))+\Pi_{K_{0}}\Big{(}x(t)-\lambda(t)T(x(t))-k(x(t))\Big{)}. (10)

The following theorem discusses the exponentially stability and shows the convergence rate for the proposed continuous method.

Theorem 5.

Assume the following:

  1. (i)

    𝒦(x):\mathcal{K}(x):\mathcal{H}\rightrightarrows\mathcal{H} is a convex set-valued mapping satisfies (3) where k(x):k(x):\mathcal{H}\to\mathcal{H} is ll-Lipschitz mapping.

  2. (ii)

    TT is μ\mu-strongly monotone and Lipschitzian with constant L>0L>0.

  3. (iii)

    α(1l)0\alpha(1-l)\geq 0, (1l)(α+β)3α(1-l)(\alpha+\beta)\geq 3\alpha.

  4. (iv)

    l<min{1, 12β, 2λ(t)[μlL]μ2λ2(t)}l<\min\{1,\ 1-2\beta,\ 2\lambda(t)\big{[}\mu-lL\big{]}-\mu^{2}\lambda^{2}(t)\}.

  5. (v)

    λ(t)C1[0,),α(t)α>0,β(t)β>0,γ(t)1.\lambda(t)\in C^{1}[0,\infty),\ \alpha(t)\equiv\alpha>0,\ \beta(t)\equiv\beta>0,\ \gamma(t)\equiv 1.

Then

x(t)x2[eβαt]1x(0)x2+α1K2e2βαt,t0\begin{split}\|x(t)-x^{*}\|^{2}\leq[e^{\frac{\beta}{\alpha}t}]^{-1}\|x(0)-x^{*}\|^{2}+\alpha^{-1}K_{2}\ e^{-\frac{2\beta}{\alpha}t},\ \ \forall t\geq 0\end{split} (11)
Proof.

Using Proposition 3, the projected dynamical system given by (10) is equivalent to the variational inequality setting

x(t)λ(t)T(x(t))k(x(t))αx˙˙˙(t)βx¨(t)x˙(t)x(t)+k(x(t)),yαx˙˙˙(t)βx¨(t)x˙(t)x(t)+k(x(t))0,\begin{split}\langle x(t)-\lambda(t)T(x(t))&-k(x(t))-\alpha\dddot{x}(t)-\beta\ddot{x}(t)-\dot{x}(t)-x(t)+k(x(t)),\\ &y-\alpha\dddot{x}(t)-\beta\ddot{x}(t)-\dot{x}(t)-x(t)+k(x(t))\rangle\leq 0,\end{split}

for all yK0y\in K_{0}. Rearranging the above inequality using simple algebraic manipulations

λ(t)T(x(t))+αx˙˙˙(t)+βx¨(t)+x˙(t),yαx˙˙˙(t)βx¨(t)x˙(t)x(t)+k(x(t))0,\begin{split}\langle\lambda(t)T(x(t))&+\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t),y-\alpha\dddot{x}(t)-\beta\ddot{x}(t)-\dot{x}(t)-x(t)+k(x(t))\rangle\geq 0,\end{split} (12)

for all yK0y\in K_{0}. Set y=xk(x)K0y=x^{*}-k(x^{*})\in K_{0} in (12) and x=α(t)x˙˙˙(t)+β(t)x¨(t)+x˙(t)+x(t)+k(x)k(x(t))𝒦(x)x=\alpha(t)\dddot{x}(t)+\beta(t)\ddot{x}(t)+\dot{x}(t)+x(t)+k(x^{*})-k(x(t))\in\mathcal{K}(x^{*}) in (1). Multiply (1) by λ(t)>0\lambda(t)>0 and combining the resulting inequalities, we obtain respectively

λ(t)T(x(t))+αx˙˙˙(t)+βx¨(t)+x˙(t),αx˙˙˙(t)βx¨(t)x˙(t)x(t)+x+k(x(t))k(x)0.\begin{split}\langle\lambda(t)T(x(t))&+\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t),-\alpha\dddot{x}(t)-\beta\ddot{x}(t)-\dot{x}(t)-x(t)+x^{*}+k(x(t))-k(x^{*})\rangle\geq 0.\end{split}

and

λ(t)T(x),αx˙˙˙(t)+βx¨(t)+x˙(t)+x(t)x+k(x)k(x(t))0.\begin{split}\lambda(t)\langle T(x^{*}),\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t)+x(t)-x^{*}+k(x^{*})-k(x(t))\rangle\geq 0.\end{split}

Adding together, we have

λ(t)T(x(t)),αx˙˙˙(t)βx¨(t)x˙(t)x(t)+x+k(x(t))k(x)αx˙˙˙(t)+βx¨(t)+x˙(t),αx˙˙˙(t)+βx¨(t)+x˙(t)+x(t)x+k(x)k(x(t))+λ(t)T(x),αx˙˙˙(t)+βx¨(t)+x˙(t)+x(t)x+k(x)k(x(t))0.\begin{split}&\langle\lambda(t)T(x(t)),-\alpha\dddot{x}(t)-\beta\ddot{x}(t)-\dot{x}(t)-x(t)+x^{*}+k(x(t))-k(x^{*})\rangle\\ &-\langle\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t),\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t)+x(t)-x^{*}+k(x^{*})-k(x(t))\rangle\\ &+\lambda(t)\langle T(x^{*}),\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t)+x(t)-x^{*}+k(x^{*})-k(x(t))\rangle\geq 0.\end{split}

Thus

αx˙˙˙(t)+βx¨(t)+x˙(t),αx˙˙˙(t)+βx¨(t)+x˙(t)+x(t)x+k(x)k(x(t))λ(t)T(x(t))T(x),αx˙˙˙(t)βx¨(t)x˙(t)x(t)+x+k(x(t))k(x)\begin{split}\langle\alpha&\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t),\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t)+x(t)-x^{*}+k(x^{*})-k(x(t))\rangle\\ &\leq\lambda(t)\langle T(x(t))-T(x^{*}),-\alpha\dddot{x}(t)-\beta\ddot{x}(t)-\dot{x}(t)-x(t)+x^{*}+k(x(t))-k(x^{*})\rangle\end{split} (13)

Next we rewrite (13), by applying the inequality aba22+b22ab\leq\displaystyle\frac{a^{2}}{2}+\displaystyle\frac{b^{2}}{2} , as

αx˙˙˙(t)+βx¨(t)+x˙(t)2+αx˙˙˙(t)+βx¨(t)+x˙(t),x(t)x+αx˙˙˙(t)+βx¨(t)+x˙(t),k(x(t))k(x)λ2(t)2T(x(t))T(x)2+12αx˙˙˙(t)βx¨(t)x˙(t)2+λ(t)T(x(t))T(x),xx(t)+λ(t)T(x(t))T(x),k(x(t))k(x)\begin{split}\|\alpha\dddot{x}(t)&+\beta\ddot{x}(t)+\dot{x}(t)\|^{2}+\langle\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t),x(t)-x^{*}\rangle+\langle\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t),\\ &k(x(t))-k(x^{*})\rangle\leq\frac{\lambda^{2}(t)}{2}\|T(x(t))-T(x^{*})\|^{2}+\frac{1}{2}\|\alpha\dddot{x}(t)-\beta\ddot{x}(t)-\dot{x}(t)\|^{2}\\ &+\lambda(t)\langle T(x(t))-T(x^{*}),x^{*}-x(t)\rangle+\lambda(t)\langle T(x(t))-T(x^{*}),k(x(t))-k(x^{*})\rangle\end{split} (14)

From the assumption TT is LL-Lipschitz and μ\mu-strongly monotone. Applying Theorem 4 and since k(x)k(x) is ll-Lipschitz, the inequality (14) implies that

αx˙˙˙(t)+βx¨(t)+x˙(t)2+2αx˙˙˙(t)+βx¨(t)+x˙(t),x(t)xlαx˙˙˙(t)+βx¨(t)+x˙(t)2lx(t)x2+λ(t)[2(L+μ)λ(t)]T(x(t))T(x),x(t)x+λ2(t)Lμx(t)x22λ(t)lLx(t)x20.\begin{split}\|\alpha\dddot{x}(t)&+\beta\ddot{x}(t)+\dot{x}(t)\|^{2}+2\langle\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t),x(t)-x^{*}\rangle-l\|\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t)\|^{2}\\ &-l\|x(t)-x^{*}\|^{2}+\lambda(t)\big{[}2-(L+\mu)\lambda(t)\big{]}\langle T(x(t))-T(x^{*}),x(t)-x^{*}\rangle\\ &+\lambda^{2}(t)L\mu\|x(t)-x^{*}\|^{2}-2\lambda(t)lL\|x(t)-x^{*}\|^{2}\leq 0.\end{split}

Since TT is μ\mu-strongly monotone then by Definition 1 (iii), the latter inequality reduced to

(1l)αx˙˙˙(t)+βx¨(t)+x˙(t)2+2αx˙˙˙(t)+βx¨(t)+x˙(t),x(t)x+μλ(t)[2(L+μ)λ(t)]x(t)x2+(λ2(t)Lμ2λ(t)lLl)x(t)x20.\begin{split}(1&-l)\|\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t)\|^{2}+2\langle\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t),x(t)-x^{*}\rangle\\ &+\mu\lambda(t)\big{[}2-(L+\mu)\lambda(t)\big{]}\|x(t)-x^{*}\|^{2}+\Big{(}\lambda^{2}(t)L\mu-2\lambda(t)lL-l\Big{)}\|x(t)-x^{*}\|^{2}\leq 0.\end{split}

Algebraically rearranging terms, we obtain

(1l)αx˙˙˙(t)+βx¨(t)+x˙(t)2+2αx˙˙˙(t)+βx¨(t)+x˙(t),x(t)x+A(t)x(t)x20,\begin{split}(1&-l)\|\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t)\|^{2}+2\langle\alpha\dddot{x}(t)+\beta\ddot{x}(t)+\dot{x}(t),x(t)-x^{*}\rangle+A(t)\|x(t)-x^{*}\|^{2}\leq 0,\end{split} (15)

where A(t)=2λ(t)[μlL]μ2λ2(t)l.A(t)=2\lambda(t)\big{[}\mu-lL\big{]}-\mu^{2}\lambda^{2}(t)-l.

Next, we employ the following helpful relations to simplify the previous inequality:

x˙,x˙\displaystyle\langle\dot{x},\dot{x}\rangle =x˙2\displaystyle=\|\dot{x}\|^{2} (16a)
x˙,x(t)x\displaystyle\langle\dot{x},x(t)-x^{*}\rangle =12ddtx(t)x2\displaystyle=\frac{1}{2}\frac{d}{dt}\|x(t)-x^{*}\|^{2} (16b)
x¨,x(t)x\displaystyle\langle\ddot{x},x(t)-x^{*}\rangle =12d2dt2x(t)x2x˙2\displaystyle=\frac{1}{2}\frac{d^{2}}{dt^{2}}\|x(t)-x^{*}\|^{2}-\|\dot{x}\|^{2} (16c)
x˙˙˙,x(t)x\displaystyle\langle\dddot{x},x(t)-x^{*}\rangle =12d3dt3x(t)x23x¨,x˙\displaystyle=\frac{1}{2}\frac{d^{3}}{dt^{3}}\|x(t)-x^{*}\|^{2}-3\langle\ddot{x},\dot{x}\rangle (16d)
2x˙˙˙,x˙\displaystyle 2\langle\dddot{x},\dot{x}\rangle =ddtx¨2+ddtx˙22x¨2ddtx¨x˙2\displaystyle=\frac{d}{dt}\|\ddot{x}\|^{2}+\frac{d}{dt}\|\dot{x}\|^{2}-2\|\ddot{x}\|^{2}-\frac{d}{dt}\|\ddot{x}-\dot{x}\|^{2} (16e)

Hence, inequality (15) can be rewritten as

(1l)[α2x˙˙˙(t)2+β2x¨(t)2+x˙(t)2+2αβx˙˙˙,x¨+2αx˙˙˙,x˙+2βx¨,x˙]+αd3dt3x(t)x26αx¨,x˙+βd2dt2x(t)x22βx˙2+ddtx(t)x2+A(t)x(t)x20,\begin{split}(1&-l)\bigg{[}\alpha^{2}\|\dddot{x}(t)\|^{2}+\beta^{2}\|\ddot{x}(t)\|^{2}+\|\dot{x}(t)\|^{2}+2\alpha\beta\langle\dddot{x},\ddot{x}\rangle+2\alpha\langle\dddot{x},\dot{x}\rangle+2\beta\langle\ddot{x},\dot{x}\rangle\bigg{]}\\ &+\alpha\frac{d^{3}}{dt^{3}}\|x(t)-x^{*}\|^{2}-6\alpha\langle\ddot{x},\dot{x}\rangle+\beta\frac{d^{2}}{dt^{2}}\|x(t)-x^{*}\|^{2}-2\beta\|\dot{x}\|^{2}+\frac{d}{dt}\|x(t)-x^{*}\|^{2}\\ &+A(t)\|x(t)-x^{*}\|^{2}\leq 0,\end{split} (17)

Using Rewriting (17), we have

αd3dt3x(t)x2+βd2dt2x(t)x2+ddtx(t)x2+A(t)x(t)x2α(1l)ddtx¨x˙2+α(1l)(1+β)ddtx¨2+((1l)(α+β)3α)ddtx˙2+(1l)α2x˙˙˙2+(1l)β2x¨2+(1l2β)x˙20,\begin{split}\alpha&\frac{d^{3}}{dt^{3}}\|x(t)-x^{*}\|^{2}+\beta\frac{d^{2}}{dt^{2}}\|x(t)-x^{*}\|^{2}+\frac{d}{dt}\|x(t)-x^{*}\|^{2}+A(t)\|x(t)-x^{*}\|^{2}\\ &-\alpha(1-l)\frac{d}{dt}\|\ddot{x}-\dot{x}\|^{2}+\alpha(1-l)(1+\beta)\frac{d}{dt}\|\ddot{x}\|^{2}+\bigg{(}(1-l)(\alpha+\beta)-3\alpha\bigg{)}\frac{d}{dt}\|\dot{x}\|^{2}\\ &+(1-l)\alpha^{2}\|\dddot{x}\|^{2}+(1-l)\beta^{2}\|\ddot{x}\|^{2}+(1-l-2\beta)\|\dot{x}\|^{2}\leq 0,\end{split} (18)

Multiply (18) by H(t)=exp0t𝑑sH(t)=\exp{\int_{0}^{t}ds}, and integrating over the interval [0,t][0,t], we obtain

αd2dt2x(t)x2+βddtx(t)x2+x(t)x2+0tA(s)H(s)x(s)x2𝑑sα(1l)x¨(t)x˙(t)2+α(1l)(1+β)x¨(t)2+((1l)(α+β)3α)x˙(t)2+(1l)α20tH(s)x˙˙˙(s)2𝑑s+(1l)β20tH(s)x¨(s)2𝑑s+(1l2β)0tH(s)x˙(s)2𝑑sK0,\begin{split}\alpha&\frac{d^{2}}{dt^{2}}\|x(t)-x^{*}\|^{2}+\beta\frac{d}{dt}\|x(t)-x^{*}\|^{2}+\|x(t)-x^{*}\|^{2}+\int_{0}^{t}A(s)H(s)\|x(s)-x^{*}\|^{2}ds\\ &-\alpha(1-l)\|\ddot{x}(t)-\dot{x}(t)\|^{2}+\alpha(1-l)(1+\beta)\|\ddot{x}(t)\|^{2}+\bigg{(}(1-l)(\alpha+\beta)-3\alpha\bigg{)}\|\dot{x}(t)\|^{2}\\ &+(1-l)\alpha^{2}\int_{0}^{t}H(s)\|\dddot{x}(s)\|^{2}ds+(1-l)\beta^{2}\int_{0}^{t}H(s)\|\ddot{x}(s)\|^{2}ds\\ &+(1-l-2\beta)\int_{0}^{t}H(s)\|\dot{x}(s)\|^{2}ds\leq K_{0},\end{split} (19)

where

K0=x(0)x2α(1l)x¨(0)x˙(0)2+α(1l)(1+β)x¨(0)2+((1l)(α+β)3α)x˙(0)2.K_{0}=\|x(0)-x^{*}\|^{2}-\alpha(1-l)\|\ddot{x}(0)-\dot{x}(0)\|^{2}+\alpha(1-l)(1+\beta)\|\ddot{x}(0)\|^{2}+\Big{(}(1-l)(\alpha+\beta)-3\alpha\Big{)}\|\dot{x}(0)\|^{2}.

All definite integrals on the left hand side of (19) are non-negative, which means that

αd2dt2x(t)x2+βddtx(t)x2+x(t)x2α(1l)x¨(t)x˙(t)2+α(1l)(1+β)x¨(t)2+((1l)(α+β)3α)x˙(t)2K0,t0\begin{split}\alpha&\frac{d^{2}}{dt^{2}}\|x(t)-x^{*}\|^{2}+\beta\frac{d}{dt}\|x(t)-x^{*}\|^{2}+\|x(t)-x^{*}\|^{2}-\alpha(1-l)\|\ddot{x}(t)-\dot{x}(t)\|^{2}\\ &+\alpha(1-l)(1+\beta)\|\ddot{x}(t)\|^{2}+\bigg{(}(1-l)(\alpha+\beta)-3\alpha\bigg{)}\|\dot{x}(t)\|^{2}\leq K_{0},\ \ \forall t\geq 0\end{split} (20)

Similarly, we integrate the inequality (20) over the interval [0,t][0,t], then (20) simplified to

αddtx(t)x2+βx(t)x2+0tx(s)x2𝑑sα(1l)0tx¨(s)x˙(s)2𝑑s+α(1l)(1+β)0tx¨(s)2𝑑s+((1l)(α+β)3α)0tx˙(s)2𝑑sK0+K1,t0\begin{split}\alpha&\frac{d}{dt}\|x(t)-x^{*}\|^{2}+\beta\|x(t)-x^{*}\|^{2}+\int_{0}^{t}\|x(s)-x^{*}\|^{2}ds-\alpha(1-l)\int_{0}^{t}\|\ddot{x}(s)-\dot{x}(s)\|^{2}ds\\ &+\alpha(1-l)(1+\beta)\int_{0}^{t}\|\ddot{x}(s)\|^{2}ds+\bigg{(}(1-l)(\alpha+\beta)-3\alpha\bigg{)}\int_{0}^{t}\|\dot{x}(s)\|^{2}ds\leq K_{0}+K_{1},\ \ \forall t\geq 0\end{split} (21)

where K1=βx(0)x2K_{1}=\beta\|x(0)-x^{*}\|^{2}. In the same manner, we observe that all integrals in the left hand side are non-negative, and therefore (21) becomes

αddtx(t)x2+βx(t)x2K0+K1=K2,t0\begin{split}\alpha\frac{d}{dt}\|x(t)-x^{*}\|^{2}+\beta\|x(t)-x^{*}\|^{2}\leq K_{0}+K_{1}=K_{2},\ \ \forall t\geq 0\end{split} (22)

Multiply (22) by α1\alpha^{-1}, we obtain the linear differential inequality

ddtx(t)x2+βαx(t)x2α1K2,t0\begin{split}\frac{d}{dt}\|x(t)-x^{*}\|^{2}+\frac{\beta}{\alpha}\|x(t)-x^{*}\|^{2}\leq\alpha^{-1}K_{2},\ \ \forall t\geq 0\end{split} (23)

Multiply (23) by e0tβα𝑑se^{\int_{0}^{t}\frac{\beta}{\alpha}ds}, so it can be rewritten, using the idea of the integrating factor, in the form

ddt[x(t)x2e0tβα𝑑s]α1K2eβαt,t0\begin{split}\frac{d}{dt}\Big{[}\|x(t)-x^{*}\|^{2}e^{\int_{0}^{t}\frac{\beta}{\alpha}ds}\Big{]}\leq\alpha^{-1}K_{2}\ e^{-\frac{\beta}{\alpha}t},\ \ \forall t\geq 0\end{split} (24)

Integrating (24) over the interval [0,t][0,t], we have

x(t)x2eβαtx(0)x2+α1K2eβαt,t0\begin{split}\|x(t)-x^{*}\|^{2}e^{\frac{\beta}{\alpha}t}\leq\|x(0)-x^{*}\|^{2}+\alpha^{-1}K_{2}\ e^{-\frac{\beta}{\alpha}t},\ \ \forall t\geq 0\end{split} (25)

Multiplying by [eβαt]1[e^{\frac{\beta}{\alpha}t}]^{-1}, it follows that

x(t)x2[eβαt]1x(0)x2+α1K2e2βαt,t0\begin{split}\|x(t)-x^{*}\|^{2}\leq[e^{\frac{\beta}{\alpha}t}]^{-1}\|x(0)-x^{*}\|^{2}+\alpha^{-1}K_{2}\ e^{-\frac{2\beta}{\alpha}t},\ \ \forall t\geq 0\end{split} (26)

Thus, we have proved the result with an estimation rate of convergence of the continuous method (9) given by inequality (26). ∎

Remark 1.

If we let the coefficient parameters of the inertial and damping terms x¨,x˙\ddot{x},\dot{x} respectively, change over time, namely α(t)α,β(t)β\alpha(t)\not\equiv\alpha,\beta(t)\not\equiv\beta then the convergence of the continuous method (9) can be improved under some mild general conditions such that

α(t),β(t)C2[0,),t0,limtα(t)>0,limtβ(t)>0.\alpha(t),\beta(t)\in C^{2}[0,\infty),\ \ t\geq 0,\ \ \lim_{t\to\infty}\alpha(t)>0,\ \lim_{t\to\infty}\beta(t)>0.

3.2 Discrete methods

The idea of iterative methods mostly is based on discretizing the space derivatives by using certain discretizations methods and the goal behind boosting iterative schemes is to accelerate the rate of convergence. In this paper we use the central finite difference, backward difference and forward difference schemes to propose explicit and implicit forms which enable us to obtain the discretized counterpart of (9) as a projected equation. Thus the dynamical system (9) may be discretized as:

αxn+22xn+1+2xn1xn22h3+βxn+12xn+xn1h2+γxnxn1h+xn+2=Π𝒦(xn)(xnλT(xn+2))\begin{split}\alpha\displaystyle\frac{x_{n+2}-2x_{n+1}+2x_{n-1}-x_{n-2}}{2h^{3}}&+\beta\frac{x_{n+1}-2x_{n}+x_{n-1}}{h^{2}}+\gamma\frac{x_{n}-x_{n-1}}{h}+x_{n+2}\\ &=\Pi_{\mathcal{K}(x_{n})}(x_{n}-\lambda T(x_{n+2}))\end{split} (27)

where hh is the step size for the iterative process. This discrete scheme (27) suggests a new implicit iterative method for solving quasi variational inequalities (1) by the third order central difference formula.

Algorithm 3.1.

For any x0,x1,x2x_{0},x_{1},x_{2}\in\mathcal{H}, and for any nonnegative integer n+n\in\mathbb{Z}_{+}, compute the update rule xn+2x_{n+2} by the iterative process

xn+2=Π𝒦(xn)[xnλT(xn+2)αxn+22(αβh)xn+12(2βhγh2)xn+2(α+βhγh2)xn1αxn22h3]\begin{split}x_{n+2}&=\Pi_{\mathcal{K}(x_{n})}\bigg{[}x_{n}-\lambda T(x_{n+2})\\ &-\frac{\alpha x_{n+2}-2(\alpha-\beta h)x_{n+1}-2(2\beta h-\gamma h^{2})x_{n}+2(\alpha+\beta h-\gamma h^{2})x_{n-1}-\alpha x_{n-2}}{2h^{3}}\bigg{]}\end{split} (28)

This algorithm is inertial proximal-type method for solving (1). Using Lemma 2, Algorithm 3.1 can be rewritten in the variational equivalent formulation:

Algorithm 3.2.

For any x0,x1,x2x_{0},x_{1},x_{2}\in\mathcal{H}, and for any nonnegative integer n+n\in\mathbb{Z}_{+}, compute xn+2x_{n+2} by the iterative process

λT(xn+2)+αxn+22(αβh)xn+12(2βhγh2)xn+2(α+βhγh2)xn1αxn22h3,yxn+20,y𝒦(x)\begin{split}\big{\langle}\lambda T(x_{n+2})&+\frac{\alpha x_{n+2}-2(\alpha-\beta h)x_{n+1}-2(2\beta h-\gamma h^{2})x_{n}+2(\alpha+\beta h-\gamma h^{2})x_{n-1}-\alpha x_{n-2}}{2h^{3}},\\ &y-x_{n+2}\big{\rangle}\geq 0,\ \ \ \forall y\in\mathcal{K}(x)\end{split} (29)

Using different discretization and taking α=1=β=γ\alpha=1=\beta=\gamma, Algorithm 3.1 can be reduced to the following iterative:

xn+22xn+1+2xn1xn22h3+xn+12xn+xn1h2+xnxn1h+xn+2=Π𝒦(xn)(xnλT(xn))\begin{split}\frac{x_{n+2}-2x_{n+1}+2x_{n-1}-x_{n-2}}{2h^{3}}&+\frac{x_{n+1}-2x_{n}+x_{n-1}}{h^{2}}+\frac{x_{n}-x_{n-1}}{h}+x_{n+2}\\ &=\Pi_{\mathcal{K}(x_{n})}(x_{n}-\lambda T(x_{n}))\end{split}

which yields to the following recurrence formula

xn+2=h^1+h^Π𝒦(xn)[(11h+2h2)xnλT(xn)(2h2)xn+1+(2+2h2h2)xn1xn22h3]x_{n+2}=\frac{\hat{h}}{1+\hat{h}}\Pi_{\mathcal{K}(x_{n})}\bigg{[}(1-\frac{1}{h}+\frac{2}{h^{2}})x_{n}-\lambda T(x_{n})-\frac{(2h-2)x_{n+1}+(2+2h-2h^{2})x_{n-1}-x_{n-2}}{2h^{3}}\bigg{]} (30)

where h^=2h3\hat{h}=2h^{3}. This is called an inertial explicit proximal method for solving quasi variational inequalities (1). Taking α=1=β=γ\alpha=1=\beta=\gamma, and h=1h=1 then Algorithm 3.1 can be reduced to the explicit iterative formula:

3xn+22xn+xn2=2Π𝒦(xn)(xnλT(xn)).3x_{n+2}-2x_{n}+x_{n-2}=2\ \Pi_{\mathcal{K}(x_{n})}(x_{n}-\lambda T(x_{n})). (31)

Following the same fashion with slightly exploring the forward/backward iterates we can suggest several explicit and implicit recursive methods for finding approximate solutions of parametric quasi variational inequalities (1). Hence, by using the central finite difference and this time with forward difference scheme rather than backward scheme as in (27), which allows us to propose a new iterative approach

αxn+22xn+1+2xn1xn22h3+βxn+12xn+xn1h2+γxn+1xnh+xn+2=Π𝒦(xn)(xnλT(xn+1))\begin{split}\alpha\displaystyle\frac{x_{n+2}-2x_{n+1}+2x_{n-1}-x_{n-2}}{2h^{3}}&+\beta\frac{x_{n+1}-2x_{n}+x_{n-1}}{h^{2}}+\gamma\frac{x_{n+1}-x_{n}}{h}+x_{n+2}\\ &=\Pi_{\mathcal{K}(x_{n})}(x_{n}-\lambda T(x_{n+1}))\end{split} (32)

which can be, equivalently, derived as the following inertial implicit proximal method:

Algorithm 3.3.

For x0,x1,x2x_{0},x_{1},x_{2}\in\mathcal{H}, and for any nonnegative integer n+n\in\mathbb{Z}_{+}, compute the update step xn+2x_{n+2} by

xn+2=Π𝒦(xn)[xnλT(xn+1)αxn+22(αβhγh2)xn+12(2βh+γh2)xn+2(α+βh)xn1αxn22h3].\begin{split}x_{n+2}&=\Pi_{\mathcal{K}(x_{n})}\bigg{[}x_{n}-\lambda T(x_{n+1})\\ &-\frac{\alpha x_{n+2}-2(\alpha-\beta h-\gamma h^{2})x_{n+1}-2(2\beta h+\gamma h^{2})x_{n}+2(\alpha+\beta h)x_{n-1}-\alpha x_{n-2}}{2h^{3}}\bigg{]}.\end{split} (33)

For α=1=β=γ\alpha=1=\beta=\gamma and h=1h=1 then Algorithm 3.3 can be reduced to the explicit version:

3xn+2+2xn+16xn+4xn1xn2=2Π𝒦(xn)(xnλT(xn+1)).3x_{n+2}+2x_{n+1}-6x_{n}+4x_{n-1}-x_{n-2}=2\ \Pi_{\mathcal{K}(x_{n})}(x_{n}-\lambda T(x_{n+1})).

On a different perspective, it is known that adding an inertial term into discrete-time algorithms will speed up and cause a significant change in the convergence rate using extrapolating factor Θn(xnxn1)\Theta_{n}(x_{n}-x_{n-1}) for several classes of smooth and strongly monotone mappings. In the light of this concept, we introduce the following two-step inertial iterative algorithms:

Algorithm 3.4.

Given x0,x1,x2x_{0},x_{1},x_{2}\in\mathcal{H} and n+n\in\mathbb{Z}_{+}, compute xn+2x_{n+2} by the iterative steps:

zn=xn+Θn(xnxn1)xn+2=23Π𝒦(xn)(3xnλTznxn1)\begin{split}z_{n}=&\ x_{n}+\Theta_{n}(x_{n}-x_{n-1})\\ x_{n+2}=&\ \frac{2}{3}\ \Pi_{\mathcal{K}(x_{n})}(3x_{n}-\lambda Tz_{n}-x_{n-1})\end{split}

where 0Θn10\leq\Theta_{n}\leq 1.

Similarly, given x0,x1,x2x_{0},x_{1},x_{2}\in\mathcal{H} and 0Θn10\leq\Theta_{n}\leq 1 for n+n\in\mathbb{Z}_{+}, in some cases letting the constraint set depend on the inertial equation would give new algorithms. Thus, we can compute xn+2x_{n+2} by two new inertial iterative methods for quasi variational inequalities, respectively:

zn=xn+Θn(xnxn1)xn+2=23Π𝒦(xn)(3znλTznxn1),\begin{split}z_{n}=&\ x_{n}+\Theta_{n}(x_{n}-x_{n-1})\\ x_{n+2}=&\ \frac{2}{3}\ \Pi_{\mathcal{K}(x_{n})}(3z_{n}-\lambda Tz_{n}-x_{n-1}),\end{split}

and

zn=xn+Θn(xnxn1)xn+2=23Π𝒦(zn)(znλTznxn1).\begin{split}z_{n}=&\ x_{n}+\Theta_{n}(x_{n}-x_{n-1})\\ x_{n+2}=&\ \frac{2}{3}\ \Pi_{\mathcal{K}(z_{n})}(z_{n}-\lambda Tz_{n}-x_{n-1}).\end{split}

Before wrapping up this section and going to the convergence analysis it is worth noting that by applying suitable discretizations based on changing the update rule explicitly or implicitly, one can establish and design a variety of inertial projection proximal-type methods for solving parametric quasi variational inequalities (1). Convergence analyses for Algorithm 3.1 of the third-order projected dynamical system (9) are derived in the remaining part of this work.

3.3 Convergence of a discrete system

In this section, we derive the convergence of a solution to the implicit iterative scheme (28) and its equivalent variational form (29) given by Algorithm (3.1). However, other implicit (32) and explicit (30) proposed methods have a very similar arguments and follow the same guidlines except that there are some minor differences which is due to the values of the scalars formatting of α,β,γ\alpha,\beta,\gamma, and also due to the existing diverse discretization schemes. In proving that the approximate solution converges to a unique accumulation point, we need the following assumption:

Assumption 1. Suppose that xnxx_{n}\to x as nn\to\infty, then for any y𝒦(x)y\in\mathcal{K}(x) there exists a sequence {yn}\{y_{n}\} sucht hat yn𝒦(xn)y_{n}\in\mathcal{K}(x_{n}) and ynyy_{n}\to y. For all sequences {xn}\{x_{n}\} and {yn}\{y_{n}\} such that yn𝒦(xn)y_{n}\in\mathcal{K}(x_{n}), then y𝒦(x)y\in\mathcal{K}(x).

Theorem 6.

Let x𝒦(x)x\in\mathcal{K}(x) be the solution of the quasi variational inequality (1) and xn+2x_{n+2} be the approximate solution using the inertial proximal method in (29). If TT is monotone, then

(αβh+γh2)xxn+22αx2xn+1+2xn1xn22αxn+22xn+1+2xn1xn22+βhxn+12xn+xn12+γh2xnxn1+xxn+22γh2xnxn12.\begin{split}(\alpha-\beta h+\gamma h^{2})\|x&-x_{n+2}\|^{2}\leq\ \alpha\|x-2x_{n+1}+2x_{n-1}-x_{n-2}\|^{2}\\ &-\alpha\|x_{n+2}-2x_{n+1}+2x_{n-1}-x_{n-2}\|^{2}+\beta h\|x_{n+1}-2x_{n}+x_{n-1}\|^{2}\\ &+\gamma h^{2}\|x_{n}-x_{n-1}+x-x_{n+2}\|^{2}-\gamma h^{2}\|x_{n}-x_{n-1}\|^{2}.\end{split} (34)
Proof.

See [14].

Theorem 7.

Let x𝒦(x)x\in\mathcal{K}(x) be the solution of (1). Let xn+2x_{n+2} be the approximate solution of Algorithm 3.1, Suppose that the operator TT is monotone and Assumption 1 satisfies, then the generated sequence from (29) converges to the solution xx of the parametric quasi variational inequality (1) , i.e., limnxn+2=x.\displaystyle\lim_{n\to\infty}x_{n+2}=x.

Proof.

See [14].

4 Applications

We investigate two scenarios. The first one is the obstacle problem which is a class of free boundary problems that observe the dynamic behavior of a state variable described by a differential equation and studies the equilibrium states over an obstacle with fixed boundary conditions arise in financial mathematics and optimal control. The second one is the generalized Nash equilibrium problem which is an extension of the classical Nash equilibrium problem, in which each player’s strategy set depends on the rival player’s strategies.

4.1 The obstacle boundary value

Consider the second-order implicit obstacle boundary value problem, which have been discussed in Noor [23] as finding xx such that

{x¨(t)f(t)onΩ=[a,b]x(t)M(x(t))onΩ=[a,b](x¨(t)f(t))(x(t)M(x(t)))=0onΩ=[a,b]x(a)=0=x(b),\left\{\begin{array}[]{ll}-\ddot{x}(t)\geq f(t)&\text{on}\;\Omega=[a,b]\\ x(t)\geq M(x(t))&\text{on}\;\Omega=[a,b]\\ \big{(}-\ddot{x}(t)-f(t)\big{)}\big{(}x(t)-M(x(t))\big{)}=0&\text{on}\;\Omega=[a,b]\\ x(a)=0=x(b),\end{array}\right.

where f(t)f(t) is a continuous function and M(x(t))M(x(t)) is the following cost function

M(x(t))=k+infixi,wherek0.M(x(t))=k+\inf_{i}x^{i},\ \ \ \text{where}\ k\geq 0.

To see the connection between the obstacle problem and the quasi variational inequalities, we define the constraint closed convex-valued set

𝒦(x)={y:y01(Ω),yM(x)}\mathcal{K}(x)=\{y:y\in\mathcal{H}_{0}^{1}(\Omega),\ \ y\geq M(x)\}

where 01(Ω)\mathcal{H}_{0}^{1}(\Omega) is a Sobolev space. Introduce the energy functional corresponding to the obstacle problem as:

I[y]=Tx,y2f,y,y𝒦(x)I[y]=\langle Tx,y\rangle-2\langle f,y\rangle,\ \ \forall y\in\mathcal{K}(x)

where

Tx,y=ab(dydt)2𝑑t,andf,y=abf(t)y𝑑t.\langle Tx,y\rangle=\int_{a}^{b}\bigg{(}\frac{dy}{dt}\bigg{)}^{2}dt,\ \ \text{and}\ \ \langle f,y\rangle=\int_{a}^{b}f(t)y\ dt.

It is clear that the operator TT defined above is linear, symmetric and positive. Using the technique of Noor [25] one can show that the minimum of the functional II associated with the problem (2) on the closed convex-valued set 𝒦(x)\mathcal{K}(x) can be characterized by the inequality

Tx,yxf,yx,y𝒦(x).\langle Tx,y-x\rangle\geq\langle f,y-x\rangle,\ \ \forall y\in\mathcal{K}(x).

which is exactly the quasi variational inequality (1).

4.2 Generalized Nash equilibrium problems

The main concept in game theory is the Nash equilibrium. A Nash equilibrium is a set of strategies (constraints) assigned to each member (player) of the game. In other words, it is one such that no constraints across players are allowed. It is known that variational inequalities or variational equilibria can be extended to the Nash equilibria game. In the same fashion, generalized Nash equilibrium problems can be formulated in terms of QVIs due to Bensoussan [6], where the author dealt with infinite-dimensional strategy sets in which not only each player’s payoff function but also their strategy set depend on the other players strategies. Necessary and sufficient conditions for Nash equilibria of a game in normal strategic form cane be constructed in terms of the generated optimization problems.

Noncooperative Games in normal strategic form: Roughly speaking, when the feasible set of the game is actually the full Cartesian product of the individual strategy sets then the composed game is called a noncooperative game. In other words, players can only impact the cost functions of the rival players but not their feasible sets. Consider a finite set I={1,,n}I=\{1,...,n\} of players such that each player iIi\in I has a set of strategies (also called actions or constraints) denoted by SiS_{i}, and this set is a compact convex subset of a Hilbert space i\mathcal{H}_{i} defined by the set-valued mapping Si:iiS_{i}:\mathcal{H}\setminus\mathcal{H}_{i}\rightrightarrows\mathcal{H}_{i} where =iIi\mathcal{H}=\prod_{i\in I}\mathcal{H}_{i} is the ambient space. Each player ii controls their decision variable xiix^{i}\in\mathcal{H}_{i} such that the vector x=(x1,,xn)x=(x^{1},...,x^{n})\in\mathcal{H} describes the decision vector of all players. We often use the notation x=(xi,xi)x=(x^{i},x^{–i}), where xi=(xj)jI{i}=(x1,x2,,xi1,xi+1,,xn).x^{–i}=(x^{j})^{j\in I\setminus\{i\}}=(x^{1},x^{2},...,x^{i–1},x^{i+1},...,x^{n}). Let

𝒦(x):=i=1nSi(xi),\mathcal{K}(x):=\displaystyle\prod_{i=1}^{n}S_{i}(x^{-i}),

represents the full Cartesian product of the strategy sets. Furthermore, every player ii has a convex smooth utility (payoff) function Ui:U_{i}:\mathcal{H}\to\mathbb{R} defined as

Ui=iISiU_{i}=\displaystyle\prod_{i\in I}S_{i}

Then it is well known that GNEP consists in finding a vector x=(x1,,xn)𝒦(x)x^{*}=(x^{*1},...,x^{*n})\in\mathcal{K}(x^{*}) such that

Tx,xx0,x𝒦(x),\langle Tx^{*},x-x^{*}\rangle\geq 0,\ \ \forall x\in\mathcal{K}(x^{*}), (35)

where TT is a vector-valued function defined as

T(x)=xiUi(xi,xi), 1in.T(x)=\nabla_{x^{i}}U_{i}(x^{i},x^{-i}),\ \ \forall\ 1\leq i\leq n.

The goal of each player is to minimize utility with respect to the only variable xix^{i} under their control

minxi𝒦Ui(xi,xi)\min_{x^{i}\in\mathcal{K}}U_{i}(x^{i},x^{-i}) (36)

where 𝒦iISi\mathcal{K}\subseteq\prod_{i\in I}S_{i}. A point x𝒦x^{*}\in\mathcal{K} is said to be a variational equilibrium of a game (U,𝒦)(U,\mathcal{K}) if

U(x),xx0,x𝒦.\langle\nabla U(x^{*}),x-x^{*}\rangle\geq 0,\ \ \forall x\in\mathcal{K}. (37)

All the previous discussed formulas in this section can be reconstructed for the utility maximization problem by replacing min with max in (36) such that the vector TT has a line search align with the anti-gradient direction on the contrary of the gradient operator direction. i.e., the maximization problem can be slightly stated in the standard form (35).

Definition 4.

We say that x𝒦x^{*}\in\mathcal{K} is a Nash equilibrium of a game (U,𝒦)(U,\mathcal{K}) if for every player iIi\in I,

Ui(xi,xi)Ui(xi,xi),(xi,xi)𝒦.U_{i}(x^{i*},x^{-i*})\leq U_{i}(x^{i},x^{-i*}),\ \ \forall(x^{i},x^{-i*})\in\mathcal{K}.

Moreover, if 𝒦=iISi\mathcal{K}=\displaystyle\prod_{i\in I}S_{i}, we therefore have

Ui(xi,xi)=minxiSiUi(xi,xi).U_{i}(x^{i*},x^{-i*})=\min_{x^{i}\in S_{i}}U_{i}(x^{i},x^{-i*}).
Theorem 8 ([8]).

Let (U,𝒦)(U,\mathcal{K}) be a noncooperative game and xx^{*} is a Nash equilibrium then

U(x),xx0,xx+I𝒦(x)𝒦.\langle\nabla U(x^{*}),x-x^{*}\rangle\geq 0,\quad\forall x\in x^{*}+I_{\mathcal{K}}(x^{*})\cap\mathcal{K}. (38)

This is called a necessary optimality condition for QVI (1) in the form of Stampacchia-type [27] for a noncooperative game. Also, this QVI (38) is equivalent to the projection equation:

ΠIK(x)(U(x))=0,\Pi_{I_{K}(x^{*})}(-\nabla U(x^{*}))=0,

where ΠIK(x)\Pi_{I_{K}(x^{*})} denotes the orthogonal projection onto the internal cone IK(x)I_{K}(x^{*}) such that its associated dynamical system is defined by

x˙(t)=ΠIK(x)(U(x)).\dot{x}(t)=\Pi_{I_{K}(x)}(-\nabla U(x)). (39)

Such projected dynamical system describes the evolution of the game from a nonstationary initial point. These dynamical systems are of great importance due to their geometrical explanation. Since the antigradient direction U(x)-\nabla U(x) offers the players the steepest cost if xx stays on the boundary of the feasible set KK, thus projecting this direction could be performed in several ways including (39). A different dynamical system is to exploit the projection on the tangent cone of KK, that is

x˙(t)=ΠTK(x)(U(x)).\dot{x}(t)=\Pi_{T_{K}(x)}(-\nabla U(x)).

Another dynamical system is attained by performing the projection on the whole set KK, namely,

x˙(t)=ΠK(xαU(x))x,\dot{x}(t)=\Pi_{K}(x-\alpha\nabla U(x))-x,

where α\alpha is a fixed positive constant. Steady states of the latter two projected systems coincide with variational equilibria of the game (37). However, a Nash equilibrium is not necessarily a steady state. In this case we remark that its steady states coincide with the solutions to the quasi variational inequality (38), and consequently, equal Nash equilibria of the game.

5 Conclusion

In this paper we proposed a new high order projected dynamical system for solving parametric quasi vairiational inequalities. This approach is novel and attribute new algorithms which can be considered interchangeably as continuous-time versions and discrete-time counterparts iterative schemes. It can be expected that the techniques described in this paper will be useful for more elaborate dynamical models, such as stochastic models, and that the connection between such dynamical models and the solutions to quasi variational inequalities will provide a deeper understanding of generalized equilibrium problems for nonconvex scenarios. Equilibrium solutions (stationary points or trajectories) of the associated dynamical system converge to the solutions of the parametric variational inequality problems by the equivalent formulation of fixed points problems and variational inequalities. The proposed implicit and explicit algorithms may be extended for a broader class of generalized quasi equilibrium problems and even beyond the convexity scope to nonconvex equilibrium variational problems. The stability analysis of the novel dynamical system technique has been investigated. This approach usually provides qualitative behaviour of the system around the equilibrium points. One of the advantages of this approach is studying changes over time for energy-like functions without solving the differential equation analytically. Even the applicability and leverage of the approaches into real-applications, combining third-order dynamics into quasi variational inequalities still carries various challenges due to the computational complexity when proposing composite optimization algorithms for solving such systems.

Future research directions may focus on developing efficient algorithms, integrating machine learning techniques for parameter estimation, and extending the framework to stochastic environments and/or to nonmonotone manners whether on operators or in line searches for linearly convergence of algorithms. Another direction left to the future research is through exploring the dynamic constraint convex-valued set for different formats and layouts including two parameters or linear operators with mild conditions on symmetric matrices.

Finally, since QVI can be used to formulate the generalized Nash game in which not only each player’s payoff function but also their strategy set depend on the other players strategies, the QVI can attract ongoing attention to game theory. Merit functions such as the gap function is a powerful tool in the equivalent optimization formulation of the variational inequality. Since gap functions possess smooth properties when the constraints are represented by nonlinear inequalities, thus constructing gap functions for the QVI is a future research aim. From the viewpoint of application, it is essentially beneficial to study generalized Nash games that use gap functions, or more generally, merit functions and this would also be worth investigation.

References

  • [1] Alvarez, F., Attouch, H.: An inertial proximal method for maximal monotone operators via discretization of a nonlinear oscillator with damping. Set-Valued Analysis. 9, 3-11 (2001).
  • [2] Antipin, A.S., Jacimovic, M., Mijajlovic, N.: Extra gradient method for solving quasi variational inequalities. Optimization. 67, 103–112 (2018).
  • [3] Antipin, A.S., Mijailovic, N., Jacimovic, M.: A second-order continuous method for solving quasi-variational inequalities. Comput. Math. and Math. Phys. 51, 1856–1863 (2011).
  • [4] Antipin, A.S., Mijajlovic, N., Jacimovic, M.: A second-order iterative method for solving quasi-variational inequalities. Comput. Math. and Math. Phys. 53, 258–264 (2013).
  • [5] Bello-Cruz, Y., Hazaimah, O.: On the weak and strong convergence of modified forward-backward-half-forward splitting methods. Optimization Letters. 17, 3 (2022)
  • [6] Bensoussan, A., Goursat, M., Lions, J.L.: Controle impulsionnel et inequations quasi-variationnelles, C. R. Acad. Sci. Paris Ser. A 276, 1279–1284 (1973).
  • [7] Blum, E., Oettli, W.: From optimization and variational inequalities to equilibrium problems, Mathematics Students. 63, 123—145 (1994).
  • [8] Cavazzuti, E., Pappalardo, M., Passacantando, M.: Nash Equilibria, Variational Inequalities and Dynamical Systems. Journal of Optimization Theory and Applications. 114, 491–506 (2002).
  • [9] Dupuis P., Nagurney A.: Dynamical Systems and Variational Inequalities. Annals of Operations Research. 44, 9-42 (1993).
  • [10] Dong, J., Zhang, D., Nagurney, A.: A projected dynamical systems model of general financial equilibrium with stability analysis. Math. Comput. Modelling. 24, 35–44 (1996). 1
  • [11] Goeleven, D: Existence and uniqueness for a linear mixed variational inequality arising in electrical circuits with transistors. J. Optim. Theory Appl. 138, 397–406 (2008). 1
  • [12] Giannessi F., Maugeri A., Pardalos P.M.: Equilibrium Problems: Nonsmooth Optimization and Variational Inequality Models. Kluwer Academics Publishers, Holland (2001).
  • [13] Hai, T.N.: Dynamical systems for solving variational inequalities. J Dyn Control Syst. 28, 681–696 (2022).
  • [14] Hazaimah O.: A Third order dynamical system for mixed variational inequalities. https://arxiv.org/abs/2406.19345.
  • [15] Jabeen S., Noor M.A., Noor K.I.: Inertial iterative methods for general quasi variational inequalities and dynamical systems, J. Math. Anal. 11, 14–29 (2020).
  • [16] Khan, A.G., Noor, M.A., Noor, K.I.: Dynamical systems for quasi variational inequalities, Ann. Funct. Anal. 6(1), 193-209 (2015).
  • [17] Kinderlehrer D., Stampacchia G.: An Introduction to Variational Inequalities and Their Applications. Society for Industrial and Applied Mathematics (SIAM). Philadelphia, (2000).
  • [18] Mijajlovic, N., Jacimovic, M.: A proximal method for solving quasi-variational inequalities. Comput. Math. and Math. Phys. 55, 1981–1985 (2015).
  • [19] Mijajlovic N., Jacimovic M.: Some Continuous Methods for Solving Quasi-Variational Inequalities. Comput. Math. and Math. Phys. Pleiades Publishing, 58 (2), 190–195 (2018)
  • [20] Mijajlovic, N., Milojica, J., Noor, M. A.: Gradient-type projection methods for quasi variational inequalities, Optim. Lett. 13, 1885–1896 (2019).
  • [21] Noor, M.A.: General variational inequalities, Appl. Math. Letters. 1, 119–122 (1988).
  • [22] Noor, M.A., Noor, K.I.: Dynamical system technique for solving quasi variational inequalities. U.P.B. Sci. Bull., Series A, 84, 4, (2022).
  • [23] Noor M.A., Noor, K.I.: Some novel aspects of quasi variational inequalities. Earthline J. Math. Sci. 10(1), 1-66 (2022).
  • [24] Noor M.A., Noor K.I., Latif R.: Dynamical Systems and Variational Inequalities. Journal of Inequalities and Special Functions. 8, (11) 22-29 (2017).
  • [25] Noor, M.A., Noor, K.I., Rassias, M.T.: New Trends in General Variational Inequalities. Acta Appl Math 170, 981–1064 (2020)
  • [26] Noor M.A., Oettli, W.: On general nonlinear complementarity problems and quasi-equilibria, Le Matematiche (Catania). 49, 313—331 (1994).
  • [27] Stampacchia, G.: Formes bilineaires coercitives sur les ensembles convexes. Comptes Rendus Acad. Sci. Paris. 258, 4413-4416 (1964).
  • [28] Vasiliev, F.P.: Methods of Optimization. Moscow: Factorial press. (2002).