Loading [MathJax]/jax/output/SVG/jax.js
Research article Special Issues

A prediction-correction based proximal method for monotone variational inequalities with linear constraints

  • The monotone variational inequalities are being widely used as mathematical tools for studying optimal control problems and convex programming. In this paper, we propose a new prediction-correction method for monotone variational inequalities with linear constraints. The method consists of two procedures. The first procedure (prediction) utilizes projections to generate a predictor. The second procedure (correction) produces the new iteration via some minor computations. The main advantage of the method is that its main computational effort only depends on evaluating the resolvent mapping of the monotone operator, and its primal and dual step sizes can be enlarged. We prove the global convergence of the method. Numerical results are provided to demonstrate the efficiency of the method.

    Citation: Feng Ma, Bangjie Li, Zeyan Wang, Yaxiong Li, Lefei Pan. A prediction-correction based proximal method for monotone variational inequalities with linear constraints[J]. AIMS Mathematics, 2023, 8(8): 18295-18313. doi: 10.3934/math.2023930

    Related Papers:

    [1] Min Wang, Jiao Teng, Lei Wang, Junmei Wu . Application of ADMM to robust model predictive control problems for the turbofan aero-engine with external disturbances. AIMS Mathematics, 2022, 7(6): 10759-10777. doi: 10.3934/math.2022601
    [2] Cuijie Zhang, Zhaoyang Chu . New extrapolation projection contraction algorithms based on the golden ratio for pseudo-monotone variational inequalities. AIMS Mathematics, 2023, 8(10): 23291-23312. doi: 10.3934/math.20231184
    [3] Lu-Chuan Ceng, Shih-Hsin Chen, Yeong-Cheng Liou, Tzu-Chien Yin . Modified inertial subgradient extragradient algorithms for generalized equilibria systems with constraints of variational inequalities and fixed points. AIMS Mathematics, 2024, 9(6): 13819-13842. doi: 10.3934/math.2024672
    [4] Pongsakorn Yotkaew, Nopparat Wairojjana, Nuttapol Pakkaranang . Accelerated non-monotonic explicit proximal-type method for solving equilibrium programming with convex constraints and its applications. AIMS Mathematics, 2021, 6(10): 10707-10727. doi: 10.3934/math.2021622
    [5] P. Dhivya, D. Diwakaran, P. Selvapriya . Best proximity points for proximal Górnicki mappings and applications to variational inequality problems. AIMS Mathematics, 2024, 9(3): 5886-5904. doi: 10.3934/math.2024287
    [6] YeongJae Kim, YongGwon Lee, SeungHoon Lee, Palanisamy Selvaraj, Ramalingam Sakthivel, OhMin Kwon . Design and experimentation of sampled-data controller in T-S fuzzy systems with input saturation through the use of linear switching methods. AIMS Mathematics, 2024, 9(1): 2389-2410. doi: 10.3934/math.2024118
    [7] Jun Moon . The Pontryagin type maximum principle for Caputo fractional optimal control problems with terminal and running state constraints. AIMS Mathematics, 2025, 10(1): 884-920. doi: 10.3934/math.2025042
    [8] Cunlin Li, Wenyu Zhang, Baojun Yang, Hooi Min Yee . A multi-player game equilibrium problem based on stochastic variational inequalities. AIMS Mathematics, 2024, 9(9): 26035-26048. doi: 10.3934/math.20241271
    [9] Regina S. Burachik, Bethany I. Caldwell, C. Yalçın Kaya . Douglas–Rachford algorithm for control- and state-constrained optimal control problems. AIMS Mathematics, 2024, 9(6): 13874-13893. doi: 10.3934/math.2024675
    [10] Na Mi, Juhe Sun, Li Wang, Yu Liu . Alternating direction method for the fixed point problem of set-valued mappings with second-order cone double constraints. AIMS Mathematics, 2023, 8(3): 6389-6406. doi: 10.3934/math.2023323
  • The monotone variational inequalities are being widely used as mathematical tools for studying optimal control problems and convex programming. In this paper, we propose a new prediction-correction method for monotone variational inequalities with linear constraints. The method consists of two procedures. The first procedure (prediction) utilizes projections to generate a predictor. The second procedure (correction) produces the new iteration via some minor computations. The main advantage of the method is that its main computational effort only depends on evaluating the resolvent mapping of the monotone operator, and its primal and dual step sizes can be enlarged. We prove the global convergence of the method. Numerical results are provided to demonstrate the efficiency of the method.



    Let Dn be a nonempty, closed and convex set and f:nn be a given continuous and monotone mapping. The variational inequality problem, denoted by Ⅵ(D,f), is to find a vector xD such that

    xx,f(x)0,xD. (1.1)

    The Ⅵ(D,f) has found many important applications in areas such as nonlinear complementarity problems (where D=n+) [1], traffic equilibrium and economic problems [2,3]. For recent applications and numerical methods of the Ⅵ(D,f), we refer the reader to [4,5,6] and the references therein.

    In this paper, we consider a special case of the general Ⅵ problem, where the set D is assumed to have the following form:

    D={xn|xX,Ax=b}, (1.2)

    here Am×n is a given matrix, bm is a given vector, Xn is a nonempty, closed and convex subset. The solution set of (1.1) and (1.2), denoted by Ω, is assumed to be nonempty. Note that the Ⅵ problem (1.1) and (1.2) is closely related to the convex optimization problem with linear equality constraints:

    minθ(x)s.t.Ax=b, xX. (1.3)

    To show this, we recall the first order optimization conditions of problem (1.3). Let x be a minimum point of the convex function θ(x) over D and ξ be any vector in θ(x), where θ() denotes the subdifferential operator of θ(). Then for any feasible direction dn at x, we have (ξ)Td>0. It means that, the following variational inequality problem is captured:

    xx,ξ0,xD. (1.4)

    Thus, solving the Ⅵ problem (1.4) amounts to solving (1.3). This Ⅵ characterization is also used in e.g., [7,8].

    The Ⅵ problem (1.1) and (1.2) or its equivalent form (1.3) is one of the fundamental problems in convex optimization. In particular, it includes linear programming, conic and semidefinite optimization as special cases. It can find many applications in compressed sensing, image processing and data mining, see, e.g., [9,10,11]. We refer to [12] for recent examples and discussions. To solve the Ⅵ problem (1.1) and (1.2), some proximal-like algorithms have been developed over the past years, see e.g., [13] for a review. One benchmark method is the augmented Lagrangian Method (ALM) [14,15] for nonlinear problems. The ALM is applicable to solve Ⅵ(D,f) (1.1) and (1.2). More specifically, for a given uk=(xk,λk), ALM uses the following procedure to carry out the new iteration uk+1=(xk+1,λk+1)X×m:

    Find xk+1X such that

    xxk+1,f(xk+1)ATλk+βAT(Axk+1b)0,xX, (1.5a)

    then update λk+1 via

    λk+1=λkβ(Axk+1b), (1.5b)

    where β0 is a given penalty parameter for the violation of the linearly constraints. To make ALM (1.5) more efficient and flexible, some alternative strategies can be used. For example, some self-adaptive rules can be carried to adjust the parameter β based on certain strategies [16,17,18,19]. We can also use some correction technologies to the output point [20,21]. Let ˜uk denote the predictor generated by ALM. A simple and effective correction scheme is

    uk+1=ukαk(uk˜uk), (1.6)

    where 0<αk<2 is the step size parameter; see, e.g., [22,23] for details.

    The main computational cost at each iteration of ALM is to solve the subproblem (1.5a), which is still a variational inequality, structurally the same as the original problem. So in many cases the ALM (1.5a) is easy to iterate only if A is an identity matrix. This is because, for this case, the solution of the subproblem (1.5a) corresponds to the proximal mappings of f and it usually has closed-form solutions or can be efficiently solved up to a high precision. However, in many applications in sparse optimization, we often encounter the case where A is not an identity matrix, and the resulting subproblem (1.5a) no longer has closed-form solutions. For this case, solving the subproblem (1.5a) could be computationally intensive because of the costly evaluation of (ATA+1βf)1(Aυ). Therefore, efficient numerical methods with implementable iterative scheme are highly desired.

    Recently, several techniques attempting to overcome this difficulty have been proposed. In the framework of the proximal point algorithm (PPA), there are two relevant approaches. The first one is regularization. By adding a customized regularization term to the saddle-point reformulation of (1.1) and (1.2), the primal subproblem becomes easy as it only involves a simple evaluation, see the customized PPA algorithms proposed in e.g., [24,25,26,27]. We refer the reader to e.g., [28,29] of the linearized regularization term for the separable case of problem (1.2). The second one employs prediction-correction technology which adds an asymmetrical proximal term to make a prediction, and then introduces a simple correction step to guarantee the convergence, see e.g., [30,31,32]. In this paper, we propose a new prediction-correction method for the Ⅵ(D,f). At each iteration, the method first makes a simple prediction step to obtain a point, and then generates a new iteration via a minor correction to the predictor. The reduced subproblems are easy to solve when the resolvent mapping of f can be efficiently evaluated. As can be seen in the Section 5, the proposed method converges faster with less iterations to achieve the same accuracy on most numerical cases.

    The rest of this paper is organized as follows. In Section 2, we review some preliminaries which are useful for further analysis. In Section 1, we present the implementable prediction-correction method for Ⅵ(D,f). In Section 4, we establish the global convergence of the proposed method. The computational experiment is presented in Section 5. Finally, we make a conclusion in Section 6.

    In this section, we reformulate the Ⅵ problem (1.1) and (1.2) in succinct form. Let Ω=X×m. By attaching a Lagrange multiplier vector λm to the linear constraints Ax=b, the Ⅵ problem (1.1) and (1.2) can be converted to the following form:

    xx,f(x)ATλλλ,Axb0,(x,λ)TΩ. (2.1)

    By denoting

    u:=(xλ),F(u):=(f(x)ATλAxb), (2.2)

    we can rewrite (2.1) into the following more compact form

    uu,F(u)0,uΩ. (2.3)

    Henceforth, we will denote the Ⅵ problem (2.2) and (2.3) by Ⅵ(Ω,F). Now, we make some basic assumptions and summarize some well known results of Ⅵ, which will be used in subsequent analysis.

    (A1) X is a simple closed convex set.

    A set is said to be simple if the projection onto it can be easily obtained. Here, the projection of a point a onto the closed convex set X, denoted by PX(a), is defined as the nearest point xX to a, i.e.,

    PX(a)=argmin{xa |xX}.

    (A2) The mapping f is point-to-point, monotone and continuous.

    A mapping F:nn is said to be monotone on Ω if

    uυ,F(u)F(υ)0,u,υΩ. (2.4)

    (A3) The solution set of (2.2) and (2.3), denoted by Ω, is nonempty.

    Remark 1. The mapping F() defined in (2.3) is monotone with respect to Ω since

    (F(u)F(˜u))T(u˜u)0,  u,˜uΩ. (2.5)

    Proof.

    (F(u)F(˜u))T(u˜u)=(f(x)f(˜x)AT(λ˜λ)A(x˜x))T(x˜xλ˜λ)=(f(x)f(˜x))T(x˜x)0, (2.6)

    where the last inequality follows from the monotone property of f.

    Let G be a symmetric positive definite matrix. The G-norm of the vector u is denoted by uG:=u,Gu. In particular, when G=I, u:=u,u is the Euclidean norm of u. For a matrix A, A denotes its norm A:=max{Ax:x1}.

    The following well-known properties of the projection operator will be used in the coming analysis. The proofs can be found in textbooks, e.g., [2,33].

    Lemma 1. Let Xn be a nonempty closed convex set, PX() be the projection operator onto X under the G-norm. Then

    yPX(y),G(xPX(y))0,yn,xX. (2.7)
    PX(x)PX(y)GxyG,x,yn. (2.8)
    xPX(y)2Gxy2GyPX(y)2G,yn,xX. (2.9)

    For any arbitrary positive scalar β and uΩ, let e(u,β) denote the residual function associated with the mapping F, i.e.,

    e(u,β)=uPΩ[uβF(u)]. (2.10)

    Lemma 2. Solving VI(Ω,F) is equivalent to finding the zero point of the mapping

    e(u,β):=(e1(u,β)e2(u,β))=(xPX{xβ[f(x)ATλ]}β(Axb)). (2.11)

    Proof. See [2], pp 267.

    We now formally present the new prediction-correction method for the Ⅵ(Ω,F) (Algorithm 1). The method can be viewed as a generalization of [31] in relaxed case.

    Algorithm 1: A prediction-correction based method (PCM) for the Ⅵ(Ω,F).
    Step 0: Initialization step.
    Given a small number ϵ>0. Take γ(0,2), u0n+m; set k=0. Choose the parameters r>0,s>0, such that
        rs>14ATA.     (3.1)
    Step 1: Prediction step.
    Generate the predictor ˜xk via solving the following projection equation:
        ˜xk=PX[xk1r(f(˜xk)ATλk)].    (3.2a)
    Then update ˜λkm via
        ˜λk=λk1s(A˜xkb).    (3.2b)
    Step 2: Correction step.
    Generate the new iteration uk+1=(xk+1,λk+1) via
        xk+1=xkγ(xk˜xk)γ2rAT(λk˜λk),    (3.3a)
    and
        λk+1=λk+γ2sA(xk˜xk)γ(I12rsAAT)(λk˜λk).    (3.3b)

    Step 3: Convergence verification.
    If uk+1ukϵ, stop; otherwise set k:=k+1; go to Step 1.

    Remark 2. Note that the regularized projection Eq (3.2a) amounts to solving the following VI problem

     x˜xk,1r(f(˜xk)ATλk)+(˜xkxk)0,xX, (3.4)

    which represents to find ˜xkX such that

    0˜xk+1rf(˜xk)(1rATλk+xk). (3.5)

    We can rewrite the above equation as

    ˜xk(I+1rf)1(1rATλk+xk). (3.6)

    Thus, the subproblem (3.2a) is equivalent to an evaluation of the resolvent operator (I+1rf)1 when X is the finite-dimensional Euclidean spaces [34]. Notice that ALM (1.5a) needs to evaluate the operator (ATA+1βf)1. Therefore, the resulting subproblem (3.2a) could be much easier to solve than ALM (1.5a). On the other hand, the correction step only consists of some elementary manipulations. Thus, the resulting method (3.2a)–(3.3b) is easily implementable.

    Remark 3. The parameters 1r and 1s in the prediction step can be viewed as two step sizes for the projection step (3.2a) and dual step, respectively. Using the step size condition rs>14ATA of this algorithm, the parameters 1r and 1s can be chosen larger values compared to the condition rs>ATA of some other primal dual algorithms, e.g., linearized ALM, customized PPA algorithms. This larger step size is usually beneficial to the effectiveness and robustness of the algorithm. In Section 5 we will empirically see that our algorithm is significantly faster than some other primal dual algorithms by allowing larger step sizes.

    In the following, we will focus our attention to solving Ⅵ(Ω,F). But at the beginning, to make the notation of the proof more succinct, we define some matrices:

    G=(rI12AT12AsI),Q=(rIAT0sI). (4.1)

    Obviously, when rs>14ATA, G is a positive definite matrix. Now, we start to prove the global convergence of the sequence {uk}. Towards this end, we here follow the work [35] to reformulate the algorithm into a prediction-correction method and establish its convergence results. We first prove some lemmas. The first lemma quantifies the discrepancy between the point ˜uk and a solution point of Ⅵ(Ω,F).

    Lemma 3. Let {uk} be generated by the PCM and {˜uk} be generated by PCM (3.2), and the matrix Q be given in (4.1). Then we have

    u˜uk,F(˜uk)Q(uk˜uk)0,uΩ. (4.2)

    Proof. Note that the sequence {˜uk} generated by (3.2) is actually solutions of the following VIs:

     x˜xk,f(˜xk)ATλkr(xk˜xk)0,xX, (4.3)

    and

    λ˜λk,A˜xkbs(λk˜λk)0,λm. (4.4)

    Combining (4.3) and (4.4) yields

    x˜xk,f(˜xk)AT˜λkAT(λk˜λk)r(xk˜xk)λ˜λ,A˜xkbs(λk˜λk)0,uΩ. (4.5)

    It can be rewritten as

    (x˜xkλ˜λk), (f(˜xk)AT˜λkA˜xkb)(rIAT0sI)(xk˜xkλk˜λk)0,uΩ. (4.6)

    Using the notation of F(u) (2.3) and Q (4.1), the assertion (4.2) is proved.

    The following lemma characterizes the correction step by a matrix form.

    Lemma 4. Let {uk} be generated by the PCM and {˜uk} be generated by PCM (3.2), Then we have

    ukuk+1=γM(uk˜uk), (4.7)

    where

    M=(I12rAT12sAIAAT2rs). (4.8)

    Proof. From the correction Step (3.3a) and (3.3b), we have

    (xk+1λk+1)=(xkλk)γ((xk˜xk)+AT2r(λk˜λk)12sA(xk˜xk)+(I12rsAAT)(λk˜λk)), (4.9)

    which can be written as

    (xkxk+1λkλk+1)=γ(I12rAT12sAIAAT2rs)(xk˜xkλk˜λk). (4.10)

    By noting the matrix M (4.8), the proof is completed.

    Using the matrices Q (4.1) and M (4.8), we define

    H=QM1, (4.11)

    then we have Q=HM. The inequality (4.2) can be written as

    γu˜uk,F(˜uk)HM(uk˜uk)0,uΩ. (4.12)

    Substituting (4.7) into (4.12) and using the monotonicity of F (see (2.5)), we obtain

    ˜ukΩ,u˜uk,F(u)u˜uk,H(ukuk+1),uΩ. (4.13)

    Now, we prove a simple fact for the matrix H in the following lemma.

    Lemma 5. The matrix H defined in (4.11) is positive definite for any r>0,s>0, rs>14ATA.

    Proof. For the matrix Q defined by (4.1), we have

    Q1=(1rI1rsAT01sI).

    Thus, it follows from (4.1) that

    H1=MQ1=(I 12rAT12sA IAAT2rs)(1rI1rsAT01sI)=(1rI12rsAT12rsA1sI). (4.14)

    For any any r>0,s>0 satisfying rs>14ATA, H1 is positive definite, and the positive definiteness of H is followed directly.

    Then we show how to deal with the right-hand side of (4.13). The following lemma gives an equivalent expression of it in terms of uukH and uuk+1H.

    Lemma 6. Let {uk} be generated by the PCM and {˜uk} be generated by PCM (3.2). Then we have

    u˜uk,H(ukuk+1)=12(uuk+12Huuk2H)+12uk˜uk2G, uΩ, (4.15)

    where the matrix G=γ(QT+Q)γ2MTHM.

    Proof. Applying the identity

    ab,H(cd)=12(ad2Hac2H)+12(cb2Hdb2H), (4.16)

    to the right term of (4.13) with a=u,b=˜uk,c=uk,d=uk+1, we obtain

    u˜uk,H(ukuk+1)=12(uuk+12Huuk2H)+12(uk˜uk2Huk+1˜uk2H). (4.17)

    For the last term of (4.17), we have

    uk˜uk2Huk+1˜uk2H=uk˜uk2H(uk˜uk)(ukuk+1)2H(4.7)=uk˜uk2H(uk˜uk)γM(uk˜uk)2H=2γuk˜uk,HM(uk˜uk)γ2uk˜uk,MTHM(uk˜uk)(4.11)=uk˜uk,(γ(QT+Q)γ2MTHM)(uk˜uk). (4.18)

    the assertion is proved.

    Now, we investigate the positive definiteness for the matrix G. Using (4.11), we have

    G=γ(QT+Q)γ2MTHM=γ(QT+Q)γ2MTQ=γ(2rIATA2sI)γ2(I 12sAT12rA IAAT2rs)(rIAT0sI)=γ(2rIATA2sI)γ2(rI12AT12AsI)=(2γγ2)(rI12AT12AsI). (4.19)

    Thus, when rs>14ATA and γ(0,2), the matrix G is guaranteed to be positive definite, and we can easily obtained the contraction property of the algorithm. This is given by the following theorem.

    Theorem 1. Suppose the condition

    rs>14ATA (4.20)

    holds. Let the relaxation factor γ(0,2). Then, for any u=(x,λ)TΩ, the sequence uk+1=(xk+1,λk+1)T generated by PCM satisfies the following inequality:

    uk+1u2Huku2Huk˜uk2G. (4.21)

    Proof.

    Combining (4.13) and (4.15), we have

    ˜ukΩ,u˜uk,F(u)12(uuk+12Huuk2H)+12uk˜uk2G,uΩ. (4.22)

    Note that uΩ. We get

    uku2Huk+1u2Huk˜uk2G+2˜uku,F(u).

    Since u is a solution of (2.3) and ˜ukΩ, we have

    ˜uku,F(u)0, (4.23)

    and thus

    uku2Huk+1u2Huk˜uk2G.

    The assertion (4.21) follows directly.

    Based on the above results, we are now ready to prove the global convergence of the algorithm.

    Theorem 2. Let {uk} be the sequence generated by PCM for the Ⅵ problem (2.2) and (2.3). Then, for any r>0,s>0 satisfying rs>14ATA and γ(0,2), the sequence {uk} converges to a solution of Ⅵ(Ω,F).

    Proof. We follows the standard analytic framework of contraction-type methods to prove the convergence of the proposed algorithm. It follows from Theorem 1 that {uk} is bounded. Then we have that

    limkuk˜ukG=0. (4.24)

    Consequently,

    limkxk˜xk=0, (4.25)

    and

    limkA˜xkb=limks(λk˜λk)=0. (4.26)

    Since

    ˜xk=PX[˜xk1r(f(˜xk)AT˜λk)+(xk˜xk)+1rAT(λk˜λk)], (4.27)

    and

    ˜λk=λk1s(A˜xkb), (4.28)

    it follows from (4.25) and (4.26) that

    {limk˜xkPX[˜xk1r(f(˜xk)AT˜λk)]=0,(4.29a)limkA˜xkb=0.(4.29b)

    Because ˜uk is also bounded, it has at least one cluster point. Let u be a cluster point of ˜uk and let ˜ukj be the subsequence converges to u. It follows from (4.29) that

    {limj˜xkjPX[˜xkj1r(f(˜xkj)AT˜λkj)]=0,(4.30a)limjA˜xkjb=0.(4.30b)

    Consequently, we have

    {xPX[x1r(f(x)ATλ)]=0,(4.31a)Axb=0.(4.31b)

    Using the continuity of F and the projection operator PΩ(), we have that u is a solution of Ⅵ(Ω,F). On the other hand, by taking limits over the subsequences in (4.24) and using limj˜ukj=u. we have that, for any k>kj,

    ukuHukjuH.

    Thus, the sequence {uk} converges to u, which is a solution of Ⅵ(Ω,F).

    In this paper, we test the performance of PCM (3.2a)–(3.3b) for solving the basis pursuit (BP) and matrix completion problem. All the simulations are performed on a Lenovo Laptop with CPU Intel with 2.81GHz and 16GB RAM memory, using Matlab R2015b.

    The BP problem arises from areas such as the communities of information theory, signal processing, statistics, machine learning. it seeks to encode a large sparse signal representations through a relatively small number of linear equations. The BP problem can be cast as the following equality-constrained l1 minimization problem

    minx1s.t.Ax=b, (5.1)

    where xn, data Am×n with m<n, bm. Here we assume that A has full row rank. By invoking the first-order optimality condition, BP is equivalent to Ⅵ (1.1) and (1.2) with f(x)=x1. Applying PCM with γ=1 for this problem, we get the following iterative scheme:

    {˜xkPn[xk1r(˜xk1ATλk)],(5.2a)˜λk=λk1s(A˜xkb),(5.2b)xk+1=˜xk12rAT(λk˜λk),(5.2c)λk+1=λk+12sA(xk˜xk)(IAAT2rs)(λk˜λk).(5.2d)

    Note that the projection (5.2a) is equivalent to the l1 shrinkage operation:

    ˜xk:=Shrink(xk+1rATλk,1r),

    where the l1 shrinkage operator, denoted by Shrink(M,ξ), is defined as

    [Shrink(M,ξ)]i:={Miξ,if Mi>ξ,Mi+ξ,if Mi<ξ,0,if |Mi|ξ,i=1,2,,n.

    In our experiments, we focus on comparing our algorithm with the linearized ALM [36] (L-ALM) and the customized PPA (C-PPA) in [27] and verifying its efficiency. Similar with PCM, L-ALM and C-PPA also depend on Shrink operation, which have the same easiness of implementation.

    The data used in this experiment is similar to the one employed in [37]. The basic setup of the problem is as follows. The data matrix A is a i.i.d. standard Gaussian matrix generated by the randn(m, n) function in MATLAB with m=n/2. The original sparse signal xtrue is sampled from i.i.d. standard Gaussian distributions with m/5 nonzero values. The output b is then created as the signs of b=Ax. It is desirable to test problems that have a precisely known solution. In fact, when the original signal xtrue is very sparse, it reduces to a minimization problem. The parameters used in the numerical experiments are similar to that in [19,38]: we set the relaxation factor γ=1, s(25,100), r=1.01ATA/(4s) for PCA and s(25,100), r=1.01ATA/s for C-PPA. (In order to ensure the convergence of L-ALM and C-PPA, the parameters r,s should satisfy rs>ATA.) We set the criterion error as min{relL=xkxk1,relS=Axkb}, and declare successful recovery when this error is less than Tol=103. In all the tests, the initial iteration is (x0,λ0)=(0,1).

    We first test the sensitivity of γ of PCM. We fix s=50 and choose different values of γ in the interval [0.6,1.8]. The number of iterations required are reported in Figure 1. The curves in Figure 1 indicate that γ(1,1.4) is preferable when we implement Algorithm PCM in practice.

    Figure 1.  Sensitivity test on the relaxation factor of γ.

    In order to investigate the stability and efficiency of our algorithms, we test 8 groups of problems with different n and we generated the model by 10 times and reported the average results. The comparisons of these algorithms for small BP problems are presented in Tables 13.

    Table 1.  Numerical Results for Basis Pursuit problem s=25.
    PCM L-ALM C-PPA
    n Iter. relL relS CPU(s) Iter. relL relS CPU(s) Iter. relL relS CPU(s)
    100 81 4.2e-04 1.9e-04 0.03 161 1.6e-04 7.5e-04 0.03 232 8.2e-05 8.6e-04 0.04
    300 88 5.1e-04 7.3e-04 0.05 207 7.5e-05 6.3e-04 0.07 311 5.3e-05 9.4e-04 0.10
    600 110 9.2e-05 7.9e-04 0.54 262 7.2e-05 8.9e-04 0.62 374 6.0e-05 9.4e-04 1.01
    1000 141 4.4e-05 9.8e-04 1.81 291 3.3e-05 7.9e-04 2.21 411 3.2e-05 9.4e-04 2.86
    1500 139 6.7e-05 6.6e-04 3.95 359 7.0e-05 9.4e-04 5.04 484 6.9e-05 9.0e-04 6.54
    2000 151 6.3e-05 7.9e-04 6.55 406 1.2e-05 9.7e-04 9.24 536 1.3e-05 9.7e-04 12.08
    2500 171 3.9e-05 8.5e-04 11.57 519 1.3e-05 8.9e-04 18.60 635 1.8e-05 9.8e-04 22.69
    3000 199 2.9e-05 8.8e-04 18.82 585 1.7e-05 9.8e-04 29.46 709 1.7e-05 9.0e-04 35.61

     | Show Table
    DownLoad: CSV
    Table 2.  Numerical Results for Basis Pursuit problem s=50.
    PCM L-ALM C-PPA
    n Iter. relL relS CPU(s) Iter. relL relS CPU(s) Iter. relL relS CPU(s)
    100 80 5.9e-04 7.0e-04 0.03 151 3.7e-04 9.6e-04 0.03 224 1.4e-04 9.4e-04 0.04
    300 87 6.1e-04 8.1e-04 0.05 203 1.2e-04 8.5e-04 0.09 299 3.3e-05 9.0e-04 0.10
    600 110 1.1e-04 6.1e-04 0.49 249 3.5e-05 9.9e-04 0.63 331 5.7e-05 8.6e-04 0.87
    1000 137 4.0e-05 8.5e-04 1.75 266 3.2e-05 7.9e-04 1.80 370 3.1e-05 8.3e-04 2.62
    1500 135 3.5e-05 8.6e-04 3.54 312 3.6e-05 8.5e-04 4.35 424 3.4e-05 8.2e-04 5.84
    2000 146 5.4e-05 7.6e-04 6.38 331 1.3e-05 9.8e-04 7.57 445 1.5e-05 9.5e-04 10.19
    2500 143 4.1e-05 9.7e-04 9.67 375 1.4e-05 9.3e-04 13.41 491 1.5e-05 9.1e-04 17.63
    3000 170 3.1e-05 9.7e-04 16.00 418 1.5e-05 9.5e-04 20.70 533 1.6e-05 8.9e-04 26.42

     | Show Table
    DownLoad: CSV
    Table 3.  Numerical Results for Basis Pursuit problem s=75.
    PCM L-ALM C-PPA
    n Iter. relL relS CPU(s) Iter. relL relS CPU(s) Iter. relL relS CPU(s)
    100 82 5.3e-04 5.7e-04 0.02 164 2.7e-04 6.2e-04 0.03 236 6.3e-05 9.2e-04 0.04
    300 99 1.9e-04 7.9e-04 0.05 213 9.4e-05 5.9e-04 0.07 284 4.2e-05 9.2e-04 0.09
    600 117 6.7e-05 8.9e-04 0.55 226 1.1e-04 9.9e-04 0.58 316 8.5e-05 9.3e-04 0.84
    1000 179 4.2e-05 8.7e-04 2.35 253 5.4e-05 9.5e-04 1.68 369 1.9e-05 9.5e-04 2.47
    1500 159 3.4e-05 9.3e-04 4.20 299 3.4e-05 8.1e-04 4.13 399 3.5e-05 7.9e-04 5.39
    2000 142 6.6e-05 9.4e-04 6.18 313 1.3e-05 9.0e-04 7.32 410 1.9e-05 9.7e-04 9.56
    2500 139 2.9e-05 9.8e-04 9.41 337 1.4e-05 1.0e-03 12.20 444 1.1e-05 9.8e-04 16.00
    3000 168 3.7e-05 9.0e-04 15.92 370 1.6e-05 8.2e-04 18.54 464 2.2e-05 9.0e-04 23.22

     | Show Table
    DownLoad: CSV

    From Tables 13, it can be seen that the PCM performs the best, both in terms of number of iterations and CPU time for all test cases. These numerical results illustrate that if the step size condition relaxed, can indeed be beneficial to yield larger step sizes, which could accelerate the convergence of algorithm.

    To verify the performance results of our algorithm, we plotted the approximation error relL=xkxk1,relS=Axkb achieved for n=1500,s=50 by each of the algorithms versus the iteration number k in Figures 2 and 3, respectively. It is clear to see that PCM outperforms all the other algorithms significantly in terms of number of iterations.

    Figure 2.  Relative errors of relL=xkxk1.
    Figure 3.  Relative errors of relS=Axkb.

    Matrix completion problem (MC) comes from many fields such as signal processing, statistics, machine learning communities. It tries to recover the low-rank matrix X from its incomplete known entries. Mathematically, its convex formula is as follows:

    minXs.t.Xij=Mij,(i,j)Ω, (5.3)

    where X is the nuclear norm of X, M is the unknown matrix with p available sampled entries and Ω is a set of pairs of indices of cardinality p. By invoking the first-order optimality condition, MC can also be equivalent to Ⅵ (1.1) and (1.2) with f(x)=X.

    The basic setup of the problem is as follows. We first generate two random matrices MLn×ra and MRn×ra, all with i.i.d. standard Gaussian entries, and then set the low-rank matrix M=MLMTR. The available index set Ω is randomly uniformly sampled in all cardinality sets |Ω|. We denote the oversampling factor (OS) by OS=|Ω|/ra(2nra), i.e., the ratio of sample size to degrees of freedom in an asymmetric matrix of rank ra. The relative error of the approximation X is defined as

    relative error=XΩMΩFMΩF. (5.4)

    We set the relative error Tol=105 as the tolerance for all algorithms. In all tests, the initial iteration is (X0,Λ0)=(0,0). The parameters used in the numerical experiments are set as follows: we set r=0.006, s=1.01/(4r), γ=1 for PCM, s=1.01/r for L-ALM and C-PPA.

    Tables 46 list the comparison between these algorithms with three different OS values. The results confirm that PCM outperforms other methods in terms of computation time and number of iterations in all cases.

    Table 4.  Comparison results of PCM, L-ALM, C-PPA (OS=5).
    Problems PCM L-ALM C-PPA
    n ra Rel.err. Iter. Rel.err. Iter. Rel.err. Iter.
    100 5 9.71e-06 60 6.22e-05 101 5.83e-05 101
    100 10 9.74e-06 18 9.17e-06 38 9.15e-06 37
    200 5 9.40e-06 69 1.41e-04 101 1.33e-04 101
    200 10 8.88e-06 27 9.14e-06 56 9.15e-06 55
    500 10 9.67e-06 43 9.29e-06 72 8.96e-06 71
    500 15 8.93e-06 31 9.87e-06 50 9.06e-06 49
    1000 10 9.21e-06 71 9.23e-06 119 9.52e-06 117
    1500 10 9.92e-06 95 9.67e-06 166 9.41e-06 165

     | Show Table
    DownLoad: CSV
    Table 5.  Comparison results of PCM, L-ALM, C-PPA (OS=10).
    Problems PCM L-ALM C-PPA
    n ra Rel.err. Iter. Rel.err. Iter. Rel.err. Iter.
    100 5 8.67e-06 14 7.66e-06 29 7.67e-06 28
    100 10 8.39e-06 13 9.10e-06 27 9.12e-06 26
    200 5 6.84e-06 21 8.85e-06 36 8.67e-06 35
    200 10 5.78e-06 7 6.17e-06 13 8.32e-06 10
    500 10 9.13e-06 24 8.22e-06 39 8.86e-06 38
    500 15 8.34e-06 17 9.13e-06 26 9.74e-06 25
    1000 10 8.22e-06 48 9.25e-06 77 9.33e-06 76
    1500 10 9.41e-06 66 9.39e-06 112 9.40e-06 111

     | Show Table
    DownLoad: CSV
    Table 6.  Comparison results of PCM, L-ALM, C-PPA (OS=15).
    Problems PCM L-ALM C-PPA
    n ra Rel.err. Iter. Rel.err. Iter. Rel.err. Iter.
    100 5 9.62e-06 11 8.92e-06 23 8.99e-06 22
    100 10 9.62e-06 13 9.15e-06 27 9.20e-06 26
    200 5 4.73e-06 13 9.27e-06 21 8.34e-06 20
    200 10 7.98e-06 6 6.15e-06 13 8.34e-06 10
    500 10 8.15e-06 17 8.51e-06 25 9.93e-06 24
    500 15 4.58e-06 10 4.22e-06 14 7.19e-06 13
    1000 10 9.35e-06 35 9.38e-06 56 9.81e-06 55
    1500 10 9.75e-06 50 9.20e-06 84 9.46e-06 83

     | Show Table
    DownLoad: CSV

    This paper proposes a new prediction-correction method for solving the monotone variational inequalities with linear constraints. At the prediction step, the implementation is carried out by a simple projection. At the correction step, the method introduces a simple updating step to generate the new iteration. We establish the global convergence of the method. The convergence condition of the method also allows larger step sizes that can in potential make the algorithm numerically converge faster. The numerical experiments approve the efficiency of the proposed methods. The future work is to explore combining self adaptive technique for the method. Besides, further applications of our method are expected.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    This research was funded by NSF of Shaanxi Province grant number 2020-JQ-485.

    The authors declare no conflict of interest.



    [1] R. Rockafellar, J. Sun, Solving monotone stochastic variational inequalities and complementarity problems by progressive hedging, Math. Program., 174 (2019), 453–471. https://doi.org/10.1007/s10107-018-1251-y doi: 10.1007/s10107-018-1251-y
    [2] D. Bertsekas, E. Gafni, Projection methods for variational inequalities with application to the traffic assignment problem, In: Nondifferential and variational techniques in optimization, Berlin: Springer, 1982, 39–159. https://doi.org/10.1007/BFb0120965
    [3] S. Dafermos, Traffic equilibrium and variational inequalities, Transport. Sci., 14 (1980), 42–54. https://doi.org/10.1287/trsc.14.1.42 doi: 10.1287/trsc.14.1.42
    [4] E. Gorbunov, N. Loizou, G. Gidel, Extragradient method: O (1/k) last-iterate convergence for monotone variational inequalities and connections with cocoercivity, Proceedings of The 25th International Conference on Artificial Intelligence and Statistics, 2022,366–402.
    [5] G. Gu, J. Yang, Tight ergodic sublinear convergence rate of the relaxed proximal point algorithm for monotone variational inequalities, J. Optim. Theory Appl., in press. https://doi.org/10.1007/s10957-022-02058-3
    [6] B. He, W. Xu, H. Yang, X. Yuan, Solving over-production and supply-guarantee problems in economic equilibria, Netw. Spat. Econ., 11 (2011), 127–138. https://doi.org/10.1007/s11067-008-9095-2 doi: 10.1007/s11067-008-9095-2
    [7] Y. Gao, W. Zhang, An alternative extrapolation scheme of PDHGM for saddle point problem with nonlinear function, Comput. Optim. Appl., 85 (2023), 263–291. https://doi.org/10.1007/s10589-023-00453-8 doi: 10.1007/s10589-023-00453-8
    [8] X. Zhao, J. Yao, Y. Yao, A nonmonotone gradient method for constrained multiobjective optimization problems, J. Nonlinear Var. Anal., 6 (2022), 693–706. https://doi.org/10.23952/jnva.6.2022.6.07 doi: 10.23952/jnva.6.2022.6.07
    [9] S. Bubeck, Convex optimization: algorithms and complexity, Found. Trends Mach. Le., 8 (2015), 231–357. https://doi.org/10.1561/2200000050 doi: 10.1561/2200000050
    [10] T. Chan, S. Esedoglu, F. Park, A. Yip, Total variation image restoration: overview and recent developments, In: Handbook of mathematical models in computer vision, Boston: Springer, 2006, 17–31. https://doi.org/10.1007/0-387-28831-7_2
    [11] G. Kutyniok, Theory and applications of compressed sensing, GAMM-Mitteilungen, 36 (2013), 79–101. https://doi.org/abs/10.1002/gamm.201310005 doi: 10.1002/gamm.201310005
    [12] E. Ryu, W. Yin, Large-scale convex optimization: algorithms & analyses via monotone operators, Cambridge: Cambridge University Press, 2022. https://doi.org/10.1017/9781009160865
    [13] N. Parikh, S. Boyd, Proximal algorithms, Foundations and Trends® in Optimization, 1 (2014), 127–239. https://doi.org/10.1561/2400000003 doi: 10.1561/2400000003
    [14] M. Hestenes, Multiplier and gradient methods, J. Optim. Theory Appl., 4 (1969), 303–320. https://doi.org/10.1007/BF00927673 doi: 10.1007/BF00927673
    [15] M. Powell, A method for nonlinear constraints in minimization problems, In: Optimization, New York: Academic Press, 1969,283–298.
    [16] B. He, H. Yang, S. Wang, Alternating direction method with self-adaptive penalty parameters for monotone variational inequalities, J. Optim. Theory Appl., 106 (2000), 337–356. https://doi.org/10.1023/A:1004603514434 doi: 10.1023/A:1004603514434
    [17] G. Qian, D. Han, L. Xu, H. Yang, Solving nonadditive traffic assignment problems: a self-adaptive projection-auxiliary problem method for variational inequalities, J. Ind. Manag. Optim., 9 (2013), 255–274. https://doi.org/10.3934/jimo.2013.9.255 doi: 10.3934/jimo.2013.9.255
    [18] X. Fu, B. He, Self-adaptive projection-based prediction-correction method for constrained variational inequalities, Front. Math. China, 5 (2010), 3–21. https://doi.org/10.1007/s11464-009-0045-1 doi: 10.1007/s11464-009-0045-1
    [19] X. Fu, A general self-adaptive relaxed-PPA method for convex programming with linear constraints, Abstr. Appl. Anal., 2013 (2013), 1–13. https://doi.org/10.1155/2013/492305 doi: 10.1155/2013/492305
    [20] F. Ma, M. Ni, W. Tong, X. Wu, Matrix completion via extended linearized augmented Lagrangian method of multipliers, Proceedings of International Conference on Informative and Cybernetics for Computational Social Systems, 2015, 45–49. https://doi.org/10.1109/ICCSS.2015.7281147 doi: 10.1109/ICCSS.2015.7281147
    [21] Y. Shen, H. Wang, New augmented Lagrangian-based proximal point algorithm for convex optimization with equality constraints, J. Optim. Theory Appl., 171 (2016), 251–261. https://doi.org/10.1007/s10957-016-0991-1 doi: 10.1007/s10957-016-0991-1
    [22] B. He, L. Liao, X. Wang, Proximal-like contraction methods for monotone variational inequalities in a unified framework Ⅱ: general methods and numerical experiments, Comput. Optim. Appl., 51 (2012), 681–708. https://doi.org/10.1007/s10589-010-9373-z doi: 10.1007/s10589-010-9373-z
    [23] B. He, F. Ma, X. Yuan, Optimal proximal augmented Lagrangian method and its application to full jacobian splitting for multi-block separable convex minimization problems, IMA J. Numer. Anal., 40 (2020), 1188–1216. https://doi.org/10.1093/imanum/dry092 doi: 10.1093/imanum/dry092
    [24] G. Gu, B. He, X. Yuan, Customized proximal point algorithms for linearly constrained convex minimization and saddle-point problems: a unified approach, Comput. Optim. Appl., 59 (2014), 135–161. https://doi.org/10.1007/s10589-013-9616-x doi: 10.1007/s10589-013-9616-x
    [25] B. He, X. Yuan, W. Zhang, A customized proximal point algorithm for convex minimization with linear constraints, Comput. Optim. Appl., 56 (2013), 559–572. https://doi.org/10.1007/s10589-013-9564-5 doi: 10.1007/s10589-013-9564-5
    [26] F. Ma, On relaxation of some customized proximal point algorithms for convex minimization: from variational inequality perspective, Comput. Optim. Appl., 73 (2019), 871–901. https://doi.org/10.1007/s10589-019-00091-z doi: 10.1007/s10589-019-00091-z
    [27] F. Ma, M. Ni, A class of customized proximal point algorithms for linearly constrained convex optimization, Comput. Appl. Math., 37 (2018), 896–911. https://doi.org/10.1007/s40314-016-0371-3 doi: 10.1007/s40314-016-0371-3
    [28] B. He, F. Ma, X. Yuan, Optimally linearizing the alternating direction method of multipliers for convex programming, Comput. Optim. Appl., 75 (2020), 361–388. https://doi.org/10.1007/s10589-019-00152-3 doi: 10.1007/s10589-019-00152-3
    [29] J. Liu, J. Chen, J. Zheng, X. Zhang, Z. Wan, A new accelerated positive-indefinite proximal ADMM for constrained separable convex optimization problems, J. Nonlinear Var. Anal., 6 (2022), 707–723. https://doi.org/10.23952/jnva.6.2022.6.08 doi: 10.23952/jnva.6.2022.6.08
    [30] B. He, F. Ma, S. Xu, X. Yuan, A generalized primal-dual algorithm with improved convergence condition for saddle point problems, SIAM J. Imaging Sci., 15 (2022), 1157–1183. https://doi.org/10.1137/21M1453463 doi: 10.1137/21M1453463
    [31] F. Ma, Y. Bi, B. Gao, A prediction-correction-based primal-dual hybrid gradient method for linearly constrained convex minimization, Numer. Algor., 82 (2019), 641–662 https://doi.org/10.1007/s11075-018-0618-8 doi: 10.1007/s11075-018-0618-8
    [32] S. Xu, A dual-primal balanced augmented Lagrangian method for linearly constrained convex programming, J. Appl. Math. Comput., 69 (2023), 1015–1035. https://doi.org/10.1007/s12190-022-01779-y doi: 10.1007/s12190-022-01779-y
    [33] E. Blum, W. Oettli, Mathematische optimierung: grundlagen und verfahren, Berlin: Springer, 1975.
    [34] P. Mahey, A. Lenoir, A survey on operator splitting and decomposition of convex programs, RAIRO-Oper. Res., 51 (2017), 17–41. https://doi.org/10.1051/ro/2015065 doi: 10.1051/ro/2015065
    [35] B. He, A uniform framework of contraction methods for convex optimization and monotone variational inequality, Scientia Sinica Mathematica, 48 (2018), 255. https://doi.org/10.1360/N012017-00034 doi: 10.1360/N012017-00034
    [36] J. Yang, X. Yuan, Linearized augmented Lagrangian and alternating direction methods for nuclear norm minimization, Math. Comput., 82 (2013), 301–329. https://doi.org/10.1090/S0025-5718-2012-02598-1 doi: 10.1090/S0025-5718-2012-02598-1
    [37] S. Boyd, N. Parikh, E. Chu, B. Peleato, J. Eckstein, Distributed optimization and statistical learning via the alternating direction method of multipliers, Found. Trends Mach. Le., 3 (2011), 1–122. https://doi.org/10.1561/2200000016 doi: 10.1561/2200000016
    [38] G. Ye, Y. Chen, X. Xie, Efficient variable selection in support vector machines via the alternating direction method of multipliers, Proceedings of the 14th International Conference on Artificial Intelligence and Statistics, 2011,832–840.
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1566) PDF downloads(64) Cited by(0)

Figures and Tables

Figures(3)  /  Tables(6)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog