1.
Introduction
The main research of this article is about the generalized absolute value equation (GAVE):
where A,B∈Rn×n,b∈Rn and |x|:=(|x1|,|x2|,⋯,|xn|)T. Actually, when B is a identity matrix, the GAVE (1.1) can be easily transformed into the absolute value equation (AVE):
The GAVE (1.1) was first proposed by Rohn as a class of nonlinear nondifferentiable optimization problems [1]. After that, the AVE (1.2) and GAVE (1.1) has applications in many areas of optimization, including linear complementarity problem, bimatrix games, constrained least squares problems, and so on, see [1,2,3,4,5,6]. For instance, the famous linear complementarity problem (LCP) [5]: Determining z such that
In [3], Mangasarian has proved that the GAVE (1.1) and LCP (1.3) can be transformed into each other under some conditions, and he also proposed the unique solvability of the AVE. Moreover, many scholars have further studied the solvability and unique solution theory of the AVE (1.2) [7,8,9]. In the early algorithm research, some efficient iteration methods have been proposed to address the GAVE (1.1) and AVE (1.2), which can be found in Mangasarian [2,10,11], Rohn[12], and so on. Then in [13], in order to overcome the difficulty of constructing the gradient of the absolute value term, Mangasarian presented the generalized Newton (GN) algorithm through using the generalized Jacobian matrix. This method may be summed up as follows:
where D(x):=diag(sgn(x)), x∈Rn, sgn(x) denotes a vector with components equal to 1, 0, −1 depending on whether the corresponding components of x is positive, zero or negative. Moreover, the GN algorithm is also able to apply on the GAVE (1.1) if we use BD(x(k)) instead of D(x(k)). Based on the GN method, Li[14] proposed a modified generalized Newton method. It should be noticed that the Jacobian matrix is changed with respect to the iteration index k, which will increase the amount of computation greatly. In order to stay away from the changed Jacobian matrix, many literature on the result of the AVE (1.2) and GAVE (1.1) are proposed. For instance, Rohn et al. presented the Picard method when A is invertible in [15]:
In [16], Wang et al. added a new term Ωx (Ω is positive semi-definite and it can be guaranteed that A+Ω is invertible) into the general problem and noticed that Ax+Ωx is differential but Ωx+B|x|+b non-differential, then studied a modified Newton-type(MN) iteration method:
To further speed up the AOR method, Li et al.[17] presented the generalization version algorithm. Actually, the above algorithms all belong to Newton-based matrix splitting methods and their relaxed versions, which is proposed in [18]. Sometimes the coefficient matrix A has some special properties, for example, A is M-matrix, under which condition Ali et al. [19] proposed two new generalized Gauss-Seidel iteration methods. In [20], a new iteration method was presented by redesigning equivalently the AVE in (1.2) as a 2×2 block nonlinear equation. Also using matrix blocking, the block-diagonal and anti-block-diagonal splitting (BAS) method[21] and modified BAS method[22] are proposed.
Recently, in order to continue to improve the efficiency of solving the AVE (1.2), some inexact iteration algorithms have been studied. Every step in iteration of the Picard method can be viewed as solving a linear system with the coefficient matrix A. Therefore, Salkuyeh[23] presented the Picard-HSS method for AVE (1.2), which used the HSS method [24] to estimate the result x(k+1) at each Picard iteration. On this basis, Miao et al. [25] used single-step HSS (SHSS) method [26] to address the linear system in Picard iteration and proposed the Picard-SHSS method, a new Picard-type method. Inspired by these, we adopt the modified Newton-type iteration method with faster convergence speed as the outer iterative method, HSS and SHSS as the inner iterative method respectively, then we obtain two new inexact iterative algorithms, abbreviated as MN-HSS method and MN-SHSS method.
The remainder of this article is laid out as follows. In Section 2, the MN-HSS method is described, and we investigate its convergence property. Section 3 is devoted to introducing the MN-SHSS method and analyzing its convergence. Section 4 contains several numerical experiments, showing the feasibility and efficiency of the two new inexact iteration methods. Section 5 draws some conclusions.
2.
The MN-HSS method for solving GAVE
At the beginning of this section, some notations are given blow. ones(n,1) represents a vector like (1,1,⋯,1)T∈R. ρ(A) indicates the spectral radius of A. The symbol (⋅)T represents the transepose of a vector or matrix. And we have limk→+∞Ak=0 if and only if ρ(A)<1 (see[27,28]).
Given A∈Rn×n be a non-Hermitian positive definite matrix. Then H=12(A+AT)S=12(A−AT), which are the Hermitian part and skew-Hermitian part of A, A=H+S. Then the HSS iteration method was presented by Bai et al.[24] for addressing positive definite system Ax=b. The iterative process of the HSS method can be expressed as follows:
where α is a positive constant.
Section 1 has introduced the MN method, which is equivalent to the following form
Similar to the Picard-HSS method, we can choose a suitable Ω to make sure the positive definiteness of A+Ω. The HSS method (2.1) can be applied as the inner iteration method of the MN method for solving the GAVE (1.1), called MN-HSS method.
Algorithm (The MN-HSS iteration method):
For k=0,1,..., until converge, do:
Set x(k,0):=x(k)
For l=0,1,..., until converge, do:
where H=12((A+Ω)+(A+Ω)T), S=12((A+Ω)−(A+Ω)T), α is a positive constant.
Endfor
Set x(k+1):=x(k,lk).
Endfor
The following theorem proves that the MN-HSS algorithm converges under necessary conditions.
Theorem 2.1. Let the matrix A,B∈Rn×n and choose suitable Ω∈Rn×n such that A+Ω is a positive definite matrix, then H=12((A+Ω)+(A+Ω)T), S=12((A+Ω)−(A+Ω)T) are the Hermitian and skew-Hermitian parts of A+Ω respectively. Let also ∥A−1B∥2 <1 and ∥(A+Ω)−1∥2<1∥B∥2+∥Ω∥2. Then the GAVE (1.1) has a unique solution x∗, and for any starting vector x(0)∈Rn and any sequence of positive integers lk, k=0,1,..., the iteration sequence {x(k)}∞k=0 produced by the MN-HSS iteration method converges to x∗ provided that l=lim infk→∞lk≥N, where N is a natural number satisfying
where Tα=M−1αNα, Mα=12α(αI+H)(αI+S), Nα=12α(αI−H)(αI−S).
Proof. Let Gα=M−1α, then according to the iteration formula (2.3), we can write the (k+1)th iterate x(k+1) of the MN-HSS iteration method as
Since ∥A−1B∥2<1, the GAVE (1.1) has a unique result x∗ [3], which satisfying
By subtracting (2.5) from (2.4), we have
It follows from [24] that ρ(Tα)<1, then
Then the Eq (2.6) turns into
It is obvious that for any x,y∈Rn, ∥|x|−|y|∥2≤∥x−y∥2. Therefore,
On the other hand, since ρ(Tα)<1, lims→∞Tsα=0. Therefore, there is a natural number N such that
Thus, assume that l=lim infk→∞lk≥N, then we can obtain the result. This completes the proof.
The MN-HSS method can also be expressed as the residual-updating form:
Algorithm (The residual-updating variant of MN-HSS iteration method):
For k=0,1,..., until converge, do:
Set d(k,0)=0 and b(k)=B|x(k)|+b−Ax(k)
For l=0,1,..., until converge, do:
where H=12((A+Ω)+(A+Ω)T), S=12((A+Ω)−(A+Ω)T), α is a positive constant.
Endfor
Set x(k+1)=x(k)+d(k,lk).
Endfor
3.
The MN-SHSS method for solving GAVE
From the iteration scheme (2.3) we can notice that the each step of HSS iteration method is equivalent to solving two linear subsystems, whose coefficient matrices are Hermitian and skew-Hermitian respectively. The first equation with Hermitian coefficient matrix can be solved quickly by conjugate gradient method or Cholesky factorization. However, the other subsystem needs much more computation. So as to accelerate the algorithm, we replace the inner iteration method with single-step HSS (SHSS) method, then we present a new inexact method in this section, named as MN-SHSS iteration method.
Compared with the HSS method, the SHSS method ignores the second subsystem with skew-Hermitian matrix, which can be expressed as
by limiting the iteration parameter α, Li et al.[26] has proved that the SHSS converges to the unique solution for any initial guess x(0).
Algorithm (The MN-SHSS iteration method):
For k=0,1,..., until converge, do:
Set x(k,0):=x(k)
For l=0,1,..., until converge, do:
where H=12((A+Ω)+(A+Ω)T), S=12((A+Ω)−(A+Ω)T), α is a positive constant.
Endfor
Set x(k+1):=x(k,lk).
Endfor
The next theorem shows that the MN-SHSS method is convergent under a restriction on the parameter α.
Theorem 3.1. Let the matrix A,B∈Rn×n and choose suitable Ω∈Rn×n such that A+Ω is a positive definite matrix, then H=12((A+Ω)+(A+Ω)T), S=12((A+Ω)−(A+Ω)T) are the Hermitian and skew-Hermitian parts of A+Ω respectively. Let also ∥A−1B∥2 <1, ∥(A+Ω)−1∥2<1∥B∥2+∥Ω∥2, and α be a constant number such that α>max{0,δ2max−λ2min2λmin}, where δmax is the largest singular value of S and λmin is the smallest eigenvalue of H. Then the GAVE (1.1) has a unique solution x∗, and for any starting vector x(0)∈Rn and any sequence of positive integers lk, k=0,1,..., the iteration sequence {x(k)}∞k=0 produced by the MN-SHSS iteration method converges to x∗ provided that l=lim infk→∞lk≥N, where N is a natural number satisfying
where Kα=M−1αNα, Mα=αI+H, Nα=αI−S.
Proof. According to the iteration formula (3.2), we can get the kth iteration x(k+1) of the MN-SHSS method:
On the other hand, the unique result x∗ of GAVE (1.1) satisfies
Subtracting (3.4) from (3.3), then we have
In [26], Li et al. has proved that ρ(Kα)<1 when α>max{0,δ2max−λ2min2λmin}, hence lk−1∑j=0KjαM−1α=(I−Klkα)(A+Ω)−1. Then (3.5) becomes
Therefore, we can calculate that
Due to the condition of ρ(Kα)<1, we can get that Ksα→0 as s→∞. So there is a natural number N such that
Then if l=lim infk→∞lk≥N, it is distinct that ∥x(k+1)−x∗∥2<∥x(k)−x∗∥2. That is to say, the iteration sequence {x(k)}∞k=0converges to x∗, which is produced by the MN-SHSS iteration method. This completes the proof.
In the same way, the MN-SHSS method also has residual form:
Algorithm (residual-updating variant of the MN-SHSS iteration method):
For k=0,1,..., until converge, do:
Set d(k,0)=0 and b(k)=B|x(k)|+b−Ax(k)
For l=0,1,..., until converge, do:
where H=12((A+Ω)+(A+Ω)T), S=12((A+Ω)−(A+Ω)T), α is a positive constant.
Endfor
Set x(k+1)=x(k)+d(k,lk).
Endfor
4.
Numerical experiments
In this section, the effectiveness of MN-HSS method and MN-SHSS method is showed by comparing them with other existing inexact iteration method for solving GAVE (1.2): the Picard-HSS method and the Picard-SHSS method. In addition, we also compare our methods with a descent method [29]. For this purpose, we compare from the following aspects: the number of outer iteration steps (denoted by 'IT'), the elapsed CPU time (denoted by 'CPU') and the residual error(denoted by 'RES'), where the 'RES' is set to be
In our computations, we all use the residual-updating version and optimal parameters α resulting in the least iteration numbers of the algorithms mentioned above and the zero vector as our starting iteration vector. We use the Cholesky factorization to solve the equations whose coefficient matrix is αI+H and the LU factorization for the other subsystem. All runs are aborted once the residual error satisfies RES<=10−7 or the maximum iteration number kmax=500 is exceeded. As for the inner iterations, we set the stopping criterion of different methods as:
and a limit on the number of iterations 10(lk=10, k=1,2,...) for inner iterations are used. The parameters used in the desent method are chosen as σ=0.2, δ=0.8, γ=0.001, p=3 and ϵ0=0.01.
The two numerical experiments below are performed in MATLAB R2020b.
Example 1.[30] The first example comes from the LCP (1.3). In [3,30], it has been known that the LCP (1.3) can be transformed into the following form:
which belongs to the GAVE (1.1) obviously. Then we let A=M+I and B=M−I, where M=ˆM+μI and the right-hand vector q is determined by q=−Mz∗, in which
is a block tridiagonal matrix,
is a tridiagonal matrix, n=m2. we can know that z∗=1.2∗ones(n,1) is the unique result of the LCP(1.3) and the exact solution of the GAVE(4.1) is x∗=0.6∗ones(n,1).
In this experiment, we pick three scenarios for the parameter μ=4,−1,−4. The matrices A and M are both symmetric positive definite when μ=4, and symmetric indefinite when μ=−4. The matrix A is symmetric positive definite if μ=−1, but M symmetric indefinite. We choose four different sizes of n (n=3600,4900,6400,8100) for each parameter μ. Considering the simplicity and efficiency of our experiments, we use the same matrix Ω=5I in the MN-HSS and MN-SHSS methods. In the methods we test, the value α is chosen to result in the fewest iteration steps, see Table 1. In Tables 2–4, we list the numerical outcomes of the studied four inexact iteration methods for μ=4,−1,−4. In these tables, "-" indicates that the relevant algorithm is not convergent to the answer under the limitation of the maximum number kmax of iterative steps or even diverge.
From Table 2, we can see that when μ=4, for all matrix scales the five evaluated methods can successfully create an estimated solution to the GAVE (4.1). And for each tested method, the CPU time also increases with the increase of the scale n of the coefficient matrix. Through the study of the numerical results, we can know that the two inexact iteration methods proposed by us are better than the Picard-HSS and Picard-SHSS methods in the number of iteration steps and CPU time. Although the descent method need less iteration steps, our methods spend less CPU time. When μ=−1,−4, in which instance the convergence of Picard-HSS and Piard-SHSS iteration methods cannot be guaranteed, hence Tables 3 and 4 show that the two Picard-type methods converge slowly or do not converge, while the MN-HSS and MN-SHSS method can still converge fast. The selectivity of matrix Ω provides higher stability for the algorithm we propose. When μ=−1, our methods still have better performance in elapsed CPU time than the descent method generally, and the descent method can not converge under constraints of maximum iteration number.
Example 2. The coefficient matrix A of the GAVE (1.1) is from the finite difference approximation the two-dimensional convection-diffusion equation
where S=(0,1)×(0,1), ∂S is its boundary, q is a constant and p is a real number. For discretization, the five-point finite difference scheme are used on the diffusive terms and the central difference scheme on the convective terms respectively, then we can obtain the coefficient matrix A.
where Im and In are the identity matrices of order m and n with n=m2, ⊗ means the Kronecker product, Tx=tridiag(a2,a1,a3)m×m and Ty=tridiag(a2,0,a3)m×m with a1=4, a2=−1−Re, a3=−1+Re. Here Re=qh2 and h=1m+1 are the mesh Reynolds number and the equidistant step size, respectively. q is a positive constant. Let the coefficient matrix B=I and the exact solution x∗=(−1,2,−3,⋯,(−1)ii,⋯,(−1)nn)T, then the right-hand side vector can also be determined. In this instance the GAVE becomes a AVE (1.2) actually.
In our experiments, we test different values of p (p=0,−1), q (q=0,1) and n (n=100,400,1600,6400). The parameters α of the four tested methods resulting in the fewest iteration steps are chosen, while the matrix Ω in our methods are chosen as Ω=ωI, see Tables 5 and 6. We refer to [25] for the parameters of Picard-HSS and Picard-SHSS methods.
From Tables 7 and 8, the iteration steps reveal that the Picard-type methods and the descent method converges faster than the MN-HSS method, but in terms of the elapsed CPU times, the MN-SHSS method we new proposed has better computing efficiency generally. In Tables 9 and 10, that is when p=−1, the iteration counts and CPU time of the Picard-type methods increase dramatically with the increase of matrix scale, which situation is more obvious on the descent method. In contrast, our methods still maintain a low iteration counts and high computational efficiency. Therefore, our new proposed method is very effective.
5.
Conclusions
In this article, to accelerate the MN iteration method, we propose the MN-HSS and MN-SHSS iteration algorithms for addressing the generalized absolute value equation. Moreover, we give the sufficient conditions to show that our methods are convergent for addressing the GAVE. In the end, we provide some experiments to comfirm the feasibility and effectiveness of our novel presented methods.
Acknowledgments
This work is supported by the National Natural Science Foundation of China (Grant no. 1227010254, Research on some efficient and fast algorithms for complex nonlinear problems).
Conflict of interest
The authors declare that they have no conflicts of interest.