Research article Special Issues

The development of new efficient iterative methods for the solution of absolute value equations

  • The use of absolute value equations (AVEs) is widespread across a wide range of fields, including scientific computing, management science, and engineering. Our aim in this study is to introduce two new methods for solving AVEs and to explore their convergence characteristics. Furthermore, numerical experiments will be carried out to demonstrate their feasibility, robustness, and efficacy.

    Citation: Rashid Ali, Fuad A. Awwad, Emad A. A. Ismail. The development of new efficient iterative methods for the solution of absolute value equations[J]. AIMS Mathematics, 2024, 9(8): 22565-22577. doi: 10.3934/math.20241098

    Related Papers:

    [1] Shu-Xin Miao, Xiang-Tuan Xiong, Jin Wen . On Picard-SHSS iteration method for absolute value equation. AIMS Mathematics, 2021, 6(2): 1743-1753. doi: 10.3934/math.2021104
    [2] Xin-Hui Shao, Wan-Chen Zhao . Relaxed modified Newton-based iteration method for generalized absolute value equations. AIMS Mathematics, 2023, 8(2): 4714-4725. doi: 10.3934/math.2023233
    [3] Wan-Chen Zhao, Xin-Hui Shao . New matrix splitting iteration method for generalized absolute value equations. AIMS Mathematics, 2023, 8(5): 10558-10578. doi: 10.3934/math.2023536
    [4] Cui-Xia Li, Long-Quan Yong . Modified BAS iteration method for absolute value equation. AIMS Mathematics, 2022, 7(1): 606-616. doi: 10.3934/math.2022038
    [5] Rashid Ali, Ilyas Khan, Asad Ali, Abdullah Mohamed . Two new generalized iteration methods for solving absolute value equations using $ M $-matrix. AIMS Mathematics, 2022, 7(5): 8176-8187. doi: 10.3934/math.2022455
    [6] Miao Guo, Qingbiao Wu . Two effective inexact iteration methods for solving the generalized absolute value equations. AIMS Mathematics, 2022, 7(10): 18675-18689. doi: 10.3934/math.20221027
    [7] ShiLiang Wu, CuiXia Li . A special shift splitting iteration method for absolute value equation. AIMS Mathematics, 2020, 5(5): 5171-5183. doi: 10.3934/math.2020332
    [8] Yang Cao, Quan Shi, Sen-Lai Zhu . A relaxed generalized Newton iteration method for generalized absolute value equations. AIMS Mathematics, 2021, 6(2): 1258-1275. doi: 10.3934/math.2021078
    [9] Mudassir Shams, Nasreen Kausar, Serkan Araci, Liang Kong . On the stability analysis of numerical schemes for solving non-linear polynomials arises in engineering problems. AIMS Mathematics, 2024, 9(4): 8885-8903. doi: 10.3934/math.2024433
    [10] Zui-Cha Deng, Fan-Li Liu, Liu Yang . Numerical simulations for initial value inversion problem in a two-dimensional degenerate parabolic equation. AIMS Mathematics, 2021, 6(4): 3080-3104. doi: 10.3934/math.2021187
  • The use of absolute value equations (AVEs) is widespread across a wide range of fields, including scientific computing, management science, and engineering. Our aim in this study is to introduce two new methods for solving AVEs and to explore their convergence characteristics. Furthermore, numerical experiments will be carried out to demonstrate their feasibility, robustness, and efficacy.



    This article concentrates on the AVE, involving a square matrix A in Rn×n, vectors y as well as b in Rn, and || signifying the absolute value. The equation is as follows:

    Ay|y|=b. (1.1)

    The formulation of the AVE provided by Eq (1.1) is a standard expression of a more general framework that can be summarized as follows:

    Ay+B|y|=b, (1.2)

    with A and B are matrices of dimensions n×n. Therefore, the general form (1.2) can be simplified to (1.1) when B=I, where I is the identity matrix. The general concept of AVE was first introduced in [1] and has since been extensively explored and examined in the literature discussed in [2,3,4] and in the references mentioned therein. AVEs are crucial nonlinear and nondifferentiable systems that are commonly found in optimization landscapes. This type of system manifests in a variety of domains, such as contact problems, bimatrix games, quadratic programming, network pricing, and linear programming; refer to [1,5,6,7,8,9] and associate literature for more knowledge. Hence, the investigation of robust numerical procedures and theoretical frameworks for AVEs has great scientific significance, as well as the potential for wide-ranging applications and significant economic benefits.

    The applications of numerical techniques to AVEs cover a wide range of structural considerations, mathematical frameworks, algebraic configurations, and innovative enactments of high-quality preconditioners alongside high-efficiency numerical strategies. In recent times, there has been a considerable surge in the exploration of numerical methods for AVEs, with numerous scholarly works suggesting diverse methodologies. For example, Yilmaz and Sahiner [10] produced a non-Lipschitz generalization of AVE and determined it utilizing two smoothing strategies. Ali [11] successively presented two fixed-point iterative strategies for AVE (1.1) and examined various kinds of convergence theorems. Zhou et al. [12] offered a Newton-based matrix splitting strategy that is capable of acquiring a linear convergence when A is an H-matrix or a positive definite matrix. Prokopyev [4] has shed light on the unique solvability of AVE, as well as its intricate ties with mixed integer programming and linear complementarity problems (LCP). Meanwhile, Hu and Huang [6] incorporated the AVE framework into the standard LCP format, providing insights into solving AVE (1.1). Salkuyeh [13] introduced the Picard-HSS iterative technique for the solution of AVEs and examined its convergence conditions. Khan et al. [14] have proposed an innovative method for solving AVEs based on Simpson's rule and generalized Newton's method. Noor et al. [15] have showcased minimization processes for Eq (1.1) and examined the convergence of these procedures under some appropriate states. Ke and Ma [16] provided an SOR (successive overrelaxation)-like method for Eq (1.1) and studied its application conditions, while Dong et al. [17] conducted an in-depth investigation of an SOR-like method to solve AVEs, exploring its convergence conditions, which differ from those presented by [16]. Tang [18] offered an innovative, inexact Newton-type method designed to deal with large-scale AVEs and outlined a detailed analysis of its convergence characteristics. Rohn et al. [19] presented another method for addressing AVE (1.1) that effectively reduces to the well-known Picard iteration method. The method is outlined below:

    yi+1=A1(|yi|+b),i=0,1,2,...,

    where y0=A1b is the initial vector. Tang and Zhou [20] have demonstrated in their study a quadratically convergent descent method to solve the AVE (1.2) problem and have discussed different properties of convergence. Mangasarian and Meyer [7] provided an important and widely used theoretical result concerning the solvability of the AVE (1.1). Their findings stated that if A1<1 (or equivalently, σmin(A)>1, where σmin(A) represents the smallest singular value of A), then a unique solution y exists to the AVE (1.1) for any bRn. Furthermore, Mangasarian [3] offered an approximated generalized Newton approach tailored for addressing the AVE as prescribed in Eq (1.2). The findings display that this approach achieves linear convergence from any initial guess, provided it reaches the unique solution of AVE (1.2) under the circumstances A1<14. Caccetta et al. [21] explored a smoothing Newton procedure designed for addressing the AVE as defined in Eq (1.2). The authors have demonstrated that this method is not only global but also has quadratic convergence, provided that the weak condition A1<1 is satisfied. In a related study, Zhang and Wei [22] developed a generalized Newton method to address the same AVE (1.2). The authors demonstrate that their method is capable of achieving both global convergence as well as finite convergence when [AI,A+I] is regular, a scenario that includes the case where A1<1. Wang et al. [23], Lian et al. [24], Cao et al. [25], Zhou et al. [26], Lv and Miao [27,28], Zhang and Miao [29], and Wu and Li [30] also investigated various approaches for AVEs and presented some fascinating convergence results.

    The aims of this paper is to provide new iterative techniques for the computation of AVEs, supported by an extensive theoretical analysis. Our contributions are as follows: Firstly, we enhance the generalized Gauss-Seidel method [31] by splitting coefficient matrix A into three parts shown by Eq (2.2) and adding an extra parameter ψ to speed up convergence speed. Subsequently, we conduct thorough examination of their properties under specific conditions to ensure they work effectively.

    The structure of this paper unfolds as follows: In Section 2, we unveil the offered methods and their convergence for addressing AVE (1.1). Moving forward, Section 3 encapsulates our numerical findings, while Section 4 contains our concluding remarks.

    We will use the following notation throughout this article. Let A=(aij)Rn×n represent a matrix. We use |A|=(|aij|) to denote its absolute value and ||A|| for its infinity norm. A matrix ARn×n is termed a Z-matrix if aij0 for ij, and it is called an M-matrix if it's a non-singular Z-matrix and satisfies A10.

    Lemma 1.1. [32] Suppose we have two vectors y and u, both belonging to Rn×n. Then |y||u|yu.

    Here, we explore the proposed methods. Both strategies are extensions of the generalized Gauss-Seidel method (GGSM). We will refer to these new methods as Extended GGSM Ⅰ (EGGSM Ⅰ) and Extended GGSM Ⅱ (EGGSM Ⅱ). There are two sections in this part. Section 2.1 examines EGGSM I and its convergence, while Section 2.2 examines EGGSM II and its convergence.

    By redefining the AVE (1.1), we obtain

    Ay|y|=b.

    If we multiply both sides by ψ, we obtain,

    ψAyψ|y|=ψb. (2.1)

    Let

    A=ADALAU, (2.2)

    where AD, AL, and AU represent the diagonal, strictly lower triangular, and upper triangular parts of matrix A, respectively. By employing Eqs (2.1) and (2.2), we propose the EGGSM I as follows:

    (ADψAL)yψ|y|=((1ψ)AD+ψAU)y+ψb. (2.3)

    Through the iterative process, the previously stated equations can be reformulated as

    (ADψAL)yi+1ψ|yi+1|=((1ψ)AD+ψAU)yi+ψb, (2.4)

    where i=0,1,2,..., and ψ(0,1.5]. In addition, the EGGSM I algorithm is presented below:

    (1) Choose a parameter ψ, pick an initial vector y0Rn, and then assign i=0.

    (2) Calculate ri=yi+1A1(|yi|+b),

    (3) Calculate

    yi+1=ψ(ADψAL)1|ri|+(ADψAL)1[((1ψ)AD+ψAU)yi+ψb].

    (4) If yi+1=yi, then stop. If not, set i=i+1 and return to Step 2.

    Now, the subsequent theorem demonstrates the convergence of EGGSM I.

    Theorem 2.1. Assume that the problem denoted as AVE (1.1) is solvable. Suppose the diagonal elements of matrix A exceed one and meet the subsequent condition:

    (ADψAL)1((1ψ)AD+ψAU)<1ψ(ADψAL)1, (2.5)

    then the sequence {yi} of EGGSM I converges to the unique solution y of AVE.

    Proof. Initially, we demonstrate that (ADψAL)1<1. Obviously, if AL=0, then (ADψAL)1=A1D<1. Now, if we suppose that AL0, then we proceed as follows:

    0|ψAL|w<(ADI)w.

    If we consider

    |ψAL|w<(ADI)w.

    When we consider both perspectives using the inverse of matrix AD, we obtain

    A1D|ψAL|w<A1D(ADI)w,|ψA1DAL|w<(IA1D)w,|ψA1DAL|w<wA1Dw,A1Dw<w|ψA1DAL|w,A1Dw<(I||)w, (2.6)

    where =ψA1DAL and w=(1,1,...,1)T. Also, we have

    0|(I)1|=|I++2+3+...+n1|,(I+||+||2+||3+...+||n1)=(I||)1. (2.7)

    So, by utilizing Eqs (2.6) and (2.7), we derive the following:

    |(ADψAL)1|w=|(I)1A1D|w|(I)1||A1D|w,<(I||)1(I||)w=w.

    This implies

    (ADψAL)1<1.

    To verify the uniqueness of the solution, consider two different solutions denoted as y and ϑ for the AVE Eq (1.1). Employing Eq (2.4), we derive

    y=ψ(ADψAL)1|y|+(ADψAL)1[((1ψ)AD+ψAU)y+ψb], (2.8)
    ϑ=ψ(ADψAL)1|ϑ|+(ADψAL)1[((1ψ)AD+ψAU)ϑ+ψb]. (2.9)

    From (2.8) and (2.9), we get

    yϑ=ψ(ADψAL)1(|y||ϑ|)+(ADψAL)1((1ψ)AD+ψAU)(yϑ).

    Applying Lemma 2.1 in conjunction with Eq (2.5), the above expression can be reformulated as

    yϑψ(ADψAL)1|y||ϑ|+(ADψAL)1((1ψ)AD+ψAU)yϑ<ψ(ADψAL)1yϑ+(1ψ(ADψAL)1)yϑ,
    yϑψ(ADψAL)1yϑ<(1ψ(ADψAL)1)yϑ,
    (1ψ(ADψAL)1)yϑ<(1ψ(ADψAL)1)yϑ,
    yϑ<yϑ.

    This implies a contradiction, so we conclude that y=ϑ.

    To ensure convergence, suppose that y represents the unique solution of AVE (1.1). As a result, by considering Eq (2.8) and

    yi+1=ψ(ADψAL)1|yi+1|+(DψAL)1[((1ψ)AD+ψAU)yi+ψb],

    we deduce

    yi+1y=ψ(ADψAL)1(|yi+1||y|)+(ADψAL)1[((1ψ)AD+ψAU)(yiy)].

    By utilizing the infinity norm alongwith Lemma 2.1, we obtain

    yi+1yψ(ADψAL)1yi+1y(ADψAL)1((1ψ)AD+ψAU)yiy,

    and since (ADψAL)1<1, it follows that

    yi+1y(ADψAL)1((1ψ)AD+ψAU)1ψ(ADψAL)1yiy.

    This inequality suggests that the EGGSM I converges when condition (2.5) is fulfilled.

    In this section, we introduce the EGGSM II. Utilizing Eqs (2.1) and (2.2), we establish the framework of EGGSM II to address AVE (1.1) in the following manner:

    (ADψAL)yi+1ψ|yi+1|=((1ψ)AD+ψAU)yi+1+ψb.

    Alternatively, we can write as follows:

    yi+1=ψ(ADψAL)1|yi+1|+(ADψAL)1[((1ψ)AD+ψAU)yi+1+ψb],i=0,1,2,... (2.10)

    Below are the procedural steps of the EGGSM II algorithm:

    (1) Choose a parameter ψ, pick an initial vector y0Rn, and then assign i=0.

    (2) Calculate ri=yi+1A1(|yi|+b).

    (3) Calculate

    yi+1=ψ(ADψAL)1|ri|+(ADψAL)1[((1ψ)AD+ψAU)ri+ψb].

    (4) If yi+1=yi, then stop. If not, set i=i+1 and go to Step 2.

    Now, to ensure the convergence of the EGGSM II, let us examine the following theorem.

    Theorem 2.2. Suppose the diagonal elements of matrix A exceed one and meet the subsequent condition:

    (ADψAL)1((1ψ)AD+ψAU)A1<1ψ(ADψAL)1, (2.11)

    then the sequence {yi} of EGGSM II converges to the unique solution y of AVE.

    Proof. The uniqueness of EGGSM II is derived directly from Theorem 2.1. To ensure convergence, suppose that y represents the unique solution of AVE (1.1). We consider Eq (2.10)

    yi+1=ψ(ADψAL)1|yi+1|+(ADψAL)1[((1ψ)AD+ψAU)yi+1+ψb],

    expressed as

    yi+1=ψ(ADψAL)1|yi+1|+(ADψAL)1[((1ψ)AD+ψAU)[A1(|yi|+b)]+ψb], (2.12)

    where yi+1A1(|yi|+b). Suppose y is the unique solution of AVE (1.1). Then, we obtain

    y=ψ(ADψAL)1|y|+(ADψAL)1[((1ψ)AD+ψAU)[A1(|y|+b)]+ψb]. (2.13)

    Subtracting Eq (2.13) from Eq (2.12) yields we deduce

    yi+1y=ψ(ADψAL)1(|yi+1||y|)+(ADψAL)1[((1ψ)AD+ψAU)A1(yiy)].

    By utilizing the infinity norm alongwith Lemma 2.1, we obtain

    yi+1yψ(ADψAL)1yi+1y(ADψAL)1((1ψ)AD+ψAU)A1yiy,

    and since (ADψAL)1<1, so we have

    yi+1y(ADψAL)1((1ψ)AD+ψAU)A11ψ(ADψAL)1yiy.

    The inequality mentioned above indicates that the EGGSM II reaches convergence when condition (2.11) is satisfied.

    The objective of this section is to showcase several numerical tests. These tests aim to exemplify the significance of new strategies from three stances:

    ● Iteration steps (ITs).

    ● Processing time in seconds (CPU).

    ● Norm of absolute residual vectors (RES).

    Where (RES) is defined as

    RES:=Ayi|yi|b2b2106.

    The computations were performed with an Intel (C) Core (TM) i5-3337U processor, 4 GB of RAM, and MATLAB (2018a). In addition, we consider the value of ψ=1.2 for all examples.

    Example 3.1. [33] Let v denote a predetermined positive integer, and n=v2. Let's delve into the AVE (1.1). Here, we assume A belongs to Rn×n and can be represented as A=M+I, where

    M=(S0.5I1.5IS0.5IS0.5I1.5IS)Rn×n

    is a block-tridiagonal matrix, and

    S=(40.51.540.540.51.54)Rv×v,

    and

    b=Ay|y| with y=(1,2,1,2,...,1,2,...)TRn.

    In Examples 3.1 and 3.2, we consider the initial estimate y=(1,0,1,0,...,1,0,...)TRn and contrast EGGSM I and EGGSM II with various iterative techniques: the AOR iteration method [34], the mixed-type splitting (MS) method [33], the fixed-point method (FM) [35], and the GGS method (GGSM) [31]. The tabulated numerical outcomes are presented in Table 1.

    Table 1.  Numerical findings of Example 3.1.
    Procedures n 100 400 900 1600 4900
    AOR ITs 97 190 336 706 384
    CPU 0.4721 2.8203 3.2174 6.3887 9.2344
    RES 9.81e–07 9.62e–07 9.74e–07 9.85e–07 9.37e–07
    MS ITs 88 157 250 386 342
    CPU 0.4043 1.7955 3.0217 5.7626 8.8967
    RES 8.93e–07 9.65e–07 9.20e–07 9.57e–07 9.88e–07
    FM ITs 42 62 79 94 103
    CPU 0.1821 0.3227 0.9642 1.3403 1.9528
    RES 9.67e–07 9.78e–07 8.65e–07 8.84e–07 8.82e–07
    GGSM ITs 34 52 67 81 93
    CPU 0.1622 0.2910 0.9442 1.0403 1.7526
    RES 9.54e–07 8.41e–07 8.43e–07 8.35e–07 8.27e–07
    EGGSM I ITs 22 32 41 49 53
    CPU 0.1193 0.1547 0.6241 0.9971 1.4911
    RES 5.43e–07 8.92e–07 8.79e–07 9.85e–07 9.07e–07
    EGGSM II ITs 18 27 35 42 47
    CPU 0.0937 0.1024 0.2241 0.7971 1.0285
    RES 6.35e–07 7.98e–07 7.90e–07 9.95e–07 9.97e–07

     | Show Table
    DownLoad: CSV

    Example 3.2. [33] Given a positive integer v, let n=v2. Suppose the expression AVE (1.1), where A is a real n×n matrix represented by A=M+4I, where

    M=(SIISISIIS)Rn×n

    is a block-tridiagonal matrix, and

    S=(411414114)Rv×v,

    and b=Ay|y| with y=((1)h,h=1,2,...,n)TRn. Table 2 lists the actual results.

    Table 2.  Numerical findings of Example 3.2.
    Procedures n 64 256 1024 4096
    AOR ITs 14 14 15 35
    CPU 0.3482 1.9789 2.3872 5.8097
    RES 5.22e–07 6.28e–07 6.53e–07 8.75e–07
    MS ITs 14 14 15 25
    CPU 0.3168 1.0954 1.9649 2.2196
    RES 4.33e–07 5.46e–07 5.08e–07 9.38e–07
    FM ITs 13 13 14 16
    CPU 0.3008 0.5327 1.6769 2.1363
    RES 9.43e–07 4.88e–07 6.33e–07 8.95e–07
    GGSM ITs 11 12 12 12
    CPU 0.2989 0.4201 1.4819 1.9929
    RES 6.39e–07 5.34e–07 7.06e–07 7.01e–07
    EGGSM I ITs 8 8 9 9
    CPU 0.1036 0.3573 1.2211 1.3492
    RES 5.34e–07 7.09e–07 8.47e–07 8.86e–07
    EGGSM II ITs 5 5 5 5
    CPU 0.0977 0.2321 0.7963 1.0837
    RES 6.15e–08 5.91e–08 5.77e–08 5.68e–08

     | Show Table
    DownLoad: CSV

    Tables 1 and 2 of our study provide a thorough comparison of the numerical findings acquired by applying multiple approaches: AOR, MS, FM, GGSM, and novel techniques we have developed. These comparisons are conducted across a variety of values for n. Upon analysis of the numerical data, it is evident that our newly proposed EGGSM I and EGGSM II show superior performance to all other methods. Specifically, they surpass the others in terms of both iterations (ITs) required for convergence and the computational time (CPU) needed to achieve these results.

    Example 3.3. [15] Let A be an n×n matrix with elements defined as follows: aii=4n, ai,i+1=ai+1,i=n, and aij=0.5 for i=1,2,,n. Consider b=(AI)e, where I is the identity matrix of order n and e is an n×1 vector with all elements equal to unity, such that y=(1,1,,1)T represents the exact solution. Specifically, we initialize the vector as y(0)=(y1,y2,,yi,)T, where yi=0.001i. In this scenario, we evaluate our proposed methods against the minimization technique introduced in [15] (referred to as MT), the modified search direction iteration method [36] (denoted as MDM), the fixed-point method (FM) [35], and GGSM [31]. The results are summarized in Table 3.

    Table 3.  Numerical findings of Example 3.3.
    Procedures n 3000 4000 5000 6000 7000
    MT ITs 26 27 27 27 27
    CPU 1.9211 4.5138 18.3728 32.4391 74.2529
    RES 5.74e–07 4.53e–07 7.42e–07 5.69e–07 6.92e–07
    MDM ITs 15 15 15 15 15
    CPU 1.0328 3.5788 8.5019 37.6694 68.4120
    RES 1.98e–07 3.63e–07 7.57e–07 3.69e–07 9.87e–07
    FM ITs 11 11 11 11 11
    CPU 0.8488 2.2201 9.3834 27.5728 31.8227
    RES 4.72e–07 8.86e–07 5.37e–07 7.24e–07 3.28e–07
    GGSM ITs 10 10 10 10 10
    CPU 0.7219 1.9810 8.3028 25.2081 27.1159
    RES 4.31e–07 6.97e–07 7.22e–07 7.98e–07 8.19e–07
    EGGSM Ⅰ ITs 6 6 6 6 6
    CPU 0.5064 1.4947 7.9371 15.7591 17.1890
    RES 4.74e–07 8.87e–7 4.47e–07 6.27e–07 4.91e–07
    EGGSM Ⅱ ITs 2 2 2 2 2
    CPU 0.3927 1.1083 5.4811 7.9627 11.2083
    RES 4.43e–10 3.86e–10 3.37e–10 3.02e–10 2.98e–10

     | Show Table
    DownLoad: CSV

    In Table 3 of our study, we present an analysis of the numerical results obtained using various methodologies: MT, MDM, FM, and GGSM, as well as innovative approaches that have been developed. Various values of n are considered in these assessments. Upon scrutinizing the numerical data, it becomes apparent that our newly introduced EGGSM I and EGGSM II exhibit superior efficacy compared to alternative methods. Moreover, they outperform others in terms of both ITs and CPU required to attain results. As a result, we deduce that the offered approaches are admirably effective and practical for implementation.

    In this study, we have explored two extended versions of the Gauss-Seidel method, known as EGGSM I and EGGSM II to solve AVEs. We meticulously investigated the requisite conditions for the convergence of these novel iterative techniques. In addition, we presented several numerical results demonstrating our approaches' effectiveness. The computational results demonstrate the relevance of the proposed methodologies, particularly when handling large, sparse AVEs, and their considerable superiority over existing approaches. In both theoretical and empirical analyses, our proposed algorithms have been shown to solve AVEs with high efficiency.

    Rashid Ali: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Resources, Supervision and Writing-original draft; Fuad A. Awwad: Software, Visualization, Funding acquisition, Writing-review and editing; Emad A. A. Ismail: Formal Analysis, Project administration, Writing-original draft and Writing-review and editing.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    Researchers supporting project number (RSPD2024R1060), King Saud University, Riyadh, Saudi Arabia. Furthermore, this study was also supported by the Zhejiang Normal University Research Fund under Grant YS304123950, China.

    The authors declare that they do not have any conflicts of interest.



    [1] J. Rohn, A theorem of the alternatives for the equation Ax+B|x|=b, Linear Multilinear A., 52 (2004), 421–426. http://dx.doi.org/10.1080/0308108042000220686 doi: 10.1080/0308108042000220686
    [2] O. Mangasarian, Absolute value programming, Comput. Optim. Appl., 36 (2007), 43–53. http://dx.doi.org/10.1007/s10589-006-0395-5 doi: 10.1007/s10589-006-0395-5
    [3] O. Mangasarian, A generalized Newton method for absolute value equations, Optim. Lett., 3 (2009), 101–108. http://dx.doi.org/10.1007/s11590-008-0094-5 doi: 10.1007/s11590-008-0094-5
    [4] O. Prokopyev, On equivalent reformulations for absolute value equations, Comput. Optim. Appl., 44 (2009), 363–372. http://dx.doi.org/10.1007/s10589-007-9158-1 doi: 10.1007/s10589-007-9158-1
    [5] R. Ali, Z. Zhang, F. Awwad, The study of new fixed-point iteration schemes for solving absolute value equations, Heliyon, in press. http://dx.doi.org/10.1016/j.heliyon.2024.e34505
    [6] S. Hu, Z. Huang, A note on absolute value equations, Optim. Lett., 4 (2010), 417–424. http://dx.doi.org/10.1007/s11590-009-0169-y doi: 10.1007/s11590-009-0169-y
    [7] O. Mangasarian, R. Meyer, Absolute value equations, Linear Algebra Appl., 419 (2006), 359–367. http://dx.doi.org/10.1016/j.laa.2006.05.004 doi: 10.1016/j.laa.2006.05.004
    [8] O. Mangasarian, Absolute value equation solution via concave minimization, Optim. Lett., 1 (2007), 3–8. http://dx.doi.org/10.1007/s11590-006-0005-6 doi: 10.1007/s11590-006-0005-6
    [9] R. Ali, A. Akgül, A new matrix splitting generalized iteration method for linear complementarity problems, Appl. Math. Comput., 464 (2024), 128378. http://dx.doi.org/10.1016/j.amc.2023.128378 doi: 10.1016/j.amc.2023.128378
    [10] N. Yilmaz, A. Sahiner, Smoothing techniques in solving non-Lipschitz absolute value equations, Int. J. Comput. Math., 100 (2023), 867–879. http://dx.doi.org/10.1080/00207160.2022.2163388 doi: 10.1080/00207160.2022.2163388
    [11] R. Ali, A. Ali, The matrix splitting fixed point iterative algorithms for solving absolute value equations, Asian-Eur. J. Math., 16 (2023), 2350106. http://dx.doi.org/10.1142/S1793557123501061 doi: 10.1142/S1793557123501061
    [12] H. Zhou, S. Wu, C. Li, Newton-based matrix splitting method for generalized absolute value equation, J. Comput. Appl. Math., 394 (2021), 113578. http://dx.doi.org/10.1016/j.cam.2021.113578 doi: 10.1016/j.cam.2021.113578
    [13] D. Salkuyeh, The Picard-HSS iteration method for absolute value equations, Optim. Lett., 8 (2016), 2191–2202. http://dx.doi.org/10.1007/s11590-014-0727-9 doi: 10.1007/s11590-014-0727-9
    [14] A. Khan, J. Iqbal, A. Akgul, R. Ali, Y. Du, A. Hussain, et al., A Newton-type technique for solving absolute value equations, Alex. Eng. J., 64 (2023), 291–296. http://dx.doi.org/10.1016/j.aej.2022.08.052 doi: 10.1016/j.aej.2022.08.052
    [15] M. Noor, J. Iqbal, S. Khattri, E. Al-Said, A new iterative method for solving absolute value equations, Int. J. Phys. Sci., 6 (2011), 1793–1797. http://dx.doi.org/10.5897/IJPS11.244 doi: 10.5897/IJPS11.244
    [16] Y. Ke, C. Ma, SOR-like iteration method for solving absolute value equations, Appl. Math. Comput., 311 (2017), 195–202. http://dx.doi.org/10.1016/j.amc.2017.05.035 doi: 10.1016/j.amc.2017.05.035
    [17] X. Dong, X. Shao, H. Shen, A new SOR-like method for solving absolute value equations, Appl. Numer. Math., 156 (2020), 410–421. http://dx.doi.org/10.1016/j.apnum.2020.05.013 doi: 10.1016/j.apnum.2020.05.013
    [18] J. Tang, Inexact Newton-type method for solving large-scale absolute value equation Ax|x|=b, Appl. Math., 69 (2024), 49–66. http://dx.doi.org/10.21136/AM.2023.0171-22 doi: 10.21136/AM.2023.0171-22
    [19] J. Rohn, V. Hooshyarbakhsh, R. Farhadsefat, An iterative method for solving absolute value equations and sufficient conditions for unique solvability, Optim. Lett., 8 (2014), 35–44. http://dx.doi.org/10.1007/s11590-012-0560-y doi: 10.1007/s11590-012-0560-y
    [20] J. Tang, J. Zhou, A quadratically convergent descent method for the absolute value equation Ax+B|x|=b, Oper. Res. Lett., 47 (2019), 229–234. http://dx.doi.org/10.1016/j.orl.2019.03.014 doi: 10.1016/j.orl.2019.03.014
    [21] L. Caccetta, B. Qu, G. Zhou, A globally and quadratically convergent method for absolute value equations, Comput. Optim. Appl., 48 (2011), 45–58. http://dx.doi.org/10.1007/s10589-009-9242-9 doi: 10.1007/s10589-009-9242-9
    [22] C. Zhang, Q. Wei, Global and finite convergence of a generalized Newton method for absolute value equations, J. Optim. Theory Appl., 143 (2009), 391–403. http://dx.doi.org/10.1007/s10957-009-9557-9 doi: 10.1007/s10957-009-9557-9
    [23] A. Wang, Y. Cao, J. Chen, Modified Newton-type iteration methods for generalized absolute value equations, J. Optim. Theory Appl., 181 (2019), 216–230. http://dx.doi.org/10.1007/s10957-018-1439-6 doi: 10.1007/s10957-018-1439-6
    [24] Y. Lian, C. Li, S. Wu, Weaker convergent results of the generalized Newton method for the generalized absolute value equations, J. Comput. Appl. Math., 338 (2018), 221–226. http://dx.doi.org/10.1016/j.cam.2017.12.027 doi: 10.1016/j.cam.2017.12.027
    [25] Y. Cao, Q. Shi, S. Zhu, A relaxed generalized Newton iteration method for generalized absolute value equations, AIMS Mathematics, 6 (2020), 1258–1275. http://dx.doi.org/10.3934/math.2021078 doi: 10.3934/math.2021078
    [26] C. Zhou, Y. Cao, Q. Shen, Q. Shi, A modified Newton-based matrix splitting iteration method for generalized absolute value equations, J. Comput. Appl. Math., 442 (2024), 115747. http://dx.doi.org/10.1016/j.cam.2023.115747 doi: 10.1016/j.cam.2023.115747
    [27] X. Lv, S. Miao, An inexact fixed point iteration method for solving absolute value equation, Japan J. Indust. Appl. Math., 41 (2024), 1137–1148. http://dx.doi.org/10.1007/s13160-023-00641-3 doi: 10.1007/s13160-023-00641-3
    [28] X. Lv, S. Miao, A new inexact fixed point iteration method for solving tensor absolute value equation, Appl. Math. Lett., 154 (2024), 109109. http://dx.doi.org/10.1016/j.aml.2024.109109 doi: 10.1016/j.aml.2024.109109
    [29] J. Zhang, S. Miao, A general fast shift-splitting iteration method for nonsymmetric saddle point problems, Comp. Appl. Math., 40 (2021), 229. http://dx.doi.org/10.1007/s40314-021-01618-z doi: 10.1007/s40314-021-01618-z
    [30] S. Wu, C. Li, A special shift splitting iteration method for absolute value equation, AIMS Mathematics, 5 (2020), 5171–5183. http://dx.doi.org/10.3934/math.2020332 doi: 10.3934/math.2020332
    [31] V. Edalatpour, D. Hezari, D. Salkuyeh, A generalization of the Gauss-Seidel iteration method for solving absolute value equations, Appl. Math. Comput., 293 (2017), 156–167. http://dx.doi.org/10.1016/j.amc.2016.08.020 doi: 10.1016/j.amc.2016.08.020
    [32] R. Varga, Matrix iterative analysis, Heidelberg: Springer, 1999. http://dx.doi.org/10.1007/978-3-642-05156-2
    [33] A. Fakharzadeh, N. Shams, An efficient algorithm for solving absolute value equations, J. Math. Ext., 15 (2021), 1–23. http://dx.doi.org/10.30495/JME.2021.1393 doi: 10.30495/JME.2021.1393
    [34] C. Li, A preconditioned AOR iterative method for the absolute value equations, Int. J. Comput. Meth., 14 (2017), 1750016. http://dx.doi.org/10.1142/S0219876217500165 doi: 10.1142/S0219876217500165
    [35] R. Ali, K. Pan, The solution of a type of absolute value equations using two new matrix splitting iterative techniques, Port. Math., 79 (2022), 241–252. http://dx.doi.org/10.4171/PM/2089 doi: 10.4171/PM/2089
    [36] Z. Yu, L. Li, Y. Yuan, A modified multivariate spectral gradient algorithm for solving absolute value equations, Appl. Math. Lett., 121 (2021), 107461. http://dx.doi.org/10.1016/j.aml.2021.107461 doi: 10.1016/j.aml.2021.107461
  • This article has been cited by:

    1. Rashid Ali, Zhao Zhang, Exploring two new iterative methods for solving absolute value equations, 2024, 1598-5865, 10.1007/s12190-024-02211-3
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(856) PDF downloads(52) Cited by(1)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog