Research article

Convergence analysis of a gradient iterative algorithm with optimal convergence factor for a generalized Sylvester-transpose matrix equation

  • Received: 01 April 2021 Accepted: 27 May 2021 Published: 03 June 2021
  • MSC : 15A12, 15A60, 46A22, 65F45

  • Consider a generalized Sylvester-transpose matrix equation with rectangular coefficient matrices. Based on gradients and hierarchical identification principle, we derive an iterative algorithm to produce a sequence of approximated solutions with a reasonable stopping rule concerning a relative norm-error. A convergence analysis via Banach fixed-point theorem reveals the sequence converges to a unique solution of the matrix equation for any given initial matrix if and only if the convergence factor is chosen appropriately in a certain range. The performance of algorithm is theoretically analysed through the convergence rate and error estimations. The optimal convergence factor is chosen to attain the fastest asymptotic behaviour. Finally, numerical experiments are provided to illustrate the capability and efficiency of the proposed algorithm, compared to recent gradient-based iterative algorithms.

    Citation: Nunthakarn Boonruangkan, Pattrawut Chansangiam. Convergence analysis of a gradient iterative algorithm with optimal convergence factor for a generalized Sylvester-transpose matrix equation[J]. AIMS Mathematics, 2021, 6(8): 8477-8496. doi: 10.3934/math.2021492

    Related Papers:

    [1] Kanjanaporn Tansri, Sarawanee Choomklang, Pattrawut Chansangiam . Conjugate gradient algorithm for consistent generalized Sylvester-transpose matrix equations. AIMS Mathematics, 2022, 7(4): 5386-5407. doi: 10.3934/math.2022299
    [2] Huiling Wang, Nian-Ci Wu, Yufeng Nie . Two accelerated gradient-based iteration methods for solving the Sylvester matrix equation AX + XB = C. AIMS Mathematics, 2024, 9(12): 34734-34752. doi: 10.3934/math.20241654
    [3] Fengxia Zhang, Ying Li, Jianli Zhao . The semi-tensor product method for special least squares solutions of the complex generalized Sylvester matrix equation. AIMS Mathematics, 2023, 8(3): 5200-5215. doi: 10.3934/math.2023261
    [4] Huamin Zhang, Hongcai Yin . Zeroing neural network model for solving a generalized linear time-varying matrix equation. AIMS Mathematics, 2022, 7(2): 2266-2280. doi: 10.3934/math.2022129
    [5] Siting Yu, Jingjing Peng, Zengao Tang, Zhenyun Peng . Iterative methods to solve the constrained Sylvester equation. AIMS Mathematics, 2023, 8(9): 21531-21553. doi: 10.3934/math.20231097
    [6] Sourav Shil, Hemant Kumar Nashine . Positive definite solution of non-linear matrix equations through fixed point technique. AIMS Mathematics, 2022, 7(4): 6259-6281. doi: 10.3934/math.2022348
    [7] Adisorn Kittisopaporn, Pattrawut Chansangiam . Approximate solutions of the $ 2 $D space-time fractional diffusion equation via a gradient-descent iterative algorithm with Grünwald-Letnikov approximation. AIMS Mathematics, 2022, 7(5): 8471-8490. doi: 10.3934/math.2022472
    [8] Abdur Rehman, Ivan Kyrchei, Muhammad Zia Ur Rahman, Víctor Leiva, Cecilia Castro . Solvability and algorithm for Sylvester-type quaternion matrix equations with potential applications. AIMS Mathematics, 2024, 9(8): 19967-19996. doi: 10.3934/math.2024974
    [9] Kanjanaporn Tansri, Pattrawut Chansangiam . Gradient-descent iterative algorithm for solving exact and weighted least-squares solutions of rectangular linear systems. AIMS Mathematics, 2023, 8(5): 11781-11798. doi: 10.3934/math.2023596
    [10] Jamilu Sabi'u, Ibrahim Mohammed Sulaiman, P. Kaelo, Maulana Malik, Saadi Ahmad Kamaruddin . An optimal choice Dai-Liao conjugate gradient algorithm for unconstrained optimization and portfolio selection. AIMS Mathematics, 2024, 9(1): 642-664. doi: 10.3934/math.2024034
  • Consider a generalized Sylvester-transpose matrix equation with rectangular coefficient matrices. Based on gradients and hierarchical identification principle, we derive an iterative algorithm to produce a sequence of approximated solutions with a reasonable stopping rule concerning a relative norm-error. A convergence analysis via Banach fixed-point theorem reveals the sequence converges to a unique solution of the matrix equation for any given initial matrix if and only if the convergence factor is chosen appropriately in a certain range. The performance of algorithm is theoretically analysed through the convergence rate and error estimations. The optimal convergence factor is chosen to attain the fastest asymptotic behaviour. Finally, numerical experiments are provided to illustrate the capability and efficiency of the proposed algorithm, compared to recent gradient-based iterative algorithms.



    It is well known that the fundamental of differential equations deals with the algebraic linear system

    x(t)=Ax(t), (1.1)

    where x(t) is an unknown vector-valued function and A is a given square matrix. To analyse the stability of an equilibrium point of the system (1.1), it suffices to find a positive definite matrix L such that ATL+LA is negative definite; see e.g., [1]. So, we need to solve the so-called Lyapunov equation

    AX+XAT=R (1.2)

    for some negative definite matrix R. To discuss the Eq (1.2), we investigate a more general form, namely, the generalized Sylvester-transpose matrix equation

    pi=1AiXBi+qj=1CjXTDj=F, (1.3)

    where Ai,Bi,Cj,Dj and F are given matrices of conforming dimensions and X is an unknown matrix to be determined. The Eq (1.3) includes important practical problems that are written as the matrix equations

    AX+XB=C, (1.4)
    AX+XTB=C, (1.5)
    AXB+CXD=E, (1.6)
    AXB+X=C, (1.7)

    known as Sylvester, Sylvester-transpose, generalized Sylvester, Kalman-Yakubovich matrix equations, respectively. The Eqs (1.2)–(1.7) have many essential applications in control theory; see e.g., [2,3,4,5,6]. In traditional method, a vectorization and the Kronecker product are used to find the unique solution. However, the large size of the Kronecker multiplication leads to computationally difficulty in that excessive computer. For this reason, iterative algorithms are received more attention.

    Many researchers attempted to meliorate such iterative algorithms for finding the approximated solutions of matrix equations (1.2)–(1.7) using many ideas, e.g., matrix sign function [7,8], recursive blocked algorithms [9,10] and Hermitian and skew-Hermitian splitting algorithms [11,12,13]. In 2005, gradient-based iterative algorithms were firstly introduced by F. Ding and T. Chen for solving (1.4), (1.6) and (1.7); see [14,15]. In a few year later, many iterative algorithms that relied on gradients and hierarchical identification principle for solving (1.2)–(1.7) are established, e.g., RGI [16], JGI [17,18], MGI [19] and AGBI [20]. See more information in [21,22,23,24,25]. The Frobenious norm F and the spectral norm 2 for matrices are used to analyse the convergence property of such algorithms. There are defined respectively for any real matrix A by

    AF=(trATA)12andA2=(λmax(ATA))12.

    A gradient-based iterative algorithm for solving (1.3) was presented as follows:

    Theorem 1.1 ([25]). Suppose that the matrix equation (1.3) has a unique solution X. For each s=1,2,,p and t=p+1,,q, construct

    Xs(k)=X(k1)+τATs[Fpi=1AiX(k1)Biqj=1CjXT(k1)Dj]BTs,Xt(k)=X(k1)+τDt[Fpi=1AiX(k1)Biqj=1CjXT(k1)Dj]TCt,X(k)=1p+q(ps=1Xs(k)+qt=p+1Xt(k)).

    A necessary and sufficient condition for which the sequence {Xk} converges to X for any given initial matrices X1(0),X2(0),,Xp+q(0) is that

    0<τ<2(pi=1As22Bs22+qt=p+1Ct22Dt22).

    Meanwhile, a least-squares based iterative algorithm for solving (1.3) was presented as follows:

    Theorem 1.2 ([25]).

    For each s=1,2,,p and t=p+1,,q, construct

    R(k)=Epi=1AiX(k1)Biqj=1CjXT(k1)Dj,Xs(k)=X(k1)+μ(ATsAs)1ATsR(k)BTs(BsBTs)1,Xt(k)=X(k1)+μ(DtDTt)1DtR(k)Ct(CTtCt)1,X(k)=1p+q(ps=1Xs(k)+qt=p+1Xt(k)).

    The sequence {Xk} converges to a unique solution X for any given initial matrices X1(0),X2(0),,Xp+q(0) if and only if 0<τ<2(p+q).

    In this work, we propose an effective iterative algorithm based on gradient and hierarchical identification principle, namely, a gradient iterative algorithm with optimal convergence factor. The proposed algorithm is applicable for the generalized Sylvester-transpose matrix equation (1.3); see Section 2. Convergence analysis (see Section 3) via Banach fixed-point theorem reveals that the iterative solutions converges to the unique solution for any initial value if and only if the convergence factor is chosen appropriately belong to an open interval. Then we study the performance of the algorithm from the convergence rate and error estimates. Moreover, we determine the fastest asymptotic convergence rate to minimize the spectral radius of the iteration matrix. Furthermore, we apply the algorithm to the Sylvester-transpose matrix equation; see Section 4. To show the applicability and the performance of the algorithm, we give numerical experiments in Section 5. In Section 6, we summarize the whole work. For benefits of reader, we include MATLAB-code for numerical experiment in Appendix.

    Let us denote by Rr×s the set of r×s real matrices. Consider the matrix equation (1.3) where AiRm×n, BiRp×q, CjRm×p, DjRn×q, FRm×q are given coefficient matrices and XRn×p is an unknown matrix to be determined. The dimension matching of matrices is assumed to be mq=np.

    Recall that the commutation matrix Kmn is defined by

    Kmn=[ImenT1ImenT2ImenTn]Rmn×mn

    where eni is the ith column of the n×n identity matrix In. Essential properties of the commutation matrices are given in the following lemma:

    Lemma 2.1. (see e.g., [26,Ch. 4]) For any ARm×n and BRp×q, we have

    vec(AT)=Kmnvec(A), (2.1)
    K1mn=KTmn=Knm, (2.2)
    (BA)=Kpm(AB)Knq. (2.3)

    A traditional method to solve the matrix equation (1.3) is to take the vector operator and utilize Lemma 2.1. Indeed, we arrive at an equivalent linear system Qx=b where

    Q=pi=1(BTiAi)+qj=1(DTjCj)KnpRnp×np,

    x=vec[X] and b=vec[F]. From now on, assume that Q is invertible. So, the matrix equation (1.3) has a unique solution

    x=Q1b. (2.4)

    However, if the dimensions of Ai,Bi,Cj,Dj are not small, e.g., 102×102, then the dimension of Q is 104×104. Such a dimension problem leads to computational difficulty for computation and inversion of large matrices. Hence, this approach is only applicable for small dimensional matrices.

    We now propose an effective iterative algorithm to solve (1.3).

    If pq then we set Cj=0 and Dj=0 for any j>q. If q>p then we set Ai=0 and Bi=0 for any i>p. For each i=1,2,,p, we assume that

    Mi:=F(si(AsXBs+CsXTDs)). (2.5)

    From the main system (1.3), we would like to solve p subsystems

    AiXBi+CiXTDi=Mii=1,2,,p, (2.6)

    so that the following errors are minimized:

    Li(X):=AiXBi+CiXTDiMi2F,i=1,2,,p. (2.7)

    We can derive the gradient of each Li as follows:

    XLi(X)=Xtr[(AiXBi+CiXTDiMi)T(AiXBi+CiXTDiMi)]=Xtr(BTiXTATiAiXBi)+Xtr(DTiXCTiAiXBi)Xtr(MTiAiXBi)+Xtr(BTiXTATiCiXTDi)+Xtr(DTiXCTiCiXTDi)Xtr(MTiCiXTDi)Xtr(BTiXTATiMi)Xtr(DTiXCTiMi)+Xtr(MTiMi)=2[ATi(AiXBi+CiXTDiMi)BTi+Di(AiXBi+CiXTDiMi)TCi]. (2.8)

    Let Xi(k) be the approximated solution of the system (2.6) at iteration k. The recursive formula of Xi(k) come from the gradient formula of Li(X) as follows:

    Xi(k)=Xi(k1)+μXLi(X),

    where μ is a step size parameter. To avoid duplicated computation, we introduce a matrix

    E=Fpi=1AiXBiqj=1CjXTDj.

    Taking the arithmetic mean of X1(k),,Xp(k) to get X(k):

    X(k)=1ppi=1Xi(k)=X(k1)+τ(pi=1ATiEBTi+qj=1DjETCj), (2.9)

    where τ:=2μ/p is called a convergence factor. The hierarchical identification principle suggests to replace the unknown solution X in (2.9) by its previous estimate X(k1). Thus we obtain the following procedure:

    For convenience, we write XT(k) and ET(k) for X(k)T and E(k)T, respectively. The matrices E(k),Ai,Bi were introduced to avoid duplicate manipulations.

    Remark 2.2. To break the procedure, if FF is close to zero, then we should consider the error E(k)F or X(k)X(k1)F instead of the relative error E(k)F/FF.

    The convergent property of the algorithm relies on the convergence factor τ, which will be discussed in the next section.

    In this section, we discuss a convergence criteria for the applicability of Algorithm 1. Then, we analyse the performance of the algorithm through its convergence rate and error estimates. The last job is to determine the convergence factor for which the algorithm fits with the fastest asymptotic behaviour. The main idea for the analysis starts with transforming a recursive equation of the error of approximated solutions into a fixed-point iteration x(k)=Sx(k1), where S:RnpRnp is a contraction mapping. Then, we apply the following Banach fixed-point theorem to analysis the algorithm.

    Theorem 3.1. (see e.g., [27]) Let (X,d) be a non-empty complete metric space with a contraction mapping T:XX. Then

    (i) The map T admits a unique fixed-point x in X i.e. T(x)=x.

    (ii) The fixed-point x can be found as follows: start with an arbitrary element x0 in X and define a sequence {xn} by xn=T(xn1) for n1. Then xnx.

    (iii) The following inequalities hold and describe the speed of convergence:

    d(x,xn)zn1zd(x1,x0) (3.1)
    d(x,xn+1)z1zd(xn+1,xn) (3.2)
    d(x,xn+1)zd(x,xn). (3.3)

    From Algorithm 1, at each k-th iteration, we start with considering the error matrix ˆX(k)=X(k)X. Indeed, we have

    ˆX(k)=X(k1)+τ(pi=1ATiE(k1)BTi+qj=1DjET(k1)Cj)X=ˆX(k1)+τ(pi=1ATiE(k1)BTi+qj=1DjET(k1)Cj), (3.4)

    and

    E(k1)=(pi=1AiˆX(k1)Bi+qj=1CjˆXT(k1)Dj).

    Thus, Lemma 2.1 implies that

    vecE(k1)=pi=1(BTiAi)vecˆX(k1)qj=1(DTjCj)vecˆXT(k1)=pi=1(BTiAi)vecˆX(k1)qj=1(DTjCj)KnpvecˆX(k1)=QvecˆX(k1). (3.5)

    Taking the vector operator to the Eq (3.4) and utilizing (3.5), we get

    vecˆX(k)=vecˆX(k1)+τ[pi=1vecATiE(k1)BTi+qj=1vecDjET(k1)Cj]=vecˆX(k1)+τ[pi=1(BiATi)vecE(k1)+qj=1(CTjDj)KmqvecE(k1)]=vecˆX(k1)τ[pi=1(BiATi)QvecˆX(k1)+qj=1Kpn(DjCTj)QvecˆX(k1)]=(InpτQTQ)vecˆX(k1).

    Letting S=InpτQTQ and x(k)=vecˆX(k1), we get a linear iteration

    x(k)=Sx(k1). (3.6)

    Using Theorem 3.1 and [28,Theorem 1], we deduce that the following are equivalent:

    (ⅰ) the sequence {X(k)} converges to X for any initial value X(0);

    (ⅱ) S is a contraction mapping (here, we view S:RnpRnp as a mapping);

    (ⅲ) the spectral norm of S is less than 1.

    Since S is symmetric, all its eigenvalue are real. Note that the eigenvalues of S are of the form 1τλ where λ is any eigenvalue of QTQ. We can compute the spectral radius of S as follows:

    ρ[S]=max{|1τλmax(QTQ)|,|1τλmin(QTQ)|}. (3.7)

    Thus, ρ[S]<1 if and only if

    0<τλmax(QTQ)<2. (3.8)

    Since Q is invertible, the matrix QTQ is positive definite and hence, Q22=λmax(QTQ)>0. The condition (3.8) now becomes

    0<τ<2Q22. (3.9)

    We summarize convergence criteria for Algorithm 1 as follows:

    Theorem 3.2. Consider the matrix equation (1.3) under the assumption that the matrix Q is invertible (i.e., Eq (1.3) has a unique solution). Let τR. Then a necessary and sufficient condition for which Algorithm 1 is applicable for any initial matrix X(0) is the condition (3.9).

    We now discuss the performance of Algorithm 1 through its convergence rate and error estimates.

    Considering Eq (3.6), we get

    X(k)XF=ˆX(k)F=vecˆX(k)F=SvecˆX(k1)FS2vecˆX(k1)F.

    Since S is symmetric, we have S2=ρ[S]. Thus the Eq (3.3) implies that

    X(k)XFρ[S]X(k1)XF, (3.10)

    By induction, we obtain

    X(k)XFρk[S]X(0)XF. (3.11)

    According to the estimate (3.11), the asymptotic convergence rate of the algorithm relies on ρ[S]. Moreover, given an error ϵ>0, we would like to find the iteration number k for which

    ρk(S)X(0)XF<ϵ. (3.12)

    By taking the 10th-base logarithms, we obtain the following equivalent requirement:

    k>logϵlogX(0)XFlogρ(S). (3.13)

    Moreover, by Theorem 3.1(ⅲ), we have

    X(k)XFρk[S]1ρ[S]X(1)X(0)F, (3.14)
    X(k+1)XFρ[S]1ρ[S]X(k+1)X(k)F. (3.15)

    The following results are discussed to concluding the convergence rate and several estimates of the proposed algorithm.

    Theorem 3.3. Suppose the convergence factor τ is chosen so that Algorithm 1 is applicable for any initial matrix X(0).

    (i) The spectral radius ρ[S] in (3.7) governs the asymptotic convergence rate of the algorithm.

    (ii) Equations (3.10) and (3.11) show the error estimates X(k)XF compared to the previous step and the first step, respectively. Meanwhile, the error at each iteration reduces from the previous one.

    (iii) The prior and posterior estimates are presented in (3.14) and (3.15), respectively.

    (iv) For each given error ε>0, we have X(k)XF<ϵ after the k-th iteration for any kN that satisfies (3.13).

    We can see that Ai,Bi,Cj,Dj affect the convergence rate of Algorithm 1 but E is not. However, the matrix E is required for stopping process. Furthermore, If we take ϵ=0.5×10n in (3.13) and k satisfies

    k>log0.5logX(0)XFnlogρ(S),

    then the approximated X(k) has an accuracy of n decimal digit.

    The fastest convergence factor of Algorithm 1 is discussed intensively. Recall that the condition number of a matrix A (relative to the spectral norm) is defined by

    κ(A)=(λmax(ATA)λmin(ATA))12.

    Assume that Eq (3.9) holds. The convergence rate of Algorithm 1 is the same as that of the linear iteration (3.6), and thus, it is given by the spectral radius (3.7) of the iteration matrix S. The fastest convergence rate is equivalent to the smallest of ρ[S]. Thus, we would like to minimize this spectral radius subject to the condition (3.9). Now, we apply the following lemma:

    Lemma 3.4. ([19]) For any real number a,b with b>a>0, we have

    min0<x<2b{max{1ax,1bx}}=bab+a.

    The minimality is reached at xopt=2a+b.

    Then the minimum value of ρ[S] is

    τopt=2λmax(QTQ)+λmin(QTQ). (3.16)

    Meanwhile, we can notice that spectral radius of the iteration matrix is

    ρ[S]=λmax(QTQ)λmin(QTQ)λmax(QTQ)+λmin(QTQ)=κ2(Q)1κ2(Q)+1. (3.17)

    Thus, we obtain the following theorem:

    Theorem 3.5. Among the convergence factors τ that meet the criteria of Algorithm 1, the one in Eq (3.16) attains the fastest asymptotic convergence rate of the algorithm, which is governed by the spectral radius (3.17).

    The above theorem tells us that if λmax(QTQ) is neighbouring to λmin(QTQ), or equivalently, the condition number is close to one then Algorithm 1 has a fast convergence.

    In this section, we consider an important special case of the matrix equation (1.3), namely, the Sylvester-transpose matrix equation. We apply GIO algorithm for this equation, and investigate the convergence property of the algorithm.

    Let m,nN. Consider the Sylvester-transpose matrix equation (1.5) where ARm×n, BRn×m, FRm×m are given constant matrices and XRn×m is an unknown matrix to be solved. Suppose that (1.5) has a unique solution, i.e., that following matrix is invertible:

    P:=(IA)+(BTI)KnmRmn×mn.

    Corollary 4.1. Consider the matrix equation (1.5) when the matrix P is invertible. Let τR. Then the given statements hold:

    (i) A equivalent condition for which Algorithm 2 is applicable for any initial matrix X(0) is

    0<τ<2P22. (4.1)

    Meanwhile, the spectral radius of the iteration matrix T=InpτPTP is given by

    ρ[T]=max{|1τλmax(PTP)|,|1τλmin(PTP)|}. (4.2)

    (ii) The spectral radius ρ[T] in (4.2) represents the asymptotic convergence rate of Algorithm 2.

    (iii) The fastest asymptotic convergence factor is determined by

    τopt=2λmax(PTP)+λmin(PTP). (4.3)

    In this section, we report some numerical results to illustrate the applicability the effectiveness of Algorithm 1. All simulations have been carried out by MATLAB R2018a, AMD Ryzen7 3700U with Redeon Vega Mobile Gfx @ 2.30 GHz, RAM 12.00 GB PC environment. In Example 5.1, we show that our algorithm is applicable for small-square-matrices in the case p>q. We also show that our algorithm is applicable and efficient for large-square-matrices and consider the effect of changing the convergence factor τ in Example 5.2. In Example 5.3, we illustrate the efficiency of Algorithm 1 when coefficients are non-square matrices of different moderate sizes. In Examples 5.4 and 5.5, we do experiments in the cases p>q and q>p with rectangular coefficient matrices. In Example 5.6, we test Algorithm 2 for the Sylvester-transpose equation with square coefficient matrices. In all examples, we compare the proposed algorithm to both the traditional method (Eq (2.4)) and recent iterative algorithms. The computational time is measured in seconds by MATLAB functions tic and toc.

    Example 5.1. Consider the matrix equation

    A1XB1+C1XTD1+A2XB2=F,

    where

    A1=[0.1230.0020.7800.5630.0090.1230.0080.0050.0970.0020.3980.0070.0230.0940.0010.0090.4780.9940.0010.0050.0130.0030.0280.0040.456],A2=[0.1120.3020.7850.3120.0490.7090.9960.7330.2190.0050.2610.0050.0030.1140.1110.2190.0050.1230.1250.0090.0010.0000.0180.9940.956],
    B1=[0.6670.2090.3460.6750.0990.0990.2180.2780.2190.0040.0020.0050.1090.6780.2340.0560.0050.0060.1950.0090.0040.0650.1870.9840.000],B2=[0.0040.0560.0050.0040.0490.5790.0960.1140.0080.1120.1130.1190.2840.0030.0140.0890.0270.0090.1450.0360.0010.0790.4560.4581.000],
    C1=[0.1630.0210.0070.1520.1930.4740.0980.0010.3840.1930.0850.1090.0930.0170.1730.8120.7420.8410.9410.4850.1970.9340.0120.8450.917],D1=[0.0020.0740.0040.0720.2840.0560.0370.4850.1880.4850.8630.0720.4750.9450.5940.0160.0340.0040.0010.8550.8540.0030.9270.9230.567].

    In fact, the unique solution is given by

    X=[1.0000.0100.2240.1110.9080.9800.7650.3650.4820.5280.6490.3090.8490.0300.6120.4950.0080.8620.0010.0040.2390.9370.2510.3640.062].

    Let us apply Algorithm 1 to compute the sequence {X(k)} of approximated solutions. Take an initial point X(0)=0. The optimal convergence factor can be computed according to Theorem 3.2 as follows:

    τopt=2λmin(QTQ)+λmax(QTQ)28.3389×106+14.50240.1379.

    Table 1 shows that the direct method consumes 15 ms to get the exact solution, while GIO algorithm takes 6 ms to perform 10 iterations in order to get an approximated solution with a small relative error. Figure 1 and Table 1 illustrate that the computational time of GIO algorithm is less than that for GI algorithm with preferable relative error. Figure 1 and Table 1 also show that GIO algorithm is outperform than LS algorithm.

    Table 1.  Numerical results for Example 5.1.
    Algorithm IT CT relative error
    Direct - 0.0150 -
    GIO 10 0.0060 0.5088
    GI with τ=0.127 10 0.0078 0.7510
    GI with τ=0.009 10 0.0069 0.9755
    LS 10 0.0223 1.0000

     | Show Table
    DownLoad: CSV
    Figure 1.  Relative error for Example 5.1.

    Example 5.2. Consider the matrix equation

    2i=1AiXBi+2j=1CjXTDj=F

    where all coefficients are 100×100 tridiagonal matrices given by

    A1=tridiag(3,1,1),A2=tridiag(1,0,4),B1=tridiag(1,3,2),B2=tridiag(1,2,1),C1=tridiag(1,0,2),C2=tridiag(1,2,3),D1=tridiag(0,2,4),D2=tridiag(1,1,1).

    In fact, the unique solution is given by X=tridiag(0,1,1). Let us apply Algorithm 1 to compute the sequence {X(k)} of approximated solutions. Take an initial point

    X(0)=106×tridiag(1,1,1).

    The optimal convergence factor can be computed according to Theorem 3.2 as follows:

    τopt=2λmin(QTQ)+λmax(QTQ)24.15×1013+7.15×1040.000028.

    The computing results of changing τ are listed in Figure 2 and Table 2. Figure 2 shows that, for the given initial point, as k increases, for τ=τopt,τ=0.00002,τ=0.000015, and τ=0.00001, we see that the terms E(k)F/FF are becoming smaller and approaching to zero. Moreover, the relative errors E(k)F/FF for τopt go faster to 0 than those for another convergence factors. Furthermore, when τ=0.00004 and τ=0.00001 which do not satisfy the criteria (3.9), the approximated solutions diverge. Table 2 confirms that the computational time of the proposed algorithm is significantly less than the time of the traditional method. Thus, in this case, Algorithm 1 is applicable and effective.

    Figure 2.  Relative errors for Example 5.2.
    Table 2.  Numerical results for Example 5.2.
    Algorithm IT CT relative error
    Direct - 15.2058 -
    GIO 100 0.5467 0.8005
    GI with τ=0.00002 100 1.6245 0.8099
    GI with τ=0.000015 100 1.7007 0.8512
    GI with τ=0.00001 100 2.8129 0.8619
    GI with τ=0.00004 100 2.9998 1.0055
    GI with τ=0.00001 100 4.0097 1.3314

     | Show Table
    DownLoad: CSV

    Example 5.3. Consider the matrix equation

    3i=1AiXBi+3j=1CjXTDj=F

    when AiR40×60, BiR20×30, CjR40×20, DjR60×30 and FR40×30 are tridiagonal matrices given by

    A1=tridiag(1,1,1),A2=tridiag(2,0,3),A3=tridiag(2,1,2),B1=tridiag(1,3,0),B2=tridiag(1,2,1),B3=tridiag(0,1,3),C1=tridiag(3,0,2),C2=tridiag(1,2,3),C3=tridiag(2,1,2),D1=tridiag(0,2,1),D2=tridiag(1,2,1),D3=tridiag(0,1,1).

    Then the unique solution is X=tridiag(0,1,1)R60×20. We take the initial matrix

    X(0)=106×tridiag(1,1,1).

    The computing results of changing τ are listed in Figure 3. As k large enough, the terms E(k)F/FF for τopt0.000085 are becoming smaller and go faster to zero than another convergence factors. Moreover, for τ=0.0003 and τ=0.000001 which do not satisfy the condition (3.9), we see that the term E(k)F/FF do not approach zero, so the approximated solutions diverge. From Table 3, it is clear that the computational time of GIO algorithm is significantly less than the time of the traditional method and another GI with different convergence factors.

    Figure 3.  Relative error for Example 5.3.
    Table 3.  Numerical results for Example 5.3.
    Algorithm IT CT relative error
    Direct - 4.7478 -
    GIO 100 0.2350 0.3012
    GI with τ=0.00006 100 1.6954 0.3465
    GI with τ=0.00003 100 2.7760 0.4802
    GI with τ=0.00001 100 3.9008 0.7219
    GI with τ=0.0003 100 4.4435 0.7906
    GI with τ=0.00001 100 4.0978 1.0244

     | Show Table
    DownLoad: CSV

    Example 5.4. Consider the matrix equation

    3i=1AiXBi+2j=1CjXTDj=F

    when AiR40×60, BiR20×30, CjR40×20, DjR60×30 and FR40×30 are tridiagonal matrices given by

    A1=tridiag(2,2,2),A2=tridiag(3,3,4),A3=tridiag(3,1,2),B1=tridiag(2,3,5),B2=tridiag(2,3,1),B3=tridiag(1,0,3),C1=tridiag(3,4,2),C2=tridiag(2,2,3),D1=tridiag(5,3,1),D2=tridiag(1,2,3).

    Then the unique solution is X=tridiag(1,2,1)R60×20. We take the initial matrix

    X(0)=106×tridiag(1,1,1)R60×20.

    The numerical results in Figure 4 show that the direct method consumes around 11 seconds, while GIO algorithm spends 0.047 seconds (50 iterations) with satisfactory error. Figure 4 and Table 4 show that the computational time of GIO algorithm is slightly more than that of GI algorithm, as the relative error of GIO is significantly less than that of GI algorithm. Figure 4 and Table 4 also show that GIO algorithm is outperform than LS algorithm in both computational time and relative error.

    Figure 4.  Relative error for Example 5.4.
    Table 4.  Numerical results for Example 5.4.
    Algorithm IT CT relative error
    Direct - 10.8753 -
    GIO 50 0.0470 0.1163
    GI with τ=0.00005 50 0.0411 0.2952
    GI with τ=0.00003 50 0.0367 0.4376
    LS 50 0.2664 1.0000

     | Show Table
    DownLoad: CSV

    Example 5.5. Consider the matrix equation

    2i=1AiXBi+3j=1CjXTDj=F

    where all coefficients are 10×10 tridiagonal matrices given by

    A1=tridiag(1,2,1),A2=tridiag(2,4,3),B1=tridiag(1,3,2),B2=tridiag(2,3,1),C1=tridiag(2,3,1),C2=tridiag(1,3,1),C3=tridiag(5,3,4),D1=tridiag(1,2,1),D2=tridiag(4,2,1),D3=tridiag(2,3,1).

    In fact, the unique solution is given by X=tridiag(1,1,1). Let us apply Algorithm 1 to compute the sequence {X(k)} of approximated solutions using an initial point

    X(0)=106×tridiag(1,1,1).

    Table 5 shows that for small size matrices, the difference between the computational times for the direct method and GIO algorithm (50 iterations) is not much as that for the moderate sizes in Example 5.4. Figure 5 and Table 5 illustrate that the computational time of GIO algorithm is less than that for GI algorithm with preferable relative error. Figure 5 and Table 5 also show that GIO algorithm is outperform than LS algorithm.

    Table 5.  Numerical results for Example 5.5.
    Algorithm IT CT relative error
    Direct - 0.0266 -
    GIO 50 0.0112 0.0370
    GI with τ=0.00005 50 0.0146 0.2813
    GI with τ=0.00008 50 0.0134 0.1526
    LS 50 0.0621 0.9942

     | Show Table
    DownLoad: CSV
    Figure 5.  Relative error for Example 5.4.

    Example 5.6. The Sylvester-transpose equation AX+XTB=F, where A,B,F,XR10×10 are given by

    A=tridiag(1,3,1),B=tridiag(2,2,4),X=tridiag(4,1,4)

    is considered by using the initial matrix X(0)=tridiag(106,106,106). The objective is to compare the capability of GIO algorithm to GI [25], LS [25], and AGBI [20] algorithms. We fix the iteration number to be 50 and investigate the relative error E(k)F/FF. Figure 6 and Table 6 indicate that Algorithm 2 is more efficient than the traditional method and another iterative algorithms in spite of a little more computational time.

    Figure 6.  Relative error for Example 5.6.
    Table 6.  Numerical results for Example 5.6.
    Algorithm IT CT relative error
    Direct - 0.0303 -
    GIO 50 0.0056 0.8621
    GI 50 0.0070 0.8770
    LS 50 0.0121 0.9884
    AGBI 50 0.0105 0.8838

     | Show Table
    DownLoad: CSV

    The proposed algorithm, a gradient iterative algorithm with an optimal convergence factor, is applicable for a generalized Sylvester-transpose matrix equation (1.3) under the assumption that the matrix equation has a unique solution. The analysis tells us that a necessary and sufficient condition for which Algorithms 1 and 2 are applicable for any initial matrix X(0) is the convergence factor is chosen appropriately belong to an open interval. The convergence rate and several estimations (e.g., error, prior, posterior) depend on the spectral radius of the associated iteration matrix. Moreover, we can determine the fastest asymptotic convergence rate. The numerical experiments illustrate that the algorithm can be applied for any matrix problems with conformable coefficient matrices of small/moderate/large sizes and square/non-square sizes. Moreover, the algorithm has more efficient than the traditional method and another recent gradient-based iterative algorithms.

    The first author would like to thank Ministry of Education, Thailand for a financial support during her PhD study.

    The authors declare that they have no conflict of interest.

    This appendix contains MATLAB-code of Example 5.1

    tic;

    τ=0.1379;

    Xt = X;

    XtT = transpose(Xt);

    Et = F - (A1*Xt*B1 + A2*Xt*B2 + A3*Xt*B3 + C1*XtT*D1 + C2*XtT*D2 + C3*XT*D3);

    EtT = transpose(Et);

    et(1) = e(1)/norm(F, 'fro');

    let(1) = le(1);

    kt = 2;

    while (kt<50)

    Et = F - (A1*Xt*B1 + A2*Xt*B2 + A3*Xt*B3 + C1*XtT*D1 + C2*XtT*D2 + C3*XT*D3);

    EtT = transpose(Et); Xt = Xt + τ*(A1T*Et*B1T+ A2T*Et*B2T + A3T*Et*B3T + D1*EtT*C1 + D2*EtT*C2 + D3*EtT*C3);

    et(kt) = norm(Et, 'fro')/norm(F, 'fro');

    let(kt) = log(norm(Et, 'fro'));

    kt = kt+1;

    end

    tt = toc;



    [1] G. Dullerud, F. Paganini, A course in robust control theory – a convex approach, New York: Springer-Verlag, 1994.
    [2] C. C. Tsui, On robust observer compensator design, Automatica, 24 (1988), 687–692. doi: 10.1016/0005-1098(88)90116-1
    [3] P. V. Dooren, Reduce order observer: A new algorithm and proof, Syst. Control Lett., 4 (1984), 243–251. doi: 10.1016/S0167-6911(84)80033-X
    [4] H. K. Wimmer, Consistency of a pair of generalized Sylvester equations, IEEE T. Automat. Contr., 39 (1994), 1014–1016. doi: 10.1109/9.284883
    [5] A. Wu, G. Duan, Y. Xue, Kronecker maps and Sylvester-polynomial matrix equations, IEEE T. Automat. Contr., 52 (2007), 905–910. doi: 10.1109/TAC.2007.895906
    [6] A. Wu, G. Duan, B. Zhou, Solution to generalized Sylvester matrix equations, IEEE T. Automat. Contr., 53 (2008), 811–815. doi: 10.1109/TAC.2008.919562
    [7] R. Bartels, G. Stewart, Solution of the matrix equation AX+XB=C, Commun. ACM, 15 (1972), 820–826. doi: 10.1145/361573.361582
    [8] P. Benner, S. Quintana, Solving stable generalized Lyapunov matrix equations with the matrix sign function, Numer. Algorithms, 20 (1999), 75–100. doi: 10.1023/A:1019191431273
    [9] I. Jonsson, B. Kagstrom, Recursive blocked algorithms for solving triangular systems-Part Ⅰ: One-sided and couple Sylvester-type matrix equations, ACM T. Math. Software, 28 (2002), 392–415. doi: 10.1145/592843.592845
    [10] I. Jonsson, B. Kagstrom, Recursive blocked algorithms for solving triangular systems-Part Ⅱ: Two-sided and generalized Sylvester and Lyapunov matrix equations, ACM T. Math. Software, 28 (2002), 416–435. doi: 10.1145/592843.592846
    [11] X. Wang, Y. Li, L. Dai, On the Hermitian and skew-Hermitian splitting iteration methods for the linear matrix equation AXB=C, Comput. Math. Appl., 65 (2013), 657–664. doi: 10.1016/j.camwa.2012.11.010
    [12] H. M. Zhang, F. Ding, A property of the eigenvalues of the symmetric positive definite matrix and the iterative algorithm for couple Sylvester matrix equations, J. Franklin I., 351 (2014), 340–357. doi: 10.1016/j.jfranklin.2013.08.023
    [13] Z. Z. Bai, On Hermitian and skew-Hermitian splitting iteration methods for continuous Sylvester equation, J. Comput. Math., 29 (2011), 185–198. doi: 10.4208/jcm.1009-m3152
    [14] F. Ding, T. Chen, Gradient based iterative algorithms for solving a class of matrix equations, IEEE T. Automat. Contr., 50 (2005), 1216–1221. doi: 10.1109/TAC.2005.852558
    [15] F. Ding, T. Chen, Iterative least squares solutions of coupled Sylvester matrix equations, Syst. Control Lett., 54 (2005), 95–107. doi: 10.1016/j.sysconle.2004.06.008
    [16] Q. Niu, X. Wang, L. Z. Lu, A relaxed gradient based algorithm for solving Sylvester equation, Asian J. Control, 13 (2011), 461–464. doi: 10.1002/asjc.328
    [17] W. Fan, C. Gu, Z. Tian, Jacobi-gradient iterative algorithms for Sylvester matrix equations, In: Linear Algebra Society Topics, Shanghai University, Shanghai, China, 2007, 16–20.
    [18] S. K. Li, T. Z. Huang, A shift-splitting Jacobi-gradient algorithm for Lyapunov matrix equation arising form control theory, J. Comput. Anal. Appl., 13 (2011), 1246–1257.
    [19] Y. J. Xie, C. F. Ma, The accelerated gradient based iterative algorithm for solving a class of generalized Sylvester-transpose matrix equation, Appl. Math. Comput., 273 (2016), 1257–1269.
    [20] X. Wang, L. Dai, D. Liao, A modified gradient based algorithm for solving Sylvester equations, Appl. Math. Comput., 218 (2012), 5620–5628.
    [21] F. Ding, P. X. Liu, T. Chen, Iterative solutions of the generalized Sylvester matrix equations by using the hierarchical identification principle, Appl. Math. Comput., 197 (2008), 41–50.
    [22] A. Kittisopaporn, P. Chansangiam, Gradient-descent iterative algorithm for solving a class of linear matrix equations with applications to heat and Poisson equations, Adv. Differ. Equ., 2020 (2020), 324. doi: 10.1186/s13662-020-02785-9
    [23] A. Kittisopaporn, P. Chansangiam, W. Lewkeeratiyukul, Convergence analysis of gradient-based iterative algorithms for a class of rectangular Sylvester matrix equation based on Banach contraction principle, Adv. Differ. Equ., 2021 (2021), 17. doi: 10.1186/s13662-020-03185-9
    [24] N. Sasaki, P. Chansangiam, Modified Jacobi-gradient iterative method for generalized Sylvester matrix equation, Symmetry, 12 (2020), 1831. doi: 10.3390/sym12111831
    [25] Y. J. Xie, C. F. Ma, Gradient based and least square based iterative algorithms for matrix equation AXB+CXTB=F, Appl. Math. Comput., 217 (2010), 2191–2199.
    [26] R. A. Horn, C. R. Johnson, Topics in matrix analysis, New York: Cambridge University Press, 1991.
    [27] E. Kreyszig, Introductory functional analysis with applications, New York: John Wiley & Sons, 1978.
    [28] L. Teck, Nonexpansive matrices with applications to solutions of linear systems by fixed point iterations, Fixed Point Theory Appl., 2010 (2009), 821928.
  • This article has been cited by:

    1. Kanjanaporn Tansri, Pattrawut Chansangiam, Conjugate Gradient Algorithm for Least-Squares Solutions of a Generalized Sylvester-Transpose Matrix Equation, 2022, 14, 2073-8994, 1868, 10.3390/sym14091868
    2. Kanjanaporn Tansri, Sarawanee Choomklang, Pattrawut Chansangiam, Conjugate gradient algorithm for consistent generalized Sylvester-transpose matrix equations, 2022, 7, 2473-6988, 5386, 10.3934/math.2022299
    3. Baohua Huang, Changfeng Ma, On the relaxed gradient-based iterative methods for the generalized coupled Sylvester-transpose matrix equations, 2022, 359, 00160032, 10688, 10.1016/j.jfranklin.2022.07.051
    4. Kanjanaporn Tansri, Pattrawut Chansangiam, Gradient-descent iterative algorithm for solving exact and weighted least-squares solutions of rectangular linear systems, 2023, 8, 2473-6988, 11781, 10.3934/math.2023596
    5. Janthip Jaiprasert, Pattrawut Chansangiam, Exact and least-squares solutions of a generalized Sylvester-transpose matrix equation over generalized quaternions, 2024, 32, 2688-1594, 2789, 10.3934/era.2024126
    6. Dimitrios Gerontitis, Panagiotis Tzekis, Solving the generalized Sylvester equation with a novel fast extended neurodynamics, 2024, 0, 2155-3289, 0, 10.3934/naco.2024026
    7. Rui Qi, Caiqin Song, 2023, The full-rank JGI algorithm for the generalized coupled Sylvester-transpose matrix equations, 978-89-93215-27-4, 832, 10.23919/ICCAS59377.2023.10316954
    8. Dimitrios Gerontitis, Panagiotis Tzekis, A New General Fast Neurodynamics (GFN) for Solving Complex Generalized Sylvester Equation With Power Systems Application, 2025, 0170-4214, 10.1002/mma.10911
  • Reader Comments
  • © 2021 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(3091) PDF downloads(106) Cited by(8)

Figures and Tables

Figures(6)  /  Tables(6)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog