1.
Introduction
Quadratic matrix equation has many different forms, such as AX2+BX+C=O arising in quasi-birth-death processes [1,2] and Riccati equation XCX−XE−AX+B=O arising in transport theory [3,4,5,6]. There are also some coupled quadratic matrix equations with two or three variables [7,8]. This article will study the general form of these equations. As can be seen, the linear part of these equations can be expressed in the form ∑3i=1C(l)iXiD(l)i, and the quadratic part of them can be expressed as ∑3i,j=1XiE(l)ijXj. Therefore, we study the following coupled quadratic matrix equation:
where all matrices are n×n real matrices. Each equation in (1.1) consists of three linear terms and nine quadratic terms. Besides, Eq (1.1) have three variables and only two equations, so the solution is not unique.
As we know, we always need some special kind of solutions in practical applications, such as symmetric solutions are widely used in control theory [8,9] and reflexive solutions which are also called generalized centro-symmetric solutions are used in information theory, linear estimate theory and numerical analysis [10,11]. Liao, Liu, etc. have studied the problem of different constrained solutions of linear matrix equations [12,13]. In this article, we will design a method to obtain different constrained solutions of a class of quadratic matrix equations.
Many researchers have studied quadratic matrix equations. For example, Bini, Iannazzo and Poloni gave a fast Newton's method for a quadratic matrix equation [4]. Long, Hu and Zhang used an improved Newton's method to solve a quadratic matrix equation [14]. Convergence rate of some iterative methods for quadratic matrix equations arising in transport theory was also described by Guo and Lin [15]. Zhang, Zhu and Liu have studied the constrained solutions of two-variable Riccati matrix equations based on Newton's method and modified conjugate gradient (MCG) method [16,17]. This article will further study the problem of different constrained solutions of coupled quadratic matrix equations with three matrix variables. The algorithm designed in the paper is superior in computing different constrained solutions.
Notation: Rn×n denotes the set of n×n real matrices. The symbols AT and tr(A) represent the transpose and the trace of the matrix A respectively. A⊗B stands for the Kronecker product of matrices A and B, ¯vec(⋅) is an operator that transforms a matrix A into a column vector by vertically stacking the columns of the matrix AT. For example, for the 2×2 matrix
the vectorization is ¯vec(A)=[a,b,c,d]T. We define an inner product of two matrices A,B∈Rn×n as [A,B]=tr(ATB), then the norm of a matrix A generated by this inner product is Frobenius norm and denoted by ‖A‖, i.e. ‖A‖=√[A,A].
Let Ω1 be the set of symmetric matrices. P1,P2∈Rn×n are said to be symmetric orthogonal matrices if Pi=PTi and P2i=I (i=1,2). X∈Rn×n is said to be a reflexive matrix with respect to P1 if P1XP1=X. Let Ω5 be the set of reflexive matrices. X∈Rn×n is said to be a symmetric reflexive matrix with respect to P2 if XT=X=P2XP2. Let Ω9 be the set of symmetric reflexive matrices. We call (X1,X2,X3) a constrained matrix in Ω1−5−9 when X1∈Ω1, X2∈Ω5 and X3∈Ω9. Besides, if the symmetric orthogonal matrices P1 and P2 are changed, we will get different constrained matrices in Ω1−5−9.
The paper is organized as follows: First, we use Newton's method to convert the quadratic matrix equations into linear matrix equations. Second, MCG method [10,13,16,18] is applied to solve the derived linear matrix equations. Finally, numerical examples are presented to support the theoretical results of this paper.
2.
Newton's method for solving quadratic matrix equations
As a matter of convenience, we first introduce some notations.
Y, Y(k) and Y∗ are defined in the same way as X, X(k) and X∗ respectively. Then let
we can obtain
where ϕ(l)X(Y): Rn×n→Rn×n is the Fréchet derivative of ψ(l)(X) at X in the direction Y [1].
Lemma 2.1. Finding the solution (X∗1,X∗2,X∗3)∈Ω1−5−9 of (1.1) can be transformed into finding the corrected value (Y1,Y2,Y3)∈Ω1−5−9 of ψ(l)(X+Y)=0 (l=1,2). We linearize and solve, to find (Y1,Y2,Y3)∈Ω1−5−9 from the coupled linear matrix equation
Proof. Supposing that the approximate solution (X1,X2,X3)∈Ω1−5−9 of Eq (1.1) has been obtained. Let X∗i=Xi+Yi (i=1,2,3), then finding (X∗1,X∗2,X∗3)∈Ω1−5−9 of (1.1) is transformed into finding the corrected value (Y1,Y2,Y3)∈Ω1−5−9 from
The Eq (2.3) is quadratic equations about Yi. As is known, when the norm of Yi is small enough, the quadratic parts ∑3i,j=1YiE(l)ijYj about Yi in (2.1) can be discarded according to Newton's method. In this way, we can get a linear approximation
Therefore, finding the solution (X∗1,X∗2,X∗3)∈Ω1−5−9 of (1.1) is transformed into finding (Y1,Y2,Y3)∈Ω1−5−9 from ψ(l)(X)+ϕ(l)X(Y)=O (l=1,2), that is, to solve (2.2).
According to [14], Newton's method (algorithm 1) is introduced to find constrained solutions in Ω1−5−9 of (1.1). Let
The convergent properties about Newton's method can be obtained as follows according to [14] (The proof is similar to Lemma 2.1 in [14]).
Theorem 2.1. Assume that the real matrix X∗ is a simple root of (1.1), i.e. ψ(X∗)=O and ϕX∗(Y) is regular. Then if the starting matrix X(1) is chosen sufficiently close to the solution X∗, the sequence {X(k)} generated by Newton's method converges quadratically to the solution X∗.
3.
MCG method for solving linear matrix equations
In algorithm 1, when X(k) is known, then Y(k) needs to be solved. In this section, MCG method will be used to solve Y(k) from Eq (2.2), that is, to solve Eq (2.4) or Eq (2.5). Consider the general form of Eq (2.2)
where all matrices in Eq (3.1) are n×n real matrices. Let
where
and P1,P2∈Rn×n are symmetric orthogonal matrices.
In order to solve Eq (3.1), the following two questions will be considered.
Problem 3.1. Assume that (3.1) has constrained solutions, find (Y1,Y2,Y3)∈Ω1−5−9 from (3.1).
Problem 3.2. Assume that (3.1) hasn't constrained solutions, find (Y1,Y2,Y3)∈Ω1−5−9, such that
3.1. MCG method for solving problem 1
Based on the MCG method, we establish the following algorithm (algorithm 2) to solve problem 3.1.
From algorithm 2, we can easily see (Y(k)1,Y(k)2,Y(k)3)∈Ω1−5−9 for k=1,2,⋯ and have the following convergent properties (The proof is similar to Theorem 2.1 in [10]):
Theorem 3.1. Assume that Eq (3.1) has constrained solutions in Ω1−5−9. Then for an arbitrary initial matrix (Y(1)1,Y(1)2,Y(1)3)∈Ω1−5−9, a solution of problem 3.1 can be obtained by algorithm 2 within finite number of iterations, which is also a constrained solution in Ω1−5−9 of (3.1).
3.2. MCG method for solving problem 2
Algorithm 2 will break if Ri≠O and Zi=O, which means that Eq (3.1) hasn't constrained solution in Ω1−5−9 according to Theorem 3.1. Therefore, we need to solve problem 3.2, that is, to find constrained least-squares solutions of (3.1).
We replace the problem of finding least-squares solutions in Ω1−5−9 of (3.1) with finding solutions in Ω1−5−9 of equivalent linear matrix equations by the Theorem 3.2, and then an iterative algorithm to find constrained least-squares solutions in Ω1−5−9 of (3.1) is constructed according to algorithm 2.
As a matter of convenience, we introduce some notations:
where
u and v are functions of Y. Then, according to [16,17], we have the following theorem.
Theorem 3.2. Iterative algorithm for solving problem 3.2 can be replaced by finding constrained solutions in Ω1−5−9 from
Indeed, (3.3) has constrained solutions in Ω1−5−9.
Proof. When (Y1,Y2,Y3)∈Ω1−5−9, we have Y1=YT1, Y2=P1Y2P1 and Y3=YT3=P2Y3P2. Therefore, solving problem 3.2 is equivalent to solving (Y1,Y2,Y3)∈Ω1−5−9 from
Now we have to prove that solving the problem (3.4) is equivalent to finding constrained solutions in Ω1−5−9 of (3.3). We let multiply operation prior to Kronecker product operation between matrices. Let
where Tn,n denotes a commutation matrix such that Tn,n¯vec(An×n)=¯vec(AT) [19] and let Tn,n only work on ¯vec. Then applying ¯vec to the following equations:
we can get the equivalent equation: My=f. Besides, MTMy=MTf, the normal equation of My=f, is the vectorization of (3.3). Therefore, the least-squares solution of My=f is also a solution of MTMy=MTf, and the vectorization of the solution of (3.3). So the solution of (3.4) is also a solution of (3.3), and vice versa.
Above all, iterative algorithm for solving problem 3.2 can be replaced by finding constrained solutions in Ω1−5−9 of (3.3).
As we all know, normal equations always have solutions, and the vectorization of Eq (3.3) is a normal equation, so Eq (3.3) also has solutions. Suppose ˜Y=(˜YT1,˜YT2,˜YT3)T (whether ˜Y∈Ω1−5−9 or not) is a solution of (3.3), then g(˜Y)=Q. Let Y∗i=qi(˜Y) (i=1,2,3), then (Y∗1,Y∗2,Y∗3)∈Ω1−5−9 and g(Y∗)=Q. Hence, (3.3) has constrained solutions in Ω1−5−9.
We use the MCG method to find constrained solutions in Ω1−5−9 of (3.3) by algorithm 3, which is also a process to solve the problem 3.2.
From algorithm 3, we can see that (Y(k)1,Y(k)2,Y(k)3)∈Ω1−5−9 for k=1,2,⋯ and have the following convergent properties (The proof is similar to Theorem 2 in [13]).
Theorem 3.3. For an arbitrary initial matrix (Y(1)1,Y(1)2,Y(1)3)∈Ω1−5−9, a solution of problem 3.2 can be obtained by algorithm 3 within finite number of iterations, and it is also a constrained least-squares solution in Ω1−5−9 of (3.1).
4.
Numerical experiments
In this section, we design two computation programmes to find constrained solutions of (1.1). Then two numerical examples are given to illustrate the proposed results. All computations are performed using MATLAB. Because of the influence of roundoff errors, we regard a matrix A as zero matrix if ‖A‖≤10−7.
Let n be the order of the matrix Xi, k,k1,k2 be the iteration numbers of algorithm 1, algorithm 2 and algorithm 3 respectively, and t be the computation time (seconds).
Programme 1.
(1) Choose an initial matrix (X(1)1,X(1)2,X(1)3)∈Ω1−5−9 and set k:=1.
(2) If ψ(X(k))=O, stop, else, solve for (Y(k)1,Y(k)2,Y(k)3)∈Ω1−5−9 from (2.4) using algorithm 2. When algorithm 2 breaks, that is (2.4) hasn't constrained solution in Ω1−5−9, solve for (Y(k)1,Y(k)2,Y(k)3)∈Ω1−5−9 from (2.5) using algorithm 3.
(3) Compute X(k+1)=X(k)+Y(k), set k:=k+1 and go to step 2.
Programme 2.
(1) Choose an initial matrix (X(1)1,X(1)2,X(1)3)∈Ω1−5−9 and set k:=1.
(2) If ψ(X(k))=O, stop, else, solve for (Y(k)1,Y(k)2,Y(k)3)∈Ω1−5−9 from (2.5) using algorithm 3. Especially, when (2.4) has constrained solutions in Ω1−5−9, the constrained least-squares solutions in Ω1−5−9 are also its constrained solutions in Ω1−5−9.
(3) Compute X(k+1)=X(k)+Y(k), set k:=k+1 and go to step 2.
Example 4.1. Consider (1.1) with the following parameters:
where
We can easily see that (1.1) has the constrained solution (X∗1,X∗2,X∗3)∈Ω1−5−9. By applying programmes 1 and 2 with the initial matrix X(1)i=eye(3), Y(1)i=zeros(3)(i=1,2,3), we obtain the constrained solution in Ω1−5−9 of (1.1) as follows:
The iteration numbers and computation time are listed in Table 1.
From the results in Table 1, we see that programme 1 is more effective when the derived linear matrix equations are always have constrained solutions in Ω1−5−9.
Example 4.2. Consider (1.1) with the following parameters:
where
By applying programmes 1 and 2 with the initial matrix X(1)i=ones(3), Y(1)i=zeros(3) (i=1,2,3), P1 and P2 are identity matrices, we obtain a special constrained solution (now, X1 and X3 are symmetric matrices, X2 is a general matrix) in Ω1−5−9 of (1.1) as follows:
When X(1)i=eye(3) (i=1,2,3), others remain unchanged, we obtain another special constrained solution in Ω1−5−9 of (1.1) as follows:
Thus it can be seen that if the constrained solution of Eq (1.1) is not unique, we can get different constrained solutions in Ω1−5−9 when choosing different initial matrices.
5.
Conclusions
In this paper, an iterative algorithm is studied to find different constrained solutions. By using the proposed algorithm, we compute a set of different constrained solution in Ω1−5−9 of multivariate quadratic matrix equations. The provided examples illustrate the effectiveness of the new iterative algorithm.
There are still some results we can obtain by changing the initial matrices X(1)i and Y(1)i, the direction matrix Zk in algorithm 2 and Eq (3.3) in algorithm 3. In this way, we can get other kind of constrained solutions, which are not only interesting but also valuable. It remains to study in our further work.
Acknowledgments
This research was funded by Doctoral Fund Project of Shandong Jianzhu University grant number X18091Z0101.
Conflict of interest
The author declares no conflict of interest.