Given matrices N∈Cs×s and S0,…,Sq∈Cs×s, we solve the linear differential equation
q∑n=0Tn(t) (d/dt)nf(t)=g(t),
where t∈R, Tn(t)=etNSne−tN, and f(t):R→Cs, using the roots of d(ν)=det D(ν), where
D(ν)=q∑n=0Sn (νIr+N)n.
For example,
N=(0−110)
implies
etN=(cost−sintsintcost),
so that Tn(t) are periodic, giving an explicit solution to a form of Floquet's theorem.
Citation: Christopher Withers, Saralees Nadarajah. Some linear differential equations generated by matrices[J]. AIMS Mathematics, 2022, 7(6): 9588-9602. doi: 10.3934/math.2022533
[1] | Yajun Xie, Changfeng Ma, Qingqing Zheng . On the nonlinear matrix equation $ X^{s}+A^{H}F(X)A = Q $. AIMS Mathematics, 2023, 8(8): 18392-18407. doi: 10.3934/math.2023935 |
[2] | Hasan Gökbaş . Some properties of the generalized max Frank matrices. AIMS Mathematics, 2024, 9(10): 26826-26835. doi: 10.3934/math.20241305 |
[3] | Xu Li, Rui-Feng Li . Shift-splitting iteration methods for a class of large sparse linear matrix equations. AIMS Mathematics, 2021, 6(4): 4105-4118. doi: 10.3934/math.2021243 |
[4] | Yige Zhao, Yibing Sun, Zhi Liu, Yilin Wang . Solvability for boundary value problems of nonlinear fractional differential equations with mixed perturbations of the second type. AIMS Mathematics, 2020, 5(1): 557-567. doi: 10.3934/math.2020037 |
[5] | J. Alberto Conejero, Antonio Falcó, María Mora–Jiménez . A pre-processing procedure for the implementation of the greedy rank-one algorithm to solve high-dimensional linear systems. AIMS Mathematics, 2023, 8(11): 25633-25653. doi: 10.3934/math.20231308 |
[6] | Efruz Özlem Mersin . Sturm's Theorem for Min matrices. AIMS Mathematics, 2023, 8(7): 17229-17245. doi: 10.3934/math.2023880 |
[7] | Pattrawut Chansangiam, Arnon Ploymukda . Riccati equation and metric geometric means of positive semidefinite matrices involving semi-tensor products. AIMS Mathematics, 2023, 8(10): 23519-23533. doi: 10.3934/math.20231195 |
[8] | Kahraman Esen Özen . A general method for solving linear matrix equations of elliptic biquaternions with applications. AIMS Mathematics, 2020, 5(3): 2211-2225. doi: 10.3934/math.2020146 |
[9] | Qin Zhong, Ling Li . Notes on the generalized Perron complements involving inverse $ {{N}_{0}} $-matrices. AIMS Mathematics, 2024, 9(8): 22130-22145. doi: 10.3934/math.20241076 |
[10] | Xiaofeng Guo, Jianyu Pan . Approximate inverse preconditioners for linear systems arising from spatial balanced fractional diffusion equations. AIMS Mathematics, 2023, 8(7): 17284-17306. doi: 10.3934/math.2023884 |
Given matrices N∈Cs×s and S0,…,Sq∈Cs×s, we solve the linear differential equation
q∑n=0Tn(t) (d/dt)nf(t)=g(t),
where t∈R, Tn(t)=etNSne−tN, and f(t):R→Cs, using the roots of d(ν)=det D(ν), where
D(ν)=q∑n=0Sn (νIr+N)n.
For example,
N=(0−110)
implies
etN=(cost−sintsintcost),
so that Tn(t) are periodic, giving an explicit solution to a form of Floquet's theorem.
There is much research of linear differential equations involving or generated by matrices, see Derevenskii [2] and Elishevich [3]. But these and other papers are limited to the first few orders. We are not aware of papers giving solutions for linear differential equations of general order generated by matrices. In the paper, we provide solutions for such a class of linear differential equations.
Let R and C denote the real and complex numbers, respectively. For any function f(t) of t∈R, set f⋅n(t)=(d/dt)nf(t). Let Y(t)=Y=(Y1,…,Yr)∈Cs×r be a function of t∈R, and a constant N=(Nj,k)∈Cr×r, such that
Y⋅1(t)=Y(t)N∈Cs×r. | (1.1) |
Choose q≥1. Suppose that for 0≤n≤q, and Tn(t)∈Cs×s, there is a constant Sn=(Sn,j,k)∈Cr×r such that
Tn(t) Y(t)=Y(t) Sn. | (1.2) |
If r=s, as in the abstract, we can take
Y(t)=etN, Tn(t)=etN Sn e−tN. | (1.3) |
If r>s, we can take Y(t) as the first s rows of etN. We wish to solve the differential equation Lf=0s for f=f(t):R→Cs, where
L=L(t)=q∑n=0Tn(t) (d/dt)n, s×s, |
that is,
Lf=q∑n=0Tnf⋅n. | (1.4) |
Theorem 2.1 gives solutions in terms of the roots of d=d(ν), where
d(ν)=det D(ν) and D(ν)=D=q∑n=0Sn (νIr+N)n∈Cr×r. | (1.5) |
We call D(ν) the characteristic matrix of the operator L. If ν is a multiple root, then other solutions are given by Theorem 2.2.
Section 3 chooses r=s=2 and N of (3.1) below. Tn(t) of (1.3) are then linear in cos2t and sin2t and so have period π. Example 3.2 gives for the first time the full solution of a well known example. Example 3.3 solves a differential equation that arises in the theory of planetary perturbations, and was the incentive for this paper. Section 4 shows how D(ν) and Tn(t) are given by the Jordan form of N when r=s.
Section 5 solves Lf=g, any given function in Cs, when r=s and Tq=Is, by converting it to the standard form X⋅1=AX+F. Its solution is then given by the variation of constants formula in terms of the solution of Lf=0s. When A is periodic, Floquet's theorem only gives the form of the solution of X⋅1=AX. We give the actual solution. Examples 3.1-3.5 all have periodic A. Set i=√−1. Set δj,k=1 or 0 for j=k or j≠k.
Theorem 2.1. Suppose that Y(t), N satisfy (1.1), and thatTn(t), Sn satisfy (1.2) for 0≤n≤q.Let ν be any of the qr roots of d(ν)=0for d(ν)of (1.5).Choose a(ν)=a=(a1,…,ar)′∈Cr such that E(ν)=0r, where
E=E(ν)=D(ν)a(ν). |
Then for f=f(t)∈Cs andL of (1.4), a solution of Lf=0s is
f0(t,ν)=eνtY(t)a(ν)=eνtr∑k=1ak(ν)Yk(t). | (2.1) |
Proof: Since Y⋅n=YNn, by Leibniz' rule, the nth derivative of eνtY is
(eνtY)⋅n=eνtY (νIr+N)n. | (2.2) |
So,
L eνtY=q∑n=0Tn(t) (eνtY)⋅n=eνtY D(ν), |
and L f0=eνtY D(ν)a(ν)=0r.
When the roots of d(ν) are distinct, this gives qr independent solutions of (2.1).
Theorem 2.2. Take ν, D=D(ν), a=a(ν) and E=E(ν) of Theorem 2.1.Suppose that for some k≥1,
E⋅m(ν)=0r | (2.3) |
for 0≤m<k, where E⋅m(ν)=∂mνE(ν), and ∂ν=d/dν.Then for 0≤n<k, Lf=0s has solutions
fn=fn(t,ν)=eνtY(t)zn=eνtr∑j=1zn,j(t,ν)Yj(t), | (2.4) |
where zn=zn(t,ν)=(t+∂ν)n a(ν).
Proof: By (2.2) and Leibniz' rule,
∂ot(eνtYzn)=o∑m=0p(om)eνtY (νIs+N)o−m∂mtzn,Lfn=eνtYGn, |
where
Gn=q∑o=0Soo∑m=0(om)(νIs+N)o−m∂mt(t+∂ν)n a=∑m,r(nm)(n−mr)tn−m−rD⋅ma⋅r. |
Transform from m to c=m+r. Then (nm)(n−mr)=(nc)(cr). So,
Gn=∑c(nc)tn−cHc, |
where
Hc=∑r(cr)D⋅c−ra⋅r=∂cνE(ν)=0s |
by (2.3).
We call fn a characteristic solution of Lf=0s. D(ν) and d(ν) expand as
D(ν)=q∑k=0νkDk, d(ν)=rq∑k=0νkdk, |
where
Dk=q∑n=k(nk)SnNn−k, D0=D(0), Dq=Sq,d0=det D(0), dq=det Sq. |
So,
D⋅m(ν)=q∑k=m(k)mνk−mDk, d⋅m(ν)=rq∑k=m(k)mνk−mdk. | (2.5) |
Corollary 2.1. Take Lf of (1.4).Consider the exceptional case when D(ν)≡0s×s.Suppose that
D⋅m(ν)=0s×s |
for 0q≤m≤n.Then for any a∈Cs, a characteristic solution of Lf=0s is
fn(t,ν)=tneνtY(t) a. |
We now transfer the condition (2.3) from E(ν) to d(ν). Let M be the adjoint of D: (−1)j+kMk,j is the determinant of D with its jth row and kth column deleted. If d≠0, then M=dD−1.
Corollary 2.2. Let e1,s,…,e2,s be any basis for Rs.For D and d of (1.5), set a(j)=Mej,s and E(j)=Da(j)=dej,s.So, for m∈Zq, E(j)⋅m=d⋅mej,s.Choose ν so that d=0.Given 1≤j≤s and a=a(j), fn,j(t)=fn(t,ν) of (2.4) is a characteristic solution of Lf=0s if
d⋅m(ν)=0 | (2.6) |
for 0≤m≤n.
By (2.5), (2.6) does not extend to n=rq if dr,q≠0. We now take ej,s as the jth unit vector in Rs. So, for 1≤j≤s, a(j) is the jth column of M. For 1≤k≤s, its kth element is a(j)k=Mk,j=aj,k say. Corollary 2.2 breaks the solution fn(x,ν) into s basis solutions fn,j(x), 1≤j≤s, of Lf=0s. For example, if s=2 then
d=D1,1D2,2−D1,2D2,1, M=(D2,2−D1,2−D2,1D1,1),e1,2=(1,0)′, e2,2=(0,1)′, a(1)=(D2,2,−D2,1)′, a(2)=(−D1,2,D1,1)′,d=(D2,2,D1,2)a(1)=(D2,1,D2,2)a(2). | (2.7) |
d=0 implies
f0,1=eνt(D2,2Y1−D2,1Y2), f0,2=eνt(−D1,2Y1+D1,1Y2). | (2.8) |
d=d⋅1=0 implies
f1,1=eνt[(tD2,2+D2,2⋅1)Y1−(tD2,1+D2,1⋅1)Y2], | (2.9) |
f1,2=eνt[−(tD1,2+D1,2⋅1)Y1+(tD1,1+D1,1⋅1)Y2]. | (2.10) |
d=d⋅1=d⋅2=0 implies
f2,1=eνt[(t2D2,2+2tD2,2⋅1+D2,2⋅2)Y1−(t2D2,1+2tD2,1⋅1+D2,1⋅2)Y2], | (2.11) |
f2,2=eνt[−(t2D1,2+2tD1,2⋅1+D1,2⋅2)Y1+(t2D1,1+2tD1,1⋅1+D1,1⋅)Y2]. | (2.12) |
d⋅m=0 for 0≤m≤3 implies f3,j=eνt(z3,1Y1+z3,2Y2), where, for j=1,
z3,1=t3D2,2+3t2D2,2⋅1+3tD2,2⋅2+D2,2⋅3, | (2.13) |
z3,2=−t3D2,1−3t2D2,1⋅1−3tD2,1⋅2−D2,1⋅3, | (2.14) |
and, for j=2,
z3,1=−t3D1,2−3t2D1,2⋅1−3tD1,2⋅2−D1,2⋅3,z3,2=t3D1,1+3t2D1,1⋅1+3tD1,1⋅2+D1,1⋅3. |
Generally, each fn,j is a linear combination of (fm,1, 0≤m≤n). We shall give details in a later paper. For example, M1,1≠0 implies, for 2≤j≤s, f0,j(t)=f0,1(t)Mj,1/M1,1. s=2 and D2,2≠0 imply f0,2(t)=−f0,1(t)D2,1/D2,2.
For L(t)=L of (1.4) and τ≠0, set
Lτ(t)=τqL(τt)=q∑n=0Tτ,n(t) (d/dt)n, |
where Tτ,n(t)=τq−nTn(τt). So, Tq=Is implies Tτ,q=Is. For example, q=s=2, T2=I2 imply Lτ(t)=(d/dt)2+τT1(τt)+τ2T0(τt).
Corollary 2.3. Take ν, D=D(ν), a=a(ν) of Theorems 2.1-2.2, andf(t)=f0(t,ν) of (2.1), or f(t)=fn(t,ν) of Theorem 2.2.Then a solution of Lτ(t)X(t)=0s is X(t)=fn(τt,ν).
We have not assumed that r=s or Tq=Is. However, if r=s and Tq=Is, then Lf=0s can be written in the standard form X⋅1=AX, where
X=(ff⋅1⋯f⋅q−1), A=(0Is0⋯000Is⋯0⋯000⋯Is−T0−T1−T2⋯−Tq−1)∈Cqs×qs, | (2.15) |
(to be read as A=−T0 if q=1), and each 0 is s×s. So, (2.15) with f of Theorems 2.1, 2.2 give all qs linearly independent solutions of X⋅1=AX.
Set ct=cost, st=sint. Here, we take r=s=2 and
N=(0−110), Y=(Y1,Y2)=etN=(ct−ststct)=ctI2+st. | (3.1) |
So,
Y1=(ctst), Y2=Y1⋅1=(−stct), NY1=Y2, NY2=−Y1. |
Set
Λ=diag (1,−1), J=(0110),R(t)=ctΛ+stJ=(Y1,−Y2)=(ctstst−ct),Q(t)=(−stctctst)=(Y2,Y1)=YJ=R⋅1(t). |
N, R(t), Q(t), Λ and J all have determinant ±1. Some properties are:
J2=Λ2=−N2=R(t)2=Q(t)2=I2,NΛ=−ΛN=J, JN=−NJ=Λ, JΛ=−ΛJ=N,R(t)Y(s)=R(t−s)=Y(t−s)Λ, R(2t)Y=YΛ=R(t),R(2t)Y1=Y1, R(2t)Y2=−Y2, ΛY=R(−t), Y(s)R(t)=R(s+t),JR(t)=¯YN=N¯Y, R(t)Λ=Y, ΛR(t)=¯Y, R(2t)=YΛY′,I2+Λ=2(1000), J−N=2(0100), J+N=2(0010), I2−Λ=2(0001). |
(1.3) holds for
(Sn,Tn)=(I2,I2), (N,N), (Λ,R(2t)), (J,Q(2t)). |
So, (1.3) also holds for any linear combination of these, say
Sn=bn,1I2+bn,2N+bn,3Λ+bn,4J=(bn,1+bn,3bn,2−bn,4−bn,2−bn,4bn,1−bn,3), | (3.2) |
Tn(t)=bn,1I2+bn,2N+bn,3R(2t)+bn,4Q(2t)=(bn,1+bn,3c2,t−bn,4s2,t−bn,2+bn,3s2,t+bn,4c2,tbn,2+bn,3s2,t+bn,4c2,tbn,1−bn,3c2,t+bn,4s2,t) | (3.3) |
for any constants bn,j∈C. Any 2×2 matrix Sn can be put in this form: set
bn,1=(Sn,1,1+Sn,2,2)/2, bn,3=(Sn,1,1−Sn,1,2)/2,bn,2=(Sn,2,1−Sn,1,2)/2, bn,4=−(Sn,2,1+Sn,1,2)/2. |
So,
Tn(t)=At(Sn), |
where
2At(S)=B(S)+c2,t C(S)+s2,t G(S),B(S)=(B1−B2B2B1), C(S)=(C1−C2−C2−C1), G(S)=(C2C1C1−C2),B1=S1,1+S2,2, B2=S1,2−S2,1, C1=S1,1−S2,2, C2=S1,2+S2,1,TnY1=(bn,1+bn,3)Y1+(bn,2+bn,4)Y2, TnY2=(−bn,2+bn,4)Y1+(bn,1−bn,3)Y2. |
Corollary 3.1. For Tn(t) of (3.3), a solution of Lf=02 is (2.1) withν any of the 2q roots of d=0,
E=02, D=c1I2+c2N+c3Λ+c4J=(c1+c3−c2+c4c2+c4c1−c3), | (3.4) |
where
cj=q∑n=0cn,j, d=c21+c22−c23−c24, | (3.5) |
where
cn,1=gnbn,1−hnbn,2, cn,2=gnbn,2+hnbn,1, cn,3=gnbn,3+hnbn,4, cn,4=gnbn,4−hnbn,3, | (3.6) |
gn=gn(ν)=Real((ν+i)n)=∑j(n2j)(−1)jνn−2j, | (3.7) |
hn=hn(ν)=Imag((ν+i)n)=∑j(n2j+1)(−1)jνn−2j−1, | (3.8) |
where the real and imaginary parts are taken as if ν were real.
Proof: Since N2=−I2,
(νI2+N)n=gnI2+hnN |
for gn and hn of (3.7), (3.8). So, for Sn of (3.2),
Sn (νI2+N)n=cn,1I2+cn,2N+cn,3Λ+cn,4J. |
So, by (1.5), (2.5) holds. gn, hn, cn,j, cj, D, d are polynomials in ν of degree n, n−1, n, q, q, 2q. So, d=0 has 2q roots ν.
By (3.7)-(3.8), when q=1, the cj needed for (3.5) are
c1=b0,1+νb1,1−b1,2, c2=b0,2+νb1,2+b1,1,c3=b0,3+νb1,3+b1,4, c4=b0,4+νb1,4−b1,3. |
By (3.7)-(3.8), when q=2, the cj needed for (3.5) are
c1=b0,1+νb1,1−b1,2+(ν2−1)b2,1−2νb2,2,c2=b0,2+νb1,2+b1,1+(ν2−1)b2,2+2νb2,1,c3=b0,3+νb1,3+b1,4+(ν2−1)b2,3+2νb2,4,c4=b0,4+νb1,4−b1,3+(ν2−1)b2,4−2νb2,3. |
When q=3, we add c3,j of (3.6) to cj for j=1,2,3,4.
Lf=02 can only be reduced to the form X⋅1=AX if eq≠0, where en=det Tn(t)=det Sn=b2n,1+b2n,2−b2n,3−b2n,4, since then we can reduce Tq(t) to I2 by multiplying by Tq(t)−1. Set
¯L=Tq(t)−1L=q∑n=0¯Tn(t) (d/dt)n, |
where ¯Tn(t)=Tq(t)−1Tn(t) is a linear combination of 1, s2,t, c2,t, s4,t, c4,t.
Example 3.1. Take q=1, S1=I2, S0=λ Λ. So, (1.2) holds with T1=I2 and T0(t)=λR(2t). Further,
D=νI2+λ Λ+N=(ν+λ−11ν−λ), |
and d=ν2−λ2+1 with roots ν=±(λ2−1)1/2=ν1, ν2, say. By (2.7), a=a(1)=(ν−λ−1) implies a⋅1=(10), z0=(ν−λ−1) and z1=(t(ν−λ)+1−t); a=a(2)=(1ν+λ) implies a⋅1=(01), z0=(1ν+λ) and z1=(tt(ν+λ)+1); d=0 implies ν=±(1−λ2)1/2. So, by (2.8), solutions are
f0,1=eνt[(ν−λ)Y1−Y2], f0,2=eνt[Y1+(ν+λ)Y2]. |
If d=0 and λ=±1, then d⋅1=ν=0 so by (2.9) and (2.10) solutions are
f0,1=−λY1−Y2, f0,2=Y1+λY2=−λf0,1,f1,1=(1−λt)Y1−tY2, f1,2=tY1+(1+λt)Y2. |
An extension is
Example 3.2. Given scalars b0, b1, b2, we solve f⋅1(t)=A(t)f(t) for
A(t)=(b0+b1c2,t−b2+b1s2,tb2+b1s2,tb0−b1c2,t)=B+b1R(2t),B=b0I2+b2N=(b0−b2b2b0). | (3.9) |
So, Lf=f⋅1−A(t)f. Take q=1, S1=I2 and
S0=(−b0−b1b2−b2−b0+b1)=−b0I2−b1Λ−b2N=−B−b1Λ. |
Then (1.4) holds with T1(t)=I2, T0(t)=−A(t). So,
D(ν)=S0+νI2+N=(ν−b0−b1b2−1−b2+1ν−b0+b1), |
and d(ν)=(ν−b0)2−b21+(b2−1)2 has roots
ν1=b0+δ1/2, ν2=b0−δ1/2 | (3.10) |
for δ=b21−(b2−1)2. By (2.8), if ν=ν1 or ν2, then d=0 and solutions are
f0,1(t)=eνt[(ν−b0+b1)Y1+(b2−1)Y2],f0,2(t)=eνt[(1−b2)Y1+(ν−b0−b1)Y2]. | (3.11) |
Consider the case δ=0. So, b2=1+λb1, where λ=±1,
f0,1(t)=b1eb0t(Y1+λY2), f0,2(t)=−b1eb0t(λY1+Y2),f1,1(t)=eb0t[(b1t+1)Y1+λb1tY2], f1,2(t)=eb0t[−λb1tY1+(1−b1t)Y2]. |
Let us rescale this example by transforming to T=t/τ, x(T)=f(τT) for τ≠0, then replacing T by t.
Example 3.3. Set Aτ(t)=τA(τt). For A(t) of (3.9), x⋅1(t)=Aτ(t)x(t) has solutions f0,j(τt) for f0,j(t) of (3.11) with ν of (3.10). If δ=0, other solutions are f1,j(τt) for f1,j(t) of (2.9) and (2.10). We consider two cases. The first case is that τ=1/2. In this case,
Aτ(t)=(b0+b1ct−b2+b1stb2+b1stb0−b1ct)/2=(b0I2+b2N+b1R(2t))/2. |
The second case is that τ=−1, b0=1/4, b2=1 and b1=−3/4. Then ν1=−1/2, ν2=1 and independent solutions are
f0,1(−t)=(−3/2)et/2(ct−st), f0,2(−t)=(3/2)e−t(stct). |
The other two f0,j are 02. In this case, Aτ(t)=−B−b1R(−2t) can be written
Aτ(t)=(−1+(3/2)c2t1−(3/2)ctst,−1−(3/2)ctst−1+(3/2)s2t), |
with period T=π. Markus and Yamabe [5] used this form of Aτ(t) but only gave the first solution f0,1(−t). This example is quoted by Chicone [1], but again the second solution f0,2(−t) is not given.
Example 3.4. Take q=2, S2=I2, S1=02×2 and S0=diag (b0+b1,b0−b1) for scalars b0, b1. Then T2=I2, T1=02×2 and
T0=b0I2+b1R(2t)=(b0+b1c2,tb1s2,tb1s2,tb0−b1c2,t). |
Then, L=I2(d/dt)2+T0. (1.2) holds with
D=(νI2+N)2+S0=(ν2−1+b0+b1−2ν2νν2−1+b0−b1),d=(ν2−1+b0)2−b21+4ν2=ν4+2bν2+c, |
where b=b0+1, c=(b0−1)2−b21. d has four roots, ν=±ν1, ±ν2, where
ν1=(−b+δ1/2)1/2, ν2=(−b−δ1/2)1/2, δ=b2−c=4b0+b21. | (3.12) |
So, ν2−1+b0=−2±δ1/2 and solutions are given by f0,1, f0,2 of (2.8) with D2,2=ν2−1+b0−b1, D2,1=2ν, D1,2=−2ν and D1,1=ν2−1+b0+b1. If δ of (3.12) is 0 and b≠0, then there are two roots of multiplicity two, ν=λν0, where λ=±1, ν0=(−b)1/2; so other solutions are f1,1 of (2.9) with ν=λν0,
D2,2=−2−b1, D2,2⋅1=D2,1=2ν, D2,1⋅1=2, | (3.13) |
and f1,2 of (2.10) with ν=λν0,
D1,2=−2ν, D1,2⋅1=−2, D1,1=b1−2, D1,1⋅1=2ν. | (3.14) |
Now suppose that b0=−1, b1=2λ, where λ=±1. Then ν=0 has multiplicity 4. So, other solutions are f2,1 of (2.11), (3.13) with D2,2⋅2=2, D2,1⋅2=0; f2,2 of (2.12), (3.14) with D1,2⋅2=0, D1,1⋅2=2; f3,1 of (2.13) with
z3,1=−(2+b1)t3+6λν0t2+6t, z3,2=2λν0t3−6t2; |
and f3,2 of (2.14) with
z3,1=2λν0t3+6t2, z3,2=(b1−2)t3+6λν0t2+6t. |
To solve Lf=g, a given function in C2, Section 5 will need its derivative, ∂t f0(t,ν)=eνtv(t,ν), where
v(t,ν)=2∑j=1aj (νYj+Yj⋅1)=a1 (νY1+Y2)+a2 (νY2−Y1)=2∑j=1ejYj,e1=a1ν−a2=−D1,2ν−D1,1=ν2+1−b0−b1,e2=a1+νa2=−D1,2+νD1,1=ν(ν2+1+b0+b1). |
By (2.15), Lf=02 can be written X⋅1=AX, where
X=(ff⋅1), A=(02×2I2−T002×2)∈C4×4. |
Now suppose that b0, b1 are real. If δ<0, all four roots are complex. Suppose that δ≥0. Then ±ν1 are real if b≤δ1/2, and ±ν2 are real if δ1/2≤−b. So, if b≤δ1/2≤−b then all four roots are real.
The differential equation, Lf=02, arises in the theory of planetary perturbations with b0=−γ/2, b1=−3γ/2, so that b=1−γ/2, c=−2γ2, δ=9γ(γ−8/9)/4, and γ>0 real. So, δ≥0 if and only if γ≥8/9. Also γ≥1 if and only if ν21≥0. For γ≥8/9, ν22<0. So, γ<8/9 implies four complex roots; 8/9≤γ<1 implies four imaginary roots; 1≤γ implies ν21≥0>ν22 so that two roots are real and two are imaginary. For large γ, ν1 and ν2 of (3.12) satisfy
ν1=(2γ)1/2[1−5γ−1/12+O(γ−2)], ν2=iγ1/2[1+γ−1/6+O(γ−2)]. |
By Corollary 2.3, we have
Example 3.5. For L of Example 3.4, Lτ(t)=(d/dt)2+τ2T0(τt), and Lτ(t)X(t)=02 has solution X(t)=fn,j(τt) when fn,j(t) is a solution to Example 3.4.
Suppose that r=s. Then D and Tn(t) can be easily written using the Jordan form of N, in terms of block matrices, or scalars when the Jordan form is diagonal. Suppose that N has Jordan form N=PJP−1, where J=diag (J1,…,Jp), Jj=Jmj(λj), Jm(λ)=λIm+Um, and Um is the m×m matrix of zeros except for ones on the first super-diagonal, that is, (Um)j,k=δk,j+1 for 1≤j<m. So, s=m1+⋯+mp and for 0≤n<m, Unm is the m×m matrix of zeros except for ones on the nth super-diagonal, that is, (Unm)j,k=δk,j+n for 1≤j<m−n. Also for n≥m, Unm=0m×m, and
Jm(λ)n=min(n,m−1)∑j=0(nj)λn−jUjm, | (4.1) |
exp{Jm(λ)t}=eλtVm(t), | (4.2) |
where
U0m=Im, Vm(t)=m−1∑j=0tjUjm/j!. |
So, Vm(t) has zeros on its subdiagonals, and the elements of its jth superdiagonal are all tj/j. That is, Vm(t)j,j+n=tj/j! for 1≤j<m−n. For example,
U2=(0100), J2(λ)=(λ10λ), J2(λ)n=(λnnλn−10λn),exp{J2(λ)t}=eλtV2(t), V2(t)=(1t01). |
Partition P, P−1, Sn and Tn=Tn(t) of (1.3) as mj×mk blocks, P=(Pj,k), P−1=(Pj,k), Sn=(Sn,j,k), Tn(t)=(Tn(t)j,k) for 1≤j, k≤p, and do similarly for D, Nn and Y=Y(t)=etN. Then
(Nn)j,k=p∑c=1Pj,cJncPc,k, Yj,k=p∑c=1Pj,cetJcPc,k,Tn(t)j,k=p∑a,b=1Yj,a(t)Sn,a,bYb,k(−t)=p∑c,d=1Pj,cetJcQcd,ne−tJdPd,k, |
where
Qcd,n=p∑a,b=1Pc,aSn,a,bPb,d,D=q∑n=0SnP diag (Mn,1,…,Mn,p) P−1=(Dj,k), Mn,j=Jmj(ν+λj)n of (4.1),Dj,k=q∑n=0p∑b,c=1Sn,j,bPb,cJmc(ν+λc)n Pc,k, |
and etJc is given by (4.2) with m=mc, λ=λc. So, Tn(t) is a mixture of polynomials in t and factors e(λc−λd)t. For example, if P=Is, then
(Nn)j,k=Jnjδj,k, Yj,k(t)=etJjδj,k, Tn(t)j,k=etJjSn,j,ke−tJk, Dj,k=q∑n=0Sn,j,k(ν+λk)n. |
Consider the two extremes: first the one Jordan block case J=Js(λ). Then p=1, m1=s, P is scalar, say 1,
N=λIs+Us, Tn(t)=Vs(t)SnVs(−t),Tn(t)j,k=s∑a=jk∑b=1[ta−j/(a−j)!] Sn,a,b (−t)k−b/(k−b)!,D=q∑n=0SnJs(ν+λ)n of (4.1). |
Second the diagonal Jordan form with J=diag (λ1,…,λs). Then p=s, mj≡1, Jc=λc,
Tn(t)j,k=s∑c,d=1e(λc−λd)tPj,cQcd,nPd,k |
with all components scalar, and
Dj,k=q∑n=0s∑b,c=1Sn,j,bPb,c(ν+λc)n Pc,k. |
For example, if P=Is, then
Tn(t)j,k=e(λj−λk)tSn,j,k, Dj,k=q∑n=0Sn,j,k(ν+λk)n. |
Example 4.1. Take J=diag (λ1,λ2). So, s=2. Set δ=λ1−λ2. Then
Tn(t)j,k=2∑c,d=1e(λc−λd)tPj,cQcd,nPd,k=2∑c=1Pj,cQcc,nPc,k+eδtPj,1Q12,nP2,k+e−δtPj,2Q21,nP1,k |
and
Tn(t)=PHnP−1, |
where
Hn=(qn,1,1eδtqn,1,2e−δtqn,2,1qn,2,2) |
and qn=P−1SnP. For N of Section 3, we can take
J=diag (i,−i), P=(11−ii)/√2, detP=i, P−1=(1i1−i)/√2. |
In Section 3, we avoided having to use P and these eigenvalues by using N2=−I2.
We now give the solution of Lf=g when Tq=Is and r=s. This can be written X⋅1=AX+F for X, A of (2.15) and F=(0′s,0′s,…,g′)′ as noted on page 90 of Hale [4] for the case s=1.
For j=1,…,qr, set fj=f(t,νj) of (2.1) if νj are distinct. Otherwise choose fj using Theorem 2.2. Set
y(t,ν)=((Ya)′,…,(Y(νIr+N)q−1a)′)′, Uj=U(t,νj), |
where
U(t,ν)=((eνtYa)′,…,(eνtYa)′⋅q−1)′=eνty(t,ν). |
So, Uj∈Cqs. For example, if s=2 and {νj} are distinct, we can take a=a(1) or a(2) of (2.7). Now suppose that r=s. Then U(t)=U=(U1,…,Uqs)∈Cqs×qs is a fundamental matrix solution of X⋅1=AX. That is, det U(t)≠0. Its (j,k) element is
Uj,k=eνjtyk(t,νj). | (5.1) |
So, ˜U(t)=U(t)U(0)−1 is the principal matrix solution at 0 as it satisfies U(0)=Iqs: see, for example, page 80 of Hale [4]. So, by the variation of constants formula, see, for example, his page 81, X⋅1=AX+F has solution
X(t)=U(t)[˜X(0)+∫t0U−1 F], | (5.2) |
where ˜X(0)=U(0)−1X(0). Write U−1 as a 2×2 block matrix with (j,k) element Uj,k∈Cs×s. Then U−1F=(U1,2U2,2)f, so that Lf=g has solution
f(t)=2∑j=1U1,j(t) [˜Xj(0)+∫t0Uj,2g], | (5.3) |
the first of the two block rows of
U(t) [˜X(0)+∫t0˜U2g], |
where (˜U1,˜U2)=U−1. So, the solutions (5.2) and (5.3) have qs unknowns X(0) for X of (2.15), that is, the initial values of f and its first q−1 derivatives.
We now illustrate a use of Floquet's theorem. Set r=qs. Suppose that r=s,Tq=Is, and {Tn} have period T. (This holds for Example 1.1 and Section 3 with T=π or π/τ for Example 3.4.) (2.15) puts Lf=0s in the standard form X⋅1=AX, where A is periodic. According to Floquet's theorem (see, for example, page 118 of Hale [4] or page 164 of Chicone [1]), since U(t) is a fundamental matrix solution of X⋅1=AX∈Cr×r for (X,A) of (2.15), and A has period T, there exists a constant B∈Cr×r and P=P(t) with period T such that U(t)=P(t)eBt. However, Floquet's theorem does not give P(t), B while our method does, as we now show. {νj} are the eigenvalues of B. Since these eigenvalues are distinct, B has diagonal Jordan form, say QΛQ−1, where Λ=diag (ν1,…,νr). So,
eBt=QeΛtQ−1, eΛt=diag (eν1t,…,eνrt), U(t)=R(t)eΛtQ−1, |
where R(t)=PQ. The (j,k) element of U(t)Q is
r∑l=1Uj,l(t)Ql,k=Rj,k(t)eνkt. |
So, Rj,k(t) is the coefficient of eνkt in U(t)Q. This gives R(t). Set (Qj,k)=Q−1. By (5.1), the coefficient of eνbt in
Uj,k(t)=r∑b=1Rj,b(t)eνbtQb,k |
is δb,jxj,k(t)=Rj,b(t)Qb,k. So,
xj,k(t)=yk(t,νj)=r∑b=1Rj,b(t)Qb,k, x(t)=R(t)Q−1, P(t)Q=R(t)=x(t)Q. |
P(t)=x(t)=(yk(t,νj)) is of (5.1). Also, U(t)=P(t)eBt, so that QeΛtQ−1=eBt=UP(t)−1=Q(t) say. So, QeΛt=Q(t)Q. Write Q as (q1,…,qr). So, for all t, qj is an eigenvector of Q(t) with eigenvalue eνjt. So, we can take qj as the eigenvector of Q(T) with eigenvalue eνjT. So, now we have Λ, Q and B.
The authors declare no conflicts of interest in this paper.
[1] | C. C. Chicone, Ordinary Differential Equations with Applications, Springer Verlag, New York, 1991. |
[2] |
V. P. Derevenskii, First-order linear ordinary differential equations over special matrices, Differential Equations, 56 (2020), 696–709. https://doi.org/10.1134/S0012266120060038 doi: 10.1134/S0012266120060038
![]() |
[3] |
M. Elishevich, Boundary-value problem for a system of linear inhomogeneous differential equations of the first order with rectangular matrices, Journal of Mathematical Sciences, 228 (2018), 226–244. https://doi.org/10.1007/s10958-017-3617-8 doi: 10.1007/s10958-017-3617-8
![]() |
[4] | J. K. Hale, Ordinary Differential Equations, 2nd edition, Krieger, Florida, 1980. |
[5] | L. Markus, H. Yamabe, Global stability criteria for differential systems, Osaka Math. Journal, 12 (1960), 305–317. |