
Aerial remote sensing images have complex backgrounds and numerous small targets compared to natural images, so detecting targets in aerial images is more difficult. Resource exploration and urban construction planning need to detect targets quickly and accurately in aerial images. High accuracy is undoubtedly the advantage for detection models in target detection. However, high accuracy often means more complex models with larger computational and parametric quantities. Lightweight models are fast to detect, but detection accuracy is much lower than conventional models. It is challenging to balance the accuracy and speed of the model in remote sensing image detection. In this paper, we proposed a new YOLO model. We incorporated the structures of YOLOX-Nano and slim-neck, then used the SPPF module and SIoU function. In addition, we designed a new upsampling paradigm that combined linear interpolation and attention mechanism, which can effectively improve the model's accuracy. Compared with the original YOLOX-Nano, our model had better accuracy and speed balance while maintaining the model's lightweight. The experimental results showed that our model achieved high accuracy and speed on NWPU VHR-10, RSOD, TGRS-HRRSD and DOTA datasets.
Citation: Lei Yang, Guowu Yuan, Hao Wu, Wenhua Qian. An ultra-lightweight detector with high accuracy and speed for aerial images[J]. Mathematical Biosciences and Engineering, 2023, 20(8): 13947-13973. doi: 10.3934/mbe.2023621
[1] | Sung Woo Choi . Explicit characteristic equations for integral operators arising from well-posed boundary value problems of finite beam deflection on elastic foundation. AIMS Mathematics, 2021, 6(10): 10652-10678. doi: 10.3934/math.2021619 |
[2] | Moh. Alakhrass . A note on positive partial transpose blocks. AIMS Mathematics, 2023, 8(10): 23747-23755. doi: 10.3934/math.20231208 |
[3] | Xinfeng Liang, Mengya Zhang . Triangular algebras with nonlinear higher Lie n-derivation by local actions. AIMS Mathematics, 2024, 9(2): 2549-2583. doi: 10.3934/math.2024126 |
[4] | Cui-Xia Li, Long-Quan Yong . Modified BAS iteration method for absolute value equation. AIMS Mathematics, 2022, 7(1): 606-616. doi: 10.3934/math.2022038 |
[5] | Sara Smail, Chafika Belabbaci . A characterization of Wolf and Schechter essential pseudospectra. AIMS Mathematics, 2024, 9(7): 17146-17153. doi: 10.3934/math.2024832 |
[6] | Yuna Zhao . Construction of blocked designs with multi block variables. AIMS Mathematics, 2021, 6(6): 6293-6308. doi: 10.3934/math.2021369 |
[7] | Wen-Ning Sun, Mei Qin . On maximum residual block Kaczmarz method for solving large consistent linear systems. AIMS Mathematics, 2024, 9(12): 33843-33860. doi: 10.3934/math.20241614 |
[8] | Shakir Ali, Amal S. Alali, Atif Ahmad Khan, Indah Emilia Wijayanti, Kok Bin Wong . XOR count and block circulant MDS matrices over finite commutative rings. AIMS Mathematics, 2024, 9(11): 30529-30547. doi: 10.3934/math.20241474 |
[9] | James Daniel, Kayode Ayinde, Adewale F. Lukman, Olayan Albalawi, Jeza Allohibi, Abdulmajeed Atiah Alharbi . Optimised block bootstrap: an efficient variant of circular block bootstrap method with application to South African economic time series data. AIMS Mathematics, 2024, 9(11): 30781-30815. doi: 10.3934/math.20241487 |
[10] | Ziqiang Wang, Qin Liu, Junying Cao . A higher-order numerical scheme for system of two-dimensional nonlinear fractional Volterra integral equations with uniform accuracy. AIMS Mathematics, 2023, 8(6): 13096-13122. doi: 10.3934/math.2023661 |
Aerial remote sensing images have complex backgrounds and numerous small targets compared to natural images, so detecting targets in aerial images is more difficult. Resource exploration and urban construction planning need to detect targets quickly and accurately in aerial images. High accuracy is undoubtedly the advantage for detection models in target detection. However, high accuracy often means more complex models with larger computational and parametric quantities. Lightweight models are fast to detect, but detection accuracy is much lower than conventional models. It is challenging to balance the accuracy and speed of the model in remote sensing image detection. In this paper, we proposed a new YOLO model. We incorporated the structures of YOLOX-Nano and slim-neck, then used the SPPF module and SIoU function. In addition, we designed a new upsampling paradigm that combined linear interpolation and attention mechanism, which can effectively improve the model's accuracy. Compared with the original YOLOX-Nano, our model had better accuracy and speed balance while maintaining the model's lightweight. The experimental results showed that our model achieved high accuracy and speed on NWPU VHR-10, RSOD, TGRS-HRRSD and DOTA datasets.
A problem that occurs frequently in a variety of mathematical contexts, is to find the common invariant subspaces of a single matrix or set of matrices. In the case of a single endomorphism or matrix, it is relatively easy to find all the invariant subspaces by using the Jordan normal form. Also, some theoretical results are given only for the invariant subspaces of two matrices. However, when there are more than two matrices, the problem becomes much harder, and unexpected invariant subspaces may occur. No systematic method is known. In a recent article [1], we have provided a new algorithms to determine common invariant subspaces of a single matrix or of a set of matrices systematically.
In the present article we consider a more general version of this problem, that is, providing two algorithms for simultaneous block triangularization and block diagonalization of sets of matrices. One of the main steps in the first two proposed algorithms, consists of finding the common invariant subspaces of matrices using the new method proposed in the recent article [1]. It is worth mentioning that an efficient algorithm to explicitly compute a transfer matrix which realizes the simultaneous block diagonalization of unitary matrices whose decomposition in irreducible blocks (common invariant subspaces) is known from elsewhere is given in [2]. An application of simultaneous block-diagonalization of normal matrices in quantum theory is presented in [3].
In this article we shall be concerned with finite dimensions only. Of course the fact that a single complex matrix can always be put into triangular form follows readily from the Jordan normal form theorem [4]. For a set of matrices, Jacobson in [5] introduced the notion of a composition series for a collection of matrices. The idea of a composition series for a group is quite familiar. The Jordan-Hölder Theorem [4] states that any two composition series of the same group have the same length and the same composition factors (up to permutation). Jacobson in [5] characterized the simultaneous block triangularization of a set of matrices by the existence of a chain {0}=V0⊂V1⊂...⊂Vt=Cn of invariant subspaces with dimension dim(Vi/Vi−1)=ni. Therefore, in the context of a collection of matrices Ω={Ai}Ni=1, the idea is to locate a common invariant subspace V of minimal dimension d of a set of matrices Ω. Assume V is generated by the (linearly independent) set B1={u1,u2,...,ud}, and let B={u1,u2,...,ud,ud+1,ud+2,...,un} be a basis of Cn containing B1. Upon setting S=(u1,u2,...,ud,ud+1,ud+2,...,un), S−1AiS has the block triangular form
S−1AiS=(Bi1,1Bi1,20Bi2,2), |
for i=1,...,n. Thereafter, one may define a quotient of the ambient vector space, and each of the matrices in the given collection will pass to this quotient. As such, one defines
Ti=Bi2,2=(0(n−d)×dIn−d)S−1AiS(0d×(n−d)In−d). |
Then one may begin again the process of looking for a common invariant subspace of minimal dimension of a set of matrices {Ti}Ni=1 and iterate the procedure. Since all spaces and matrices are of finite dimension, the procedure must terminate at some point. Again, any two such composition series will be isomorphic. When the various quotients and submatrices are lifted back to the original vector space, one obtains precisely the block-triangular form for the original set of matrices. It is important to find a composition series in the construction in order to make the set of matrices as "block-triangular as possible."
Dubi [6] gave an algorithmic approach to simultaneous triangularization of a set of matrices based on the idea of Jacobson in [5]. In the case of simultaneous triangularization, it can be understood as the existence of a chain {0}=V0⊂V1⊂...⊂Vt=Cn of invariant subspaces with dimension dim(Vi)=i. We generalize his study to cover simultaneous block triangularization of a set of matrices. The generalized algorithm depends on the novel algorithm for constructing invariant subspaces of a set of matrices given in the recent article [1].
Specht [7] (see also [8]) proved that if the associative algebra L generated by a set of matrices Ω over C satisfies L=L∗, then Ω admits simultaneous block triangularization if and only if it admits simultaneous block diagonalization, in both cases via a unitary matrix. Following a result of Specht, we prove that a set of matrices Ω admits simultaneous block diagonalization if and only if the set Γ=Ω∪Ω∗ admits simultaneous block triangularization. Finally, an algorithmic approach to simultaneous block diagonalization of a set of matrices based on this fact is proposed.
The latter part of this paper presents an alternate approach for simultaneous block diagonalization of a set of n×n matrices {As}Ns=1 by an invertible matrix that does not require finding the common invariant subspaces. Maehara et al. [9] introduced an algorithm for simultaneous block diagonalization of a set of matrices by a unitary matrix based on the existence of a Hermitian commuting matrix. Here, we extend their algorithm to simultaneous block diagonalization of a set of matrices by an invertible matrix based on the existence of a commuting matrix which is not necessarily Hermitian. For example, consider the set of matrices Ω={Ai}2i=1 where
A1=(100220111),A2=(000210010). | (1.1) |
The only Hermitian matrix commuting with the set Ω is the identity matrix. Therefore, we cannot apply the proposed algorithm given in [9]. However, one can verify that the following non Hermitian matrix C commutes with all the matrices {Ai}2i=1
C=(000210010). | (1.2) |
The matrix C has distinct eigenvalues λ1=0,λ2=1 with algebraic multiplicities n1=2,n2=1, respectively. Moreover, the matrix C is not diagonalizable. Therefore, we cannot construct the eigenvalue decomposition for the matrix C. However, one can decompose the matrix C by its generalized eigen vectors as follows:
S−1CS=(010000001)=(0100)⊕(1), | (1.3) |
where
S=(0−120011101). | (1.4) |
Initially, it is noted that the matrices {Ai}2i=1 can be decomposed into two diagonal blocks by the constructed invertible matrix S where
S−1A1S=(11201)⊕(2),S−1A2S=(0100)⊕(1). | (1.5) |
Then, a new algorithm is developed for simultaneous block diagonalization by an invertible matrix based on the generalized eigenvectors of a commuting matrix. Moreover, a new characterization is presented by proving that the existence of a commuting matrix that possesses at least two distinct eigenvalues is the necessary and sufficient condition to guarantee the simultaneous block diagonalization by an invertible matrix.
An outline of the paper is as follows. In Section 2 we review several definitions pertaining to block-triangular and block-diagonal matrices and state several elementary consequences that follow from them. In Section 3, following a result of Specht [7] (see also [8]), we provide conditions for putting a set of matrices into block-diagonal form simultaneously. Furthermore, we apply the theoretical results to provide two algorithms that enable a collection of matrices to be put into block-triangular form or block-diagonal form simultaneously by a unitary matrix based on the existence of invariant subspaces. In Section 4, a new characterization is presented by proving that the existence of a commuting matrix that possesses at least two distinct eigenvalues is the necessary and sufficient condition to guarantee the simultaneous block diagonalization by an invertible matrix. Furthermore, we apply the theoretical results to provide an algorithm that enables a collection of matrices to be put into block-diagonal form simultaneously by an invertible matrix based on the existence of a commuting matrix. Sections 3 and 4 also provide concrete examples using the symbolic manipulation system Maple.
Let Ω be a set of n×n matrices over an algebraically closed field F, and let L denote the algebra generated by Ω over F. Similarly, let Ω∗ be the set of the conjugate transpose of each matrix in Ω and L∗ denote the algebra generated by Ω∗ over F.
Definition 2.1. An n×n matrix A is given the notation BT(n1,...,nt) provided A is block upper triangular with t square blocks on the diagonal, of sizes n1,...,nt, where t≥2 and n1+...+nt=n. That is, a block upper triangular matrix A has the form
A=(A1,1A1,2⋯A1,t0A2,2⋯A2,t⋮⋮⋱⋮00⋯At,t) | (2.1) |
where Ai,j is a square matrix for all i=1,...,t and j=i,...,t.
Definition 2.2. A set of n×n matrices Ω is BT(n1,...,nt) if all of the matrices in Ω are BT(n1,...,nt).
Remark 2.3. A set of n×n matrices Ω admits a simultaneous triangularization if it is BT(n1,...,nt) with ni=1 for i=1,...,t.
Remark 2.4. A set of n×n matrices Ω is BT(n1,...,nt) if and only if the algebra L generated by Ω is BT(n1,...,nt).
Proposition 2.5. [7] (see also [8]) Let Ω be a nonempty set of complex n×n matrices. Then, there is a nonsingular matrix S such that SΩS−1 is BT(n1,...,nt) if and only if there is a unitary matrix U such that UΩU∗ is BT(n1,...,nt).
Theorem 2.6. [5,Chapter Ⅳ] Let Ω be a nonempty set of complex n×n matrices. Then, there is a unitary matrix U such that UΩU∗ is BT(n1,...,nt) if and only if the set Ω has a chain {0}=V0⊂V1⊂...⊂Vt=Cn of invariant subspaces with dimension dim(Vi/Vi−1)=ni.
Definition 2.7. An n×n matrix A is given the notation BD(n1,...,nt) provided A is block diagonal with t square blocks on the diagonal, of sizes n1,...,nt, where t≥2, n1+...+nt=n, and the blocks off the diagonal are the zero matrices. That is, a block diagonal matrix A has the form
A=(A10⋯00A2⋯0⋮⋮⋱⋮00⋯At) | (2.2) |
where Ak is a square matrix for all k=1,...,t. In other words, matrix A is the direct sum of A1,...,At. It can also be indicated as A1⊕A2⊕...⊕At.
Definition 2.8. A set of n×n matrices Ω is BD(n1,...,nt) if all of the matrices in Ω are BD(n1,...,nt).
Remark 2.9. A set of n×n matrices Ω admits a simultaneous diagonalization if it is BD(n1,...,nt) with ni=1 for i=1,...,t.
Remark 2.10. A set of n×n matrices Ω is BD(n1,...,nt) if and only if the algebra L generated by Ω is BD(n1,...,nt).
Proposition 2.11. [7] (see also [8]) Let Ω be a nonempty set of complex n×n matrices and let L be the algebra generated by Ω over C. Suppose L=L∗. Then, there is a nonsingular matrix S such that SLS−1 is BT(n1,...,nt) if and only if there is a unitary matrix U such that ULU∗ is BD(n1,...,nt).
Dubi [6] gave an algorithmic approach to simultaneous triangularization of a set of n×n matrices. In this section, we will generalize his study to cover simultaneous block triangularization and simultaneous block diagonalization of a set of n×n matrices. The generalized algorithms depend on the novel algorithm for constructing invariant subspaces of a set of matrices given in the recent article [1] and Theorem 3.3.
Lemma 3.1. Let Ω be a nonempty set of complex n×n matrices, Ω∗ be the set of the conjugate transpose of each matrix in Ω and L be the algebra generated by Γ=Ω∪Ω∗. Then, L=L∗.
Proof. Let A be a matrix in L. Then, A=P(B1,...,Bm) for some multivariate noncommutative polynomial P(x1,...,xm) and matrices {Bi}mi=1∈Γ. Therefore, A∗=P∗(B1,...,Bm)=Q(B∗1,...,B∗m) for some multivariate noncommutative polynomial Q(x1,...,xm) where the matrices {B∗i}mi=1∈Γ∗=Γ. Hence, the matrix A∗∈L
Lemma 3.2. Let Ω be a nonempty set of complex n×n matrices and Ω∗ be the set of the conjugate transpose of each matrix in Ω, and Γ=Ω∪Ω∗. Then, there is a unitary matrix U such that UΓU∗ is BD(n1,...,nt) if and only if there is a unitary matrix U such that UΩU∗ is BD(n1,...,nt).
Proof. Assume that there exists a unitary matrix U such that UΩU∗ is BD(n1,...,nt). Then, (UΩU∗)∗=UΩ∗U∗ is BD(n1,...,nt). Hence, UΓU∗ is BD(n1,...,nt).
Theorem 3.3. Let Ω be a nonempty set of complex n×n matrices and Ω∗ be the set of the conjugate transpose of each matrix in Ω, and Γ=Ω∪Ω∗. Then, there is a unitary matrix U such that UΩU∗ is BD(n1,...,nt) if and only if there is a unitary matrix U such that UΓU∗ is BT(n1,...,nt).
Proof. Let L be the algebra generated by Γ. Then, L=L∗ using Lemma 3.1. Now, by applying Proposition 2.11 and Lemma 3.2, the following statements are equivalent :
There is a unitary matrix U such that UΓU∗ is BT(n1,...,nt).
⟺ There is a unitary matrix U such that ULU∗ is BT(n1,...,nt).
⟺ There is a unitary matrix U such that ULU∗ is BD(n1,...,nt).
⟺ There is a unitary matrix U such that UΓU∗ is BD(n1,...,nt).
⟺ There is a unitary matrix U such that UΩU∗ is BD(n1,...,nt).
(1) Input: the set Ω={Ai}Ni=1.
(2) Set k=0,B=ϕ,s=n,Ti=Ai,S2=I.
(3) Search for a d-dimensional invariant subspace V=⟨v1,v2,...,vd⟩ of a set of matrices {Ti}Ni=1 starting from d=1 up to d=s−1. If one does not exist and k=0, abort and print "no simultaneous block triangularization". Else, if one does not exist and k≠0, go to step (8). Else, go to next step.
(4) Set Vk+1=(S2v1S2v2...S2vd),B=B∪{S2v1,S2v2,...,S2vd},S1=(V1V2...Vk+1).
(5) Find a basis {u1,u2,...,ul} for the orthogonal complement of B.
(6) Set S2=(u1u2...ul),S=(S1S2), and
Ti=(0(s−d)×dIs−d)S−1AiS(0d×(s−d)Is−d).
(7) Set k=k+1,s=s−d, and return to step (3).
(8) Compute the QR decomposition of the invertible matrix S, by means of the Gram–Schmidt process, to convert it to a unitary matrix Q.
(9) Output: a unitary matrix U as the conjugate transpose of the resulting matrix Q.
Remark 3.4. If one uses any non-orthogonal complement in step 5 of Algorithm A, then the matrix S is invertible such that S−1ΩS is BT(n1,...,nt). However, in such a case, one cannot guarantee that UΩU∗ is BT(n1,...,nt).
Example 3.5. The set of matrices Ω={Ai}2i=1 admits simultaneous block triangularization where
A1=(321011050000014012131113020025010006),A2=(44124−484036000−1012320444168524404−102880400040). | (3.1) |
Applying Algorithm A to the set Ω can be summarized as follows:
● Input: Ω.
● Initiation step:
We have k=0,B=ϕ,s=6,T1=A1,T2=A2,S2=I.
● In the first iteration:
We found two-dimensional invariant subspace V=⟨e1,e4⟩ of a set of matrices {Ti}2i=1. Therefore, B={e1,e4},S1=(e1,e4),S2=(e2,e3,e5,e6),
T1=(5000141220251006),T2=(3600−11232444−128840040), | (3.2) |
k=1, and s=4.
● In the second iteration: We found two-dimensional invariant subspace V=⟨e2,e3⟩ of a set of matrices {Ti}2i=1. Therefore, B={e1,e4,e3,e5},S1=(e1,e4,e3,e5),S2=(e2,e6),
T1=(5016),T2=(36−1440), | (3.3) |
k=2, and s=2.
● In the third iteration: There is no one-dimensional invariant subspace of a set of matrices {Ti}2i=1. Therefore, S=(e1e4e3e5e2e6), and the corresponding unitary matrix is
U=(100000000100001000000010010000000001) |
such that the set UΩU∗={UAiU∗}2i=1 is BT(2,2,2) where
UA1U∗=(301121111133004112000225000050000016),UA2U∗=(44−448124452841640032412400−12848000036−10000440). | (3.4) |
(1) Input: the set Ω={Ai}Ni=1.
(2) Construct the set Γ=Ω∪Ω∗.
(3) Find a unitary matrix U such that UΓU∗ is BT(n1,...,nt) using Algorithm A.
(4) Output: a unitary matrix U.
Remark 3.6. Algorithm B provides the finest block-diagonalization. Moreover, the number of the blocks equals the number the of the invariant subspaces, and the size of each block is ni×ni, where ni is the dimension of the invariant subspace.
Example 3.7. The set of matrices Ω={Ai}2i=1 admits simultaneous block diagonalization where
A1=(3000000020000000200000001000000010000000100000003),A2=(0000000000000001000000000000000000000010001000000). | (3.5) |
Applying Algorithm B to the set Ω can be summarized as follows:
● Input: Γ=Ω∪Ω∗.
● Initiation step:
We have k=0,B=ϕ,s=7,T1=A1,T2=A2,T3=AT2,S2=I.
● In the first iteration:
We found one-dimensional invariant subspace V=⟨e5⟩ of a set of matrices {Ti}3i=1. Therefore, B={e5},S1=(e5),S2=(e1,e2,e3,e4,e6,e7),
T1=(300000020000002000000100000010000003),T2=(000000000000010000000000000100100000),T3=TT2, | (3.6) |
k=1, and s=6.
● In the second iteration: We found two-dimensional invariant subspace V=⟨e4,e5⟩ of a set of matrices {Ti}3i=1. Therefore, B={e5,e4,e6},S1=(e5e4e6),S2=(e1,e2,e3,e7),
T1=(3000020000200003),T2=(0000000001001000),T3=TT2, | (3.7) |
k=2, and s=4.
● In the third iteration: We found two-dimensional invariant subspace V=⟨e2,e3⟩ of a set of matrices {Ti}3i=1. Therefore, B={e5,e4,e6,e2,e3},S1=(e5e4e6e2e3),S2=(e1,e7),
T1=(3003),T2=(0010),T3=(0100), | (3.8) |
k=3, and s=2.
● In the fourth iteration: There is no one-dimensional invariant subspace of a set of matrices {Ti}3i=1. Therefore, S=(e5e4e6e2e3e1e7), and the corresponding unitary matrix is
U=(0000100000100000000100100000001000010000000000001) |
such that the set UΩU∗={UAiU∗}2i=1 is BD(1,2,2,2) where
UA1U∗=(1)⊕(1001)⊕(2002)⊕(3003),UA2U∗=(0)⊕(0010)⊕(0010)⊕(0010). | (3.9) |
Example 3.8. The set of matrices Ω={Ai}2i=1 admits simultaneous block diagonalization where
A1=(3000000020000000200000001000000010000000100000003),A2=(0000000000100001000000000000000010000001001000000). | (3.10) |
Similarly, applying Algorithm B to the set Ω provides the matrix S=(e6e5e7e1e3e2e4). Therefore, the corresponding unitary matrix is
U=(0000010000010000000011000000001000001000000001000) |
such that the set UΩU∗={UAiU∗}2i=1 is BD(2,2,3) where
UA1U∗=(1001)⊕(3003)⊕(200020001),UA2U∗=(0101)⊕(0100)⊕(010001000). | (3.11) |
Example 3.9. The set of matrices Ω={Ai}3i=1 admits simultaneous block diagonalization where
A1=(000000000020000000001000000000−20000000000000000000−1000000000−100000000010000000000),A2=(000100000−100010000000001000000000000000−100000000000000000000000000000−100000000000),A3=(0−100000000000000000000000001000−100000100000000010000000000000−10000000000000000000). | (3.12) |
Similarly, applying Algorithm B to the set Ω provides the matrix S=(e1+e5e9e3e6e8−e7e1−e5,e2e4). Therefore, the corresponding unitary matrix is
U=(12√200012√20000000000001001000000000001000000000010000000−10012√2000−12√20000010000000000100000) |
such that the set UΩU∗={UAiU∗}3i=1 is BD(1,1,2,2,3) where
UA1U∗=(0)⊕(0)⊕(100−1)⊕(100−1)⊕(00002000−2),UA2U∗=(0)⊕(0)⊕(0100)⊕(0100)⊕(00√2−√200000),UA3U∗=(0)⊕(0)⊕(0010)⊕(0010)⊕(0−√20000√200). | (3.13) |
This section focuses on an alternate approach for simultaneous block diagonalization of a set of n×n matrices {As}Ns=1 by an invertible matrix that does not require finding the common invariant subspaces as Algorithm B given in the previous section. Maehara et al. [9] introduced an algorithm for simultaneous block diagonalization of a set of matrices by a unitary matrix based on the eigenvalue decomposition of a Hermitian commuting matrix. Here, we extend their algorithm to be applicable for a non-Hermitian commuting matrix by considering its generalized eigen vectors. Moreover, a new characterization is presented by proving that the existence of a commuting matrix that possesses at least two distinct eigenvalues is the necessary and sufficient condition to guarantee the simultaneous block diagonalization by an invertible matrix.
Proposition 4.1. Let V be a vector space, and let T:V→V be a linear operator. Let λ1,...,λk be distinct eigenvalues of T. Then, each generalized eigenspace Gλi(T) is T-invariant, and we have the direct sum decomposition
V=Gλ1(T)⊕Gλ2(T)⊕...⊕Gλk(T). |
Lemma 4.2. Let V be a vector space, and let T:V→V, L:V→V be linear commuting operators. Let λ1,...,λk be distinct eigenvalues of T. Then, each generalized eigenspace Gλi(T) is L-invariant.
Proof. Let V be a vector space and λ1,...,λk be distinct eigenvalues of T with the minimal polynomial μ(x)=(x−λ1)n1(x−λ2)n2...(x−λk)nk. Then, we have the direct sum decomposition V=Gλ1(T)⊕Gλ2(T)⊕...⊕Gλk(T).
For each i=1,..,k, let x∈Gλi(T), and then (T−λiI)nix=0. Then, (T−λiI)niLx=L(T−λiI)nix=0. Hence, Lx∈Gλi(T).
Theorem 4.3. Let {As}Ns=1 be a set of n×n matrices. Then, the set {As}Ns=1 admits simultaneous block diagonalization by an invertible matrix S if and only if the set {As}Ns=1 commutes with a matrix C that possesses two distinct eigenvalues.
Proof. ⇒ Assume that the set {As}Ns=1 admits simultaneous block diagonalization by the an invertible matrix S such that
S−1AsS=Bs,1⊕Bs,2⊕...⊕Bs,k, |
where the number of blocks k≥2, and the matrices Bs,1,Bs,2,...,Bs,k have sizes n1×n1,n2×n2,...,nk×nk, respectively, for all s=1,..,N.
Now, define the matrix C as
C=S(λ1In1×n1⊕λ2In2×n2⊕...⊕λkInk×nk)S−1, |
where λ1,λ2,...,λk are any distinct numbers.
Clearly, the matrix C commutes with the set {As}Ns=1. Moreover, it has the distinct eigenvalues λ1,λ2,...,λk.
⇐ Assume that the set {As}Ns=1 commutes with a matrix C that posseses distinct eigenvalues λ1,λ2,...,λk.
Using Proposition 4.1, one can use the generalized eigenspace Gλi(C) of the matrix C associated to these distinct eigenvalues to decompose the matrix C as a direct sum of k matrices. This can be achieved by restricting the matrix C on the invariant subspaces Gλi(C) as follows:
S−1CS=[C]Gλ1(C)⊕[C]Gλ2(C)⊕...⊕[C]Gλk(C) |
where
S=(Gλ1(C),Gλ2(C),...,Gλk(C)). |
Using Lemma 4.2, one can restrict each matrix As on the invariant subspaces Gλi(C) to decompose the matrix As as a direct sum of k matrices as follows:
S−1AsS=[As]Gλ1(C)⊕[As]Gλ2(C)⊕...⊕[As]Gλk(C). |
Remark 4.4. For a given set of n×n matrices {As}Ns=1, if the set {As}Ns=1 commutes only with the matrices having only one eigenvalue, then it does not admit a simultaneous block diagonalization by an invertible matrix.
Algorithm C:
(1) Input: the set Ω={As}Ns=1.
(2) Construct the the following matrix:
X=(I⊗A1−AT1⊗II⊗A2−AT2⊗I...I⊗AN−ATN⊗I). |
(3) Compute the null space of the matrix X and reshape the obtained vectors as n×n matrices. These matrices commute with all the matrices {As}Ns=1.
(4) Choose a matrix C from the obtained matrices that possesses two distinct eigenvalues.
(5) Find the distinct eigenvalues λ1,...,λk of the matrix C and the corresponding algebraic multiplicity n1,n2,...,nk.
(6) Find each generalized eigenspace Gλi(C) of the matrix C associated to the eigenvalue λi by computing the null space of (C−λiI)ni.
(7) Construct the invertible matrix S as
S=(Gλ1(C),Gλ2(C),...,Gλk(C)). |
(8) Verify that
S−1AsS=Bs,1⊕Bs,2⊕...⊕Bs,k, |
where the matrices Bs,1,Bs,2,...,Bs,k have sizes n1×n1,n2×n2,...,nk×nk, respectively, for all s=1,..,N.
(9) Output: an invertible matrix S.
Remark 4.5. Algorithm C provides the finest block-diagonalization if one chooses a matrix C with maximum number of distinct eigenvalues. Moreover, the number of the blocks equals the number the of the distinct eigenvalues, and the size of each block is ni×ni, where ni is the algebraic multiplicity of the eigenvalue λi.
Example 4.6. Consider the set of matrices Ω={Ai}6i=1 where
A1=(0000000001000000100−1000000−1000000000),A2=(000−10000000000000110000000000000−1000),A3=(0000−1000000−1000000000000100000010000),A4=(010000−1000000000000000000000010000−10),A5=(001000000000−10000000000−1000000000100),A6=(0000000010000−10000000010000−100000000). | (4.1) |
The set Ω admits simultaneous block diagonalization by an invertible matrix. An invertible matrix can be obtained by applying algorithm C to the set Ω as summarized below:
● A matrix C that commutes with all the matrices {Ai}6i=1 can be obtained as
C=(0000010000−100001000010000−10000100000). | (4.2) |
.
● The distinct eigenvalues of the matrix C are λ1=−1,λ2=1 with algebraic multiplicities n1=3,n2=3, respectively..
● The generalized eigenspaces of the matrix C associated to the distinct eigenvalues are
Gλ1(C)=N(C−λ1I)3=⟨e6−e1,e2+e5,e4−e3⟩,Gλ2(C)=N(C−λ2I)3=⟨e1+e6,e5−e2,e3+e4⟩. | (4.3) |
● The invertible matrix S=(Gλ1(C),Gλ2(C)) is
S=(−1001000100−1000−1001001001010010100100). | (4.4) |
● The set S−1ΩS={S−1AiS}6i=1 contains block diagonal matrices where
S−1A1S=(0000010−10)⊕(00000−1010),S−1A2S=(001000−100)⊕(00−1000100),S−1A3S=(010−100000)⊕(0−10100000),S−1A4S=(0−10100000)⊕(0−10100000),S−1A5S=(001000−100)⊕(001000−100),S−1A6S=(00000−1010)⊕(00000−1010). | (4.5) |
It is well known that a set of non-defective matrices can be simultaneously diagonalized if and only if the matrices commute. In the case of non-commuting matrices, the best that can be achieved is simultaneous block diagonalization. Both Algorithm B and the Maehara et al. [9] algorithm are applicable for simultaneous block diagonalization of a set of matrices by a unitary matrix. Algorithm C can be applied for block diagonalization by an invertible matrix when finding a unitary matrix is not possible. In case block diagonalization of a set of matrices is not possible by a unitary or an invertible matrix, then one may utilize block triangularization by Algorithm A. Algorithms A and B are based on the existence of invariant subspaces; however, Algorithm C is based on the existence of a commuting matrix which is not necessarily Hermitian, unlike the Maehara et al. algorithm.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
Ahmad Y. Al-Dweik and M. T. Mustafa would like to thank Qatar University for its support and excellent research facilities. R. Ghanam and G. Thompson are grateful to VCU Qatar and Qatar Foundation for their support.
The authors declare that they have no conflicts of interest.
[1] |
M. Lu, Y. Xu, H. Li, Vehicle Re-Identification based on UAV viewpoint: dataset and method, Remote Sens., 14 (2022), 4630. https://doi.org/10.3390/rs14184603 doi: 10.3390/rs14184603
![]() |
[2] |
S. Ijlil, A. Essahlaoui, M. Mohajane, N. Essahlaoui, E. M. Mili, A. V. Rompaey, Machine learning algorithms for modeling and mapping of groundwater pollution risk: A study to reach water security and sustainable development (Sdg) goals in a editerranean aquifer system, Remote Sens., 14 (2022), 2379. https://doi.org/10.3390/rs14102379 doi: 10.3390/rs14102379
![]() |
[3] |
Z. Jiang, Z. Song, Y. Bai, X. He, S. Yu, S. Zhang, et al., Remote sensing of global sea surface pH based on massive underway data and machine mearning, Remote Sens., 14 (2022), 2366. https://doi.org/10.3390/rs14102366 doi: 10.3390/rs14102366
![]() |
[4] |
Y. Zhao, L. Ge, H. Xie, G. Bai, Z. Zhang, Q. Wei, et al., ASTF: Visual abstractions of time-varying patterns in radio signals, IEEE Trans. Visual Comput. Graphics, 29 (2023), 214–224. https://doi.org/10.1109/TVCG.2022.3209469 doi: 10.1109/TVCG.2022.3209469
![]() |
[5] | R. Girshick, J. Donahue, T. Darrell J. Malik, Rich feature hierarchies for accurate object detection and semantic segmentation, in 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2014), 580–587. https://doi.org/10.1109/CVPR.2014.81 |
[6] | R. Girshick, Fast R-CNN, in 2015 IEEE International Conference on Computer Vision (ICCV), (2015), 1440–1448. https://doi.org/10.1109/ICCV.2015.169 |
[7] |
S. Ren, K. He, R. Girshick, J. Sun, Faster R-CNN: Towards real-time object detection with region proposal networks, IEEE Trans. Pattern Anal. Mach. Intell., 39 (2017), 1137–1149. https://doi.org/10.1109/TPAMI.2016.2577031 doi: 10.1109/TPAMI.2016.2577031
![]() |
[8] | J. Redmon, S. Divvala, R. Girshick, A. Farhadi, You only look once: Unified, real-time object detection, in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2016), 779–788. https://doi.org/10.1109/CVPR.2016.91 |
[9] | J. Redmon, A. Farhadi, YOLO9000: Better, Faster, Stronger, in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2017), 6517–6525. https://doi.org/10.1109/CVPR.2017.690 |
[10] | J. Redmon, A. Farhadi, YOLOv3: an incremental improvement, arXiv preprint, (2018), arXiv: 1804.02767. http://arXiv.org/abs/1804.02767 |
[11] | A. Bochkovskiy, C. Y. Wang, H. Liao, YOLOv4: optimal speed and accuracy of object detection, arXiv preprint, (2020), arXiv: 2004.10934. http://arXiv.org/abs/2004.10934 |
[12] | G. Jocher, Yolov5, 2020. Available from: https://github.com/ultralytics/yolov5. |
[13] | Z. Ge, S. Liu, F. Wang, Z. Li, J. Sun, YOLOX: Exceeding YOLO series in 2021, arXiv preprint, (2021), arXiv: 2107.08430. https://arXiv.org/abs/2107.08430 |
[14] |
Y. Li, X. Liu, H. Zhang, X. Li, X. Sun, Optical remote sensing image retrieval based on convolutional neural networks (in Chinese), Opt. Precis. Eng., 26 (2018), 200–207. https://doi.org/10.3788/ope.20182601.0200 doi: 10.3788/ope.20182601.0200
![]() |
[15] | A. Van Etten, You only look twice: Rapid multi-scale object detection in satellite imagery, arXiv preprint, (2018), arXiv: 1805.09512. https://doi.org/10.48550/arXiv.1805.09512 |
[16] |
M. Ahmed, Y. Wang, A. Maher, X. Bai, Fused RetinaNet for small target detection in aerial images, Int. J. Remote Sens., 43 (2022), 2813–2836. https://doi.org/10.1080/01431161.2022.2071115 doi: 10.1080/01431161.2022.2071115
![]() |
[17] |
H. Liu, G. Yuan, L. Yang, K. Liu, H. Zhou, An appearance defect detection method for cigarettes based on C‐CenterNet, Electronics, 11 (2022), 2182. https://doi.org/10.3390/electronics11142182 doi: 10.3390/electronics11142182
![]() |
[18] |
S. Du, B. Zhang, P. Zhang, P. Xiang, H. Xue, FA-YOLO: An improved YOLO model for infrared occlusion object detection under confusing background, Wireless Commun. Mobile Comput., 2021 (2021). https://doi.org/10.1155/2021/1896029 doi: 10.1155/2021/1896029
![]() |
[19] | A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, et al., MobileNets: Efficient convolutional neural networks for mobile vision applications, arXiv preprint, (2017), arXiv: 1704.04861. https://doi.org/10.48550/arXiv.1704.04861 |
[20] | M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, L. C. Chen, MobileNetV2: Inverted residuals and linear bottlenecks, in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (2018), 4510–4520. https://doi.org/10.1109/CVPR.2018.00474 |
[21] | A. Howard, M. Sandler, B. Chen, W. Wang, L. C. Chen, M. Tan, et al., Searching for mobileNetV3, in 2019 IEEE/CVF International Conference on Computer Vision (ICCV), (2019), 1314–1324. https://doi.org/10.1109/ICCV.2019.00140 |
[22] | X. Zhang, X. Zhou, M. Lin, J. Sun, ShuffleNet: An extremely efficient convolutional neural network for mobile devices, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2018), 6848–6856. |
[23] | N. Ma, X. Zhang, H. T. Zheng, J. Sun, ShuffleNet V2: Practical guidelines for efficient CNN architecture design, in European Conference on Computer Vision (ECCV), (2018), 122–138. https://doi.org/10.1109/CVPR.2018.00716 |
[24] | RangiLyu, NanoDet-Plus: Super fast and high accuracy lightweight anchor-free object detection model, 2021. Available from: https://github.com/RangiLyu/nanodet. |
[25] | C. Y. Wang, H. Liao, Y. H. Wu, P. Y. Chen, J. W. Hsieh, I. H. Yeh, CSPNet: A bew backbone that can enhance learning capability of CNN, in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), (2020), 1571–1580. https://doi.org/10.1109/CVPRW50498.2020.00203 |
[26] |
X. Luo, Y. Wu, L. Zhao, YOLOD: A target detection method for UAV aerial imagery, Remote Sens., 14 (2022), 3240. https://doi.org/10.3390/rs14143240 doi: 10.3390/rs14143240
![]() |
[27] |
D. Yan, G. Li, X. Li, H. Zhang, H. Lei, K. Lu, et al., An improved faster R-CNN method to detect tailings ponds from high-resolution remote sensing images, Remote Sens. 13 (2021), 2052. https://doi.org/10.3390/rs13112052 doi: 10.3390/rs13112052
![]() |
[28] | F. C. Akyon, S. O. Altinuc, A. Temizel, Slicing aided hyper inference and fine-tuning for small object detection, in 2022 IEEE International Conference on Image Processing (ICIP), (2022), 966–970. https://doi.org/10.1109/ICIP46576.2022.9897990 |
[29] |
L. Yang, G. Yuan, H. Zhou, H. Liu, J. Chen, H. Wu, RS-YOLOX: A high-precision detector for object detection in satellite remote sensing images, Appli. Sci., 12 (2022), 8707. https://doi.org/10.3390/app12178707 doi: 10.3390/app12178707
![]() |
[30] |
J. Liu, C. Liu, Y. Wu, Z. Sun, H. Xu, Insulators' identification and missing defect detection in aerial images based on cascaded YOLO models, Comput. Intell. Neurosci., 2022 (2022). https://doi.org/10.1155/2022/7113765 doi: 10.1155/2022/7113765
![]() |
[31] | X. Li, Y. Qin, F. Wang, F. Guo, J. T. W. Yeow, Pitaya detection in orchards using the MobileNet-YOLO model, in 2020 39th Chinese Control Conference (CCC), (2020), 6274–6278. https://doi.org/10.23919/CCC50068.2020.9189186 |
[32] | Z. Tian, C. Shen, H. Chen, T. He, FCOS: Fully convolutional one-stage object detection, in 2019 IEEE/CVF International Conference on Computer Vision (ICCV), (2019), 9626–9635. https://doi.org/10.1109/ICCV.2019.00972 |
[33] |
H. Law, J. Deng, CornerNet: Detecting objects as paired keypoints, Int. J. Comput. Vision, 128 (2020), 642–656. https://doi.org/10.1007/s11263-019-01204-1 doi: 10.1007/s11263-019-01204-1
![]() |
[34] | G. Song, Y. Liu, X. Wang, Revisiting the sibling head in object detector, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (2020), 11563–11572. |
[35] |
K. He, X. Zhang, S. Ren, J. Sun, Spatial pyramid pooling in deep convolutional networks for visual recognition, IEEE Trans. Pattern Anal. Mach. Intell., 37 (2015), 1904–1916. https://doi.org/10.1109/TPAMI.2015.2389824 doi: 10.1109/TPAMI.2015.2389824
![]() |
[36] |
L. C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, A. L. Yuille, DeepLab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs, IEEE Trans. Pattern Anal. Mach. Intell., 40 (2018), 834–848. https://doi.org/10.1109/TPAMI.2017.2699184 doi: 10.1109/TPAMI.2017.2699184
![]() |
[37] | C. Y. Wang, A. Bochkovskiy, H. Liao, YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (2023), 7464–7475. |
[38] | H. Li, J. Li, H. Wei, Z. Liu, Z. Zhan, Q. Ren, Slim-neck by GSConv: A better design paradigm of detector architectures for autonomous vehicles, arXiv preprint, (2022), arXiv: 2206.02424. https://doi.org/10.48550/arXiv.2206.02424 |
[39] | V. Dumoulin, F. Visin, A guide to convolution arithmetic for deep learning, arXiv preprint, (2018), arXiv: 1603.07285. https://doi.org/10.48550/arXiv.1603.07285 |
[40] | F. Yu, V. Koltun, T. Funkhouser, Dilated residual networks, in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2017), 636–644. https://doi.org/10.1109/CVPR.2017.75 |
[41] | Q. Wang, B. Wu, P. Zhu, P. Li, W. Zuo, Q. Hu, ECA-Net: Efficient channel attention for deep convolutional neural networks, in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (2020), 11531–11539. https://doi.org/10.1109/CVPR42600.2020.01155 |
[42] | B. Jiang, R. Luo, J. Mao, T. Xiao, Y. Jiang, Acquisition of localization confidence for accurate object detection, in Proceedings of the European Conference on Computer Vision (ECCV), (2018), 784–799. |
[43] | J. He, S. Erfani, X. Ma, J. Bailey, Y. Chi, X. S. Hua, Alpha-IoU: A family of power intersection over union losses for bounding box regression, in NeurIPS 2021 Conference, 2021. |
[44] | H. Rezatofighi, N. Tsoi, J. Gwak, A. Sadeghian, I. Reid, S. Savarese, Generalized intersection over union: A metric and a loss for bounding box regression, in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (2019), 658–666. https://doi.org/10.1109/CVPR.2019.00075 |
[45] | Z. Zheng, P. Wang, W. Liu, J. Li, R. Ye, D. Ren, Distance-IoU loss: Faster and better learning for bounding box regression, in Proceedings of the AAAI Conference on Artificial Intelligence, 2020. https://doi.org/10.1609/aaai.v34i07.6999 |
[46] | Z. Gevorgyan, SIoU loss: More powerful learning for bounding box regression, arXiv preprint, (2022), arXiv: 2205.12740. https://doi.org/10.48550/arXiv.2205.12740 |
[47] |
G. Cheng, P. Zhou, J. Han, Learning rotation-invariant convolutional neural networks for object detection in VHR optical remote sensing images, IEEE Trans. Geosci. Remote Sens., 54 (2016), 7405–7415. https://doi.org/10.1109/TGRS.2016.2601622 doi: 10.1109/TGRS.2016.2601622
![]() |
[48] |
Y. Long, Y. Gong, Z. Xiao, Q. Liu, Accurate object localization in remote sensing images based on convolutional neural networks, IEEE Trans. Geosci. Remote Sens., 55 (2017), 2486–2498. https://doi.org/10.1109/TGRS.2016.2645610 doi: 10.1109/TGRS.2016.2645610
![]() |
[49] |
X. Lu, Y. Zhang, Y. Yuan, Y. Feng, Gated and axis-concentrated localization network for remote sensing object detection, IEEE Trans. Geosci. Remote Sens., 58 (2020), 179–192. https://doi.org/10.1109/TGRS.2019.2935177 doi: 10.1109/TGRS.2019.2935177
![]() |
[50] | L. Yang, R. Y. Zhang, L. Li, X. Xie, SimAM: A simple, parameter-free attention module for convolutional neural networks, in Proceedings of the 38th International Conference on Machine Learning, 139 (2021), 11863–11874. |
[51] | Z. Zhong, Z. Q. Lin, R. Bidart, X. Hu, I. B. Daya, Z. Li, et al., Squeeze-and-attention networks for semantic segmentatio, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (2020), 13065–13074. |
[52] | R. Saini, N. K. Jha, B. Das, S. Mittal, C. K. Mohan, ULSAM: Ultra-lightweight subspace attention module for compact convolutional neural networks, in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), (2020), 1616–1625. https://doi.org/10.1109/WACV45572.2020.9093341 |
[53] | Y. Liu, Z. Shao, Y. Teng, N. Hoffmann, NAM: Normalization-based attention module, arXiv preprint, (2021), arXiv: 2111.12419. https://doi.org/10.48550/arXiv.2111.12419 |
[54] | X. Ma, Yolo-Fastest: yolo-fastest-v1.1.0, 2021. Available from: https://github.com/dog-qiuqiu/Yolo-Fastest. |
[55] | X. Ma, FastestDet: Ultra lightweight anchor-free real-time object detection algorithm, 2022. Available from: https://github.com/dog-qiuqiu/FastestDet. |
[56] | X. Yang, J. Yan, Z. Feng, T. He, R3Det: Refined single-stage detector with feature refinement for rotating object, in Proceedings of the AAAI Conference on Artificial Intelligence, 35 (2021), 3163–3173. https://doi.org/10.1609/aaai.v35i4.16426 |
[57] |
J. Han, J. Ding, J. Li, G. S. Xia, Align deep features for oriented object detection, IEEE Trans. Geosci. Remote Sens., 60 (2022), 1–11. https://doi.org/10.1109/TGRS.2021.3062048 doi: 10.1109/TGRS.2021.3062048
![]() |
[58] | X. Xie, G. Cheng, J. Wang, X. Yao, J. Han, Oriented R-CNN for object detection, in 2021 IEEE/CVF International Conference on Computer Vision (ICCV), (2021), 3500–3509. https://doi.org/10.1109/ICCV48922.2021.00350 |
[59] | J. Ding, N. Xue, Y. Long, G. S. Xia, Q. Lu, Learning RoI transformer for oriented object detection in aerial images, in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (2019), 2844–2853. https://doi.org/10.1109/CVPR.2019.00296 |
[60] |
S. Zhong, H. Zhou, Z. Ma, F. Zhang, J. Duan, Multiscale contrast enhancement method for small infrared target detection, Optik, 271 (2022), 170134. https://doi.org/10.1016/j.ijleo.2022.170134 doi: 10.1016/j.ijleo.2022.170134
![]() |
[61] |
S. Zhong, H. Zhou, X. Cui, X. Cao, F. Zhang, J. Duan, Infrared small target detection based on local-image construction and maximum correntropy, Measurement, 211 (2023), 112662. https://doi.org/10.1016/j.measurement.2023.112662 doi: 10.1016/j.measurement.2023.112662
![]() |