The escalating sophistication of malware poses a formidable security challenge, as it evades traditional protective measures. Static analysis, an initial step in malware investigation, involves code scrutiny without actual execution. One static analysis approach employs the conversion of executable files into image representations, harnessing the potency of deep learning models. Convolutional neural networks (CNNs), particularly adept at image classification, have potential for malware detection. However, their inclination towards structured data requires a preprocessing phase to convert software into image-like formats. This paper outlines a methodology for malware detection that involves applying deep learning models to image-converted executable files. Experimental evaluations have been performed by using CNN models, autoencoder-based models, and pre-trained counterparts, all of which have exhibited commendable performance. Consequently, employing deep learning for image-converted executable analysis emerges as a fitting strategy for the static analysis of software. This research is significant because it utilized the largest dataset to date and encompassed a wide range of deep learning models, many of which have not previously been tested together.
Citation: Mesut GUVEN. Leveraging deep learning and image conversion of executable files for effective malware detection: A static malware analysis approach[J]. AIMS Mathematics, 2024, 9(6): 15223-15245. doi: 10.3934/math.2024739
[1] | Maha M. Althobaiti, José Escorcia-Gutierrez . Weighted salp swarm algorithm with deep learning-powered cyber-threat detection for robust network security. AIMS Mathematics, 2024, 9(7): 17676-17695. doi: 10.3934/math.2024859 |
[2] | Ibrahim R. Alzahrani, Randa Allafi . Integrating Ebola optimization search algorithm for enhanced deep learning-based ransomware detection in Internet of Things security. AIMS Mathematics, 2024, 9(3): 6784-6802. doi: 10.3934/math.2024331 |
[3] | Sultanah M. Alshammari, Nofe A. Alganmi, Mohammed H. Ba-Aoum, Sami Saeed Binyamin, Abdullah AL-Malaise AL-Ghamdi, Mahmoud Ragab . Hybrid arithmetic optimization algorithm with deep learning model for secure Unmanned Aerial Vehicle networks. AIMS Mathematics, 2024, 9(3): 7131-7151. doi: 10.3934/math.2024348 |
[4] | Alaa O. Khadidos . Advancements in remote sensing: Harnessing the power of artificial intelligence for scene image classification. AIMS Mathematics, 2024, 9(4): 10235-10254. doi: 10.3934/math.2024500 |
[5] | Khaled Tarmissi, Hanan Abdullah Mengash, Noha Negm, Yahia Said, Ali M. Al-Sharafi . Explainable artificial intelligence with fusion-based transfer learning on adverse weather conditions detection using complex data for autonomous vehicles. AIMS Mathematics, 2024, 9(12): 35678-35701. doi: 10.3934/math.20241693 |
[6] | Ebtesam Al-Mansor, Mohammed Al-Jabbar, Arwa Darwish Alzughaibi, Salem Alkhalaf . Dandelion optimization based feature selection with machine learning for digital transaction fraud detection. AIMS Mathematics, 2024, 9(2): 4241-4258. doi: 10.3934/math.2024209 |
[7] | Mashael M Asiri, Abdelwahed Motwakel, Suhanda Drar . Robust sign language detection for hearing disabled persons by Improved Coyote Optimization Algorithm with deep learning. AIMS Mathematics, 2024, 9(6): 15911-15927. doi: 10.3934/math.2024769 |
[8] | Thavavel Vaiyapuri, Prasanalakshmi Balaji, S. Shridevi, Santhi Muttipoll Dharmarajlu, Nourah Ali AlAseem . An attention-based bidirectional long short-term memory based optimal deep learning technique for bone cancer detection and classifications. AIMS Mathematics, 2024, 9(6): 16704-16720. doi: 10.3934/math.2024810 |
[9] | Mohammed Aljebreen, Hanan Abdullah Mengash, Khalid Mahmood, Asma A. Alhashmi, Ahmed S. Salama . Enhancing cybersecurity in cloud-assisted Internet of Things environments: A unified approach using evolutionary algorithms and ensemble learning. AIMS Mathematics, 2024, 9(6): 15796-15818. doi: 10.3934/math.2024763 |
[10] | Wahida Mansouri, Amal Alshardan, Nazir Ahmad, Nuha Alruwais . Deepfake image detection and classification model using Bayesian deep learning with coronavirus herd immunity optimizer. AIMS Mathematics, 2024, 9(10): 29107-29134. doi: 10.3934/math.20241412 |
The escalating sophistication of malware poses a formidable security challenge, as it evades traditional protective measures. Static analysis, an initial step in malware investigation, involves code scrutiny without actual execution. One static analysis approach employs the conversion of executable files into image representations, harnessing the potency of deep learning models. Convolutional neural networks (CNNs), particularly adept at image classification, have potential for malware detection. However, their inclination towards structured data requires a preprocessing phase to convert software into image-like formats. This paper outlines a methodology for malware detection that involves applying deep learning models to image-converted executable files. Experimental evaluations have been performed by using CNN models, autoencoder-based models, and pre-trained counterparts, all of which have exhibited commendable performance. Consequently, employing deep learning for image-converted executable analysis emerges as a fitting strategy for the static analysis of software. This research is significant because it utilized the largest dataset to date and encompassed a wide range of deep learning models, many of which have not previously been tested together.
Let
G=(V(G),E(G)) |
be a simple undirected graph on n vertices such that V(G) and E(G) denote the vertex set and edge set of G, respectively. Let V(G)={v1,v2,⋯,vn}, for vi∈V(G), and dG(vi)=d(vi)=di denotes the degree of vertex vi and the set of vertices adjacent to vi, denoted by N(vi), refers to the neighborhood of vi. The maximum and the minimum degree of G are denoted by Δ=Δ(G) and δ=δ(G), respectively. λ(M), λk(M) and λmin(M) denote the largest eigenvalue, the k-th largest eigenvalue, and the smallest eigenvalue of a matrix M, respectively. The set of all eigenvalues for a matrix M together with their multiplicities is called the M-spectrum. A real symmetric matrix M is called positive semidefinite if λmin(M)≥0. Denote by Kn, Pn, Cn and Ks,n−s the complete graph, path, cycle, and complete bipartite graph with n vertices, respectively. Let A(G) and D(G) be the adjacency matrix and the diagonal matrix of the degrees of G, respectively. The signless Laplacian Q(G) of G is defined as
Q(G)=D(G)+A(G). |
The Laplacian L(G) of G is defined as
L(G)=D(G)−A(G). |
We denote the eigenvalues of Q(G) and L(G) by ρk(G) and λk(G), respectively, where 1≤k≤n (we drop G when it is clear from the context). In particular, the largest eigenvalue of Q(G) is called the signless Laplacian spectral radius of G, denoted by ρ(G), and the largest eigenvalue of L(G) is called the Laplacian spectral radius of G, denoted by λ(G).
In [1], Nikiforov introduced the Aα-matrix of G, which is the convex linear combination of A(G) and D(G) defined by
Aα(G)=αD(G)+(1−α)A(G),0≤α≤1. |
Clearly,
A0(G)=A(G), A1(G)=D(G)and2A12(G)=Q(G). |
Thus, the matrix Aα(G) is a generalization of the adjacency matrix and the signless Laplacian matrix. The largest eigenvalue of Aα(G) is called the Aα-spectral radius of G, denoted by ρα(G). Nikiforov [1] had studied the matrix Aα(G), and he has investigated many properties on Aα(G), including bounds on the k-th largest (especially, the largest, the second largest, and the smallest) eigenvalue of Aα(G), the positive semidefiniteness of Aα(G), etc. From then on, the study of Aα-spectra and the Aα-spectral radius for the graph has attracted the attention of many researchers. In [2], Nikiforov et al. gave an upper bound of the spectral radius of Aα(TΔ), where TΔ is the tree of maximal degree Δ. Also, he obtained several bounds on the spectral radius of Aα of general graphs. The graphs with
λk(Aα(G))=αn−1,2≤k≤n |
had been characterized by Lin et al. in [3]. In [4], Guo and Zhou gave upper bounds for the Aα-spectral radius for some graphs under some conditions. Guo and Zhang [5]; obtained a sharp upper bound on the Aα-spectral radius for α∈[12,1), and proved that for two connected graphs G1 and G2, ρα(G1)>ρα(G2) under certain conditions. For further related studies, one may see [6,7,8].
The research on L(G) has shown that it is a remarkable matrix in many respects. Since; L(G) is just the difference between D(G) and A(G), then to understand to what extent each of the summands −A(G) and D(G) determines the properties of L(G), and motivated by the above works, we introduce the Aα−-matrix defined by
Aα−(G)=αD(G)+(α−1)A(G),0≤α≤1. | (1.1) |
Through our study, several facts indicate that the study of the Aα−(G) family is of a unique importance. First of all, obviously
A(G)=−A0−(G),D(G)=A1−(G)andL(G)=2A12−(G). |
Since A12−(G) is essentially equivalent to L(G), then the matrix Aα−(G) can be regarded as a generalization of the Laplacian matrix L(G). Since
A1(G)=A1−(G)=D(G); |
unless otherwise specified, we only consider the case of 0≤α<1 in this paper.
In this paper, we study some properties on Aα−(G) and obtain some bounds of the largest eigenvalue of Aα−(G). For the relation between the Q-spectrum and L-spectrum of a connected graph G, it is well known that G is bipartite if and only if Q(G) and L(G) has the same spectrum ([9, Proposition 1.3.10]), and we extend this relation on the Aα- and Aα−-spectrum. In [10], Akbari et al. proved that, for a graph G of order n and any real number β, if 0<β≤1 or 2≤β≤3, then
n∑k=1ρβk≥n∑k=1λβk; |
and if 1≤β≤2, then
n∑k=1ρβk≤n∑k=1λβk. |
In this work we present the relation between the sum of powers of Aα and Aα− eigenvalues, which is a generalization of the relation between the sum of powers of Q(G) and L(G) eigenvalues in [10].
The rest of this paper is organized as follows. In Section 2, we give some preliminaries and lemmas used later. In Section 3, we derive new basic properties of Aα−(G). In Section 4, we present the Aα−-spectra of the complete graphs and the complete split graphs. Section 5 gives some bounds of the largest eigenvalue of Aα−(G). In Section 6, we deduce the relation between the Aα- and Aα−-spectral radius. Finally, in Section 7, we prove the relation between the sum of powers of Aα and Aα− eigenvalues.
Although Weyl's inequalities have been known for almost a century (see, e.g., [11]), their equality case was first established by So in [12], and for convenience we state Weyl and So's complete theorem as follows:
Theorem 2.1. Let A and B be Hermitian matrices of order n, and let 1≤i≤n and 1≤j≤n. Then,
λi(A)+λj(B)≤λi+j−n(A+B),ifi+j≥n+1, | (2.1) |
λi(A)+λj(B)≥λi+j−1(A+B),ifi+j≤n+1. | (2.2) |
In either of these inequalities, the equality holds if and only if there exists a nonzero n-vector that is an eigenvector to each of the three eigenvalues involved.
A simplified version of (2.1) and (2.2) gives
λk(A)+λmin(B)≤λk(A+B)≤λk(A)+λ(B). | (2.3) |
We shall need the following lemmas for our new results.
Lemma 2.2. [3] Let G be a graph of order n. If e∈E(G) and α≥12, then,
ρi(Aα(G))≥ρi(Aα(G−e)),for1≤i≤n. |
Lemma 2.3. [13] Let B be a real symmetric nonnegative irreducible matrix and λ be the largest eigenvalue of B. Z∈Rn. If ZtBZ=λ and ‖Z‖=1, then BZ=λZ.
Lemma 2.4. [5] Let G be a connected graph with n≥4 vertices and m edges. If α∈[12,1), then,
ρα(G)≤max{αΔ(G),(1−α)(m−n−12)}+2α. |
Equality holds if and only if α=n−1n+1 and G=Kn.
For a graph G of order n, suppose that λ is an eigenvalue of Aα−(G) and x is an eigenvector of Aα−(G) with respect to λ. We use x(v) to denote the entry of x corresponding to the vertex v∈V(G). It is clear that the system of eigenequations for the matrix Aα−(G) is
λx(v)=αdG(v)x(v)+(α−1)∑u∈N(v)x(u). | (3.1) |
If G is a graph of order n with Aα−(G)=Aα−, and x is a real vector, the quadratic form ⟨Aα−x,x⟩ can be represented in several equivalent ways:
⟨Aα−x,x⟩=∑uv∈E(G)(αx(u)2+2(α−1)x(u)x(v)+αx(v)2), | (3.2) |
⟨Aα−x,x⟩=(2α−1)∑u∈V(G)x(u)2dG(u)+(1−α)∑uv∈E(G)(x(u)−x(v))2, | (3.3) |
⟨Aα−x,x⟩=α∑u∈V(G)x(u)2dG(u)+2(α−1)∑uv∈E(G)x(u)x(v). | (3.4) |
Each of these representations can be useful in proofs.
Now, we give some of the spectral properties of the Aα−-matrix. Let us call the largest eigenvalue of Aα−(G) the Aα−-spectral radius of G, and denote it as λα−(G). Let us also denote the smallest eigenvalue of Aα−(G) as μα−(G). Since Aα−(G) is a real symmetric matrix, and by using Rayleigh's principle, the following result holds:
Proposition 3.1. If α∈[0,1) and G is a graph of order n, then
λα−(G)=max‖x‖2=1⟨Aα−(G)x,x⟩andμα−(G)=min‖x‖2=1⟨Aα−(G)x,x⟩. | (3.5) |
Moreover, if x is a unit n-vector, then,
λα−(G)=⟨Aα−(G)x,x⟩ |
if and only if x is an eigenvector to λα−(G), and
μα−(G)=⟨Aα−(G)x,x⟩ |
if and only if x is an eigenvector to μα−(G).
By using these relations, the following result is evident:
Proposition 3.2. If α∈[0,1) and G is a graph, then
λα−(G)=max{λα−(H):HisacomponentofG}, |
μα−(G)=min{μα−(H):HisacomponentofG}. |
It is clear that if G is a d-regular graph of order n, then
Aα−(G)=αdIn+(α−1)A(G), |
and so there is a linear correspondence between the spectra of Aα−(G) and of A(G),
λk(Aα−(G))=αd+(α−1)λn−k+1(A(G)), 1≤k≤n. | (3.6) |
In particular, if k=n, then
μα−(G)=(2α−1)d |
for any α∈[0,1]. As a consequence of Weyl's inequality (2.3), the following result is immediate:
Proposition 3.3. If α∈[0,1] and G is a graph with
A(G)=AandAα−(G)=Aα−, |
then,
αδ+(α−1)λn−k+1(A)≤λk(Aα−)≤αΔ+(α−1)λn−k+1(A). |
An important property of the Laplacian L(G) is that it is positive semidefinite. This is certainly not true for Aα−(G) if α<12 and G is a regular graph, but if α≥12, then Aα−(G) is like L(G). We give this result in the following:
Proposition 3.4. If α≥12, and G is a graph, then Aα−(G) is positive semidefinite, and if α>12 and G has no isolated vertices, then Aα−(G) is positive definite. Moreover, if α<12 and G is a regular graph, then Aα−(G) is not positive semidefinite.
Proof. Let x be a nonzero vector. If α≤12, then for any edge uv∈E(G), we see that
⟨Aα−(G)x,x⟩≥(2α−1)(x(u)2+x(v)2)+(1−α)(x(u)−x(v))2≥0. | (3.7) |
Hence, Aα−(G) is positive semidefinite. Now, assume that G has no isolated vertices. Choose a vertex u with x(u)≠0 and let uv∈E(G). Then, (3.7) becomes a strict inequality, and so Aα−(G) is positive definite. Finally, suppose that G is a d−regular graph, α<12, and let A be its adjacency matrix, then λ(A)=d. Thus
μα−(G)=(2α−1)d<0. |
By the above proposition we get the following lemma which gives a relation between the Aα− eigenvalues of G and the Aα− eigenvalues of spanning subgraphs of G.
Lemma 3.5. Let G be a graph of order n and let α∈[12,1). If G′=G−e, where e∈E(G), then,
λi(G′)≤λi(G)for1≤i≤n. |
Proof. Let e=uv such that u,v∈V(G). It is easy to see that
Aα−(G)=Aα−(G′)+M, |
where M is the matrix of order n indexed by the vertices of G having (u,v)th and (v,u)th entries both equal to α−1, and the (u,u)th and (v,v)th entries both equal to α and all other entries equal to zero, hence M is an Aα−-matrix of a graph containing only one edge. Since α∈[12,1), then Aα−(G), Aα−(G′) and M are positive semidefinite and Weyl's inequalities (2.3) imply that λi(G′)≤λi(G).
Equality (3.6) and the fact the eigenvalues of A(Kn) are {n−1,−1,⋯,−1} give the spectrum of Aα−(Kn) as follows:
Proposition 4.1. The eigenvalues of Aα−(Kn) are
μα−(Kn)=(2α−1)(n−1)andλk(Aα−(Kn))=α(n−2)+1for1≤k≤n−1. |
If S⊆V(G), then we use G[S] to denote the subgraph of G induced by S. Recall that G[S] is an independent set if no two vertices of S are adjacent and G[S] is a clique if it is a complete subgraph of G. The graph Kr∨(n−r)K1 is called a complete split graph, denoted by CSr,n−r. The work in the following proposition is motivated by the proof of [3, Proposition 2.4].
Proposition 4.2. Let G be a graph with Aα−(G), and α∈[0,1). Let S⊆V(G) and |S|=k. Suppose that dG(u)=d for each vertex u∈S, and N(v)∖{w}=N(w)∖{v} for any two vertices v,w∈S. Then, we have the following statements:
(i) If G[S] is a clique, then α(d−1)+1 is an eigenvalue of Aα−(G) with multiplicity at least k−1.
(ii) If G[S] is an independent set, then αd is an eigenvalue of Aα−(G) with multiplicity at least k−1.
Proof. Let S={v1,v2,⋯,vk}. Clearly, d1=⋯=dk=d. Let z1,z2,⋯,zk−1 be vectors such that
{zi(v1)=1,zi(vi+1)=−1,zi(v)=0,if v∈V(G)∖{v1,vi+1}, |
for i=1,⋯,k−1. Assume that G[S] is a clique. It is easy to obtain that
Aα−(G)zi=(α(d1−1)+1,0,⋯,0,α(1−di)−1,0,⋯,0)′=(α(d−1)+1)zi, |
for i=1,⋯,k−1. Hence, α(d−1)+1 is an eigenvalue of Aα−(G) and z1,z2,⋯,zk−1 are eigenvectors of Aα−(G) corresponding to α(d−1)+1. In addition, since z1,z2,⋯,zk−1 are linearly independent, the multiplicity of α(d−1)+1 is at least k−1. Now, supposing that G[S] is an independent set, we have
Aα−(G)zi=(αd1,0,⋯,0,−αdi,0,⋯,0)′=αdzi, |
for i=1,⋯,k−1. Since z1,z2,⋯,zk−1 are linearly independent, it follows that αd is an eigenvalue of Aα−(G) with multiplicity at least k−1. Thus, the proof is completed.
Consider an n×n real symmetric matrix
S=(S11S12⋯S1tS21S22⋯S2t⋮⋮⋱⋮St1St2⋯Stt), |
whose rows and columns are partitioned according to a partitioning P1, P2, ⋯, Pt of {1,2,⋯,n}. The quotient matrix B of the matrix S is the t×t matrix whose entries are the average row sums of the blocks Sij of S. The partition is equitable if each block Sij of S has constant row sum.
Lemma 4.3. [14] Let B be an equitable quotient matrix of a symmetric real matrix S. If λ is an eigenvalue of B, then λ is also an eigenvalue of S.
Now, we can determine all Aα−-eigenvalues of CSr,n−r as follows:
Proposition 4.4. The Aα−-spectrum of CSr,n−r contains α(n−2)+1 with multiplicity r−1, αr with multiplicity n−r−1, and the remaining two Aα−-eigenvalues are
α(n+2(r−1))+1±√(α(n+2(r−1))+1)2+4r(1−2α)(n−r+α(r−1))2. | (4.1) |
Proof. We can write Aα−(CSr,n−r) as
Aα−(CSr,n−r)=((1−α)Jr,r+(αn−1)Ir(1−α)Jr,n−r(1−α)Jn−r,rαrIn−r,n−r). |
Then, the quotient matrix of Aα−(CSr,n−r) is equitable and it can be written in the form
B(Aα−(CSr,n−r))=((n−r)α+r−1(1−α)(n−r)(1−α)rαr). |
Thus, by Lemma 4.3, the eigenvalues of B(Aα−(CSr,n−r)) are eigenvalues of Aα−(CSr,n−r), and according to Proposition 4.2, we get the result.
In this section we give a few general bounds on λα−(G).
Proposition 5.1. Let G be a graph, with
Δ(G)=Δ, A(G)=AandD(G)=D. |
The following inequalities hold for λα−(G):
λα−(G)≥αΔ+(α−1)λ(A), | (5.1) |
λα−(G)≥(2α−1)λ(A). | (5.2) |
Proof. Inequality (5.1) follows by Weyl's inequalities (2.3) because
αΔ+(α−1)λ(A)=αλ(D)+(α−1)λ(A)=λ(αD)+λmin((α−1)A)≤λ(αD+(α−1)A)=λα−(G). |
To prove the inequality (5.2), let H be a component of G such that λ(A)=λ(A(H)). Let x be a positive unit vector to λ(A(H)). For every edge uv∈E(H), the AM-GM inequality implies that
2x(u)x(v)=2αx(u)x(v)+2(1−α)x(u)x(v)≤α(x(u)+x(v))22+2(1−α)x(u)x(v)=12(αx(u)2+2(α−1)x(u)x(v)+αx(v)2)+(3−2α)x(u)x(v). |
Summing this inequality over all edges uv∈E(H), and using (3.2), we get
λ(A)=λ(A(H))=⟨A(H)x,x⟩≤12⟨Aα−(H)x,x⟩+12(3−2α)λ(A); |
and then we get
2λ(A)−(3−2α)λ(A)≤⟨Aα−(H)x,x⟩≤λα−(G); |
hence
(2α−1)λ(A)≤λα−(G). |
Having inequality (5.2) in hand, if α≥12, then every lower bound of λ(A) gives a lower bound on λα−(G), but if α<12, then every upper bound of λ(A) gives a lower bound on λα−(G). We mention just two such bounds.
Corollary 5.2. Let G be a graph such that α≥12. If G is of order n and has m edges, then
λα−(G)≥(2α−1)2mn. |
Corollary 5.3. Let G be a connected graph such that α<12. If G is of order n and has m edges, then
λα−(G)≥(2α−1)√2m−n+1. |
Proposition 5.4. Let G be a graph of order n. If α∈(12,1], then,
λα−(G)≤α(n−2)+1. |
Moreover,
λα−(G)=α(n−2)+1 |
with multiplicity k−1 if G has k vertices of degree n−1.
Proof. Applying Lemma 3.5 leads to
λα−(G)≤λα−(Kn)=α(n−2)+1. |
If G has k vertices of degree n−1, then it follows from Proposition 4.2 that α(n−2)+1 is an eigenvalue of Aα−(G) with multiplicity at least k−1, and since
λα−(G)≤α(n−2)+1, |
we get
λα−(G)=α(n−2)+1 |
with multiplicity k−1.
Let G be a connected graph. Merris [15] pointed out λ(L(G))≤ρ(Q(G)), and the equality holds if G is a bipartite graph. In the next result, we generalize this result to the Aα−- and Aα-spectral radius of a connected graph.
Theorem 6.1. Let G be a graph of order n, α∈(0,1), λα−(G)=λα− and ρα(G)=ρα. We have λα−≤ρα. Moreover, if G is connected, then the equality holds if and only if G is bipartite.
Proof. Let
V(G)={v1,v2,⋯,vn}, x=(x1,x2,⋯,xn)t∈Rn |
be an arbitrary vector such that ‖x‖=1. Let
y=(y1,y2,⋯,yn)t∈Rn |
be a unit eigenvector of Aα−(G) belonging to λα− and
y′=(y′1,y′2,⋯,y′n)t∈Rn |
be a unit eigenvector of Aα(G) belonging to ρα. Let
|y|=(|y1|,|y2|,⋯,|yn|)t. |
First, we prove that λα−≤ρα.
λα−=maxxtAα−x=maxxt(αD+(α−1)A)x=max(αn∑i=1x2idi+2(α−1)∑vivj∈E(G)xixj)=ytAα−y=αn∑i=1y2idi+2(α−1)∑vivj∈E(G)yiyj |
and
ρα=maxxtAαx=maxxt(αD+(1−α)A)x=max(αn∑i=1x2idi+2(1−α)∑vivj∈E(G)xixj)=y′tAαy′=αn∑i=1y′i2di+2(1−α)∑vivj∈E(G)y′iy′j. |
Thus,
λα−=αn∑i=1y2idi+2(α−1)∑vivj∈E(G)yiyj≤αn∑i=1y2idi+2(1−α)∑vivj∈E(G)|yiyj|=|yt|Aα|y|≤maxxtAαx=ρα. | (6.1) |
Now, if G is bipartite, then the matrix Aα− and the matrix Aα are similar by a diagonal matrix D′ with diagonal entries ±1, that is, Aα=D′Aα−D′−1. Therefore, Aα− and Aα have the same spectrum, and thus we get λα−=ρα. Finally, when G is connected and λα−=ρα, all inequalities (6.1) must be equalities. By Lemma 2.3 and the equality
|yt|Aα|y|=ρα; |
we know that |y| is an eigenvector of Aα belonging to ρα. So, |y|=±y′. Using the Perron-Frobenius' theorem for Aα(G), we have y′>0, |y|=y′, and |yi|>0, i=1,2,⋯,n.
Since
αn∑i=1y2idi+2(α−1)∑vivj∈E(G)yiyj=αn∑i=1y2idi+2(1−α)∑vivj∈E(G)|yiyj|; |
we get
−∑vivj∈E(G)yiyj=∑vivj∈E(G)|yiyj|, |
hence, |yiyj|=−yiyj when vivj∈E(G). Therefore, yiyj<0 if vivj∈E(G).
Let
U={vi:yi>0}andW={vj:yj<0}. |
For each edge e=vivj, we have yiyj<0, one of the vertices of edge e is in U, and the other is in W. So, G is a bipartite graph.
Remark 6.2. If G is not bipartite and is not connected, and λα− is the largest eigenvalue of Aα− for a bipartite connected component of G, then the equality in Theorem 6.1 still holds.
Example 6.3. Take
G=C3∪C4. |
Then, G is not bipartite and not connected. Now, A14−(G) has a spectrum 2[1],1.25[2],0.5[2],−1[2] (where λ[i] means the eigenvalue λ is repeated i times in the spectrum). On the other hand, A14(G) has a spectrum 2[2],0.5[2],−0.25[2],−1[1]. Then, we have
λ14−(G)=ρ14(G)=2. |
Note that 2 is the largest eigenvalue of A14−(C4) as well.
Now, we introduce the relationship between the Aα−- and Aα-spectra of bipartite graphs, which is a generalization [9, Proposition 1.3.10], and it follows from the proof of the above theorem.
Corollary 6.4. Let G be a connected graph. Then, G is bipartite if and only if the Aα−-spectrum and Aα-spectrum are equal.
Remark 6.5. In fact, if G is bipartite and is not connected, the Aα−-spectrum still equals the Aα-spectrum.
Example 6.6. Take G=P3∪C4. Then G is bipartite and not connected. We have that A34−(G) has a spectrum 2[1],1.64039[1],1.5[2],1[1],0.75[1],0.609612[1]. In contrast, A34(G) has the same spectrum as A34−(G), although G is not connected.
According to Corollary 6.4 and [1, Propositions 38 and 39], we get the following two results:
Corollary 6.7. The Aα−-spectrum and the Aα-spectrum of the complete bipartite graph Ka,b are equal, that is, if a≥b≥1 and α∈(0,1), the eigenvalues of Aα−(Ka,b) are
λα−(Ka,b)=12(α(a+b)+√α2(a+b)2+4ab(1−2α)), |
μα−(Ka,b)=12(α(a+b)−√α2(a+b)2+4ab(1−2α)), |
λk(Aα−(Ka,b))=αafor1<k≤b, |
λk(Aα−(Ka,b))=αbforb<k<a+b. |
Corollary 6.8. The Aα−-spectrum and the Aα-spectrum of the star K1,n−1 are equal, that is, the eigenvalues of Aα−(K1,n−1) are
λα−(K1,n−1)=12(αn+√α2n2+4(n−1)(1−2α)), |
μα−(K1,n−1)=12(αn−√α2n2+4(n−1)(1−2α)), |
λk(Aα−(K1,n−1))=αfor1<k<n. |
Indeed, many practical results of Theorem 6.1 can be deduced, and here are some of them.
Proposition 6.9. Let G be a connected graph with n≥4 vertices and m edges. If α∈[12,1), then
λα−(G)<max{αΔ(G),(1−α)(m−n−12)}+2α. | (6.2) |
Proof. Let
λα−(G)=λα−andρα(G)=ρα. |
Then, Theorem 6.1 and Lemma 2.4 lead to
λα−≤max{αΔ(G),(1−α)(m−n−12)}+2α. | (6.3) |
Suppose that the equality in (6.3) holds, thus ρα=λα−, and so G is bipartite and
ρα=max{αΔ(G),(1−α)(m−n−12)}+2α. |
Therefore, by Lemma 2.4, G=Kn, and thus G is bipartite if and only if n=2, but n≥4, and hence the inequality is strict.
Theorem 6.1 and [2, Theorem 2] lead directly to the next result.
Proposition 6.10. If T is a tree of order n and α∈[0,1], then
λα−(T)≤nα+√n2α2+4(n−1)(1−2α)2. |
The equality holds if and only if T is the star K1,n−1.
By Theorem 6.1 and [1, Proposition 20], we get the next result.
Proposition 6.11. If G is a graph with no isolated vertices, then
λα−(G)≤maxu∈V(G){αd(u)+1−αd(u)∑uv∈E(G)d(v)}. |
If α∈(12,1) and G is connected, the equality holds if and only if G is a regular bipartite graph.
In this part we give two explicit expressions for the sums and the sum of squares of the eigenvalues of Aα− and Aα, considering [1, Propositions 34 and 35].
Proposition 6.12. If G is a graph of order n and has m edges, then
n∑i=1λi(Aα−(G))=n∑i=1λi(Aα(G))=trAα−(G)=trAα(G); |
where
n∑i=1λi(Aα−(G))=α∑u∈V(G)d(u)=2αm. |
A similar formula for the sum of the squares of the Aα− and Aα-eigenvalues is given as follows:
Proposition 6.13. If G is a graph of order n and has m edges, then
n∑i=1λ2i(Aα−(G))=n∑i=1λ2i(Aα(G))=trA2α−(G)=trA2α(G); |
where
n∑i=1λ2i(Aα−(G))=α2∑u∈V(G)d2(u)+2(1−α)2m. |
Proof. Let
Aα−=Aα−(G), A=A(G)andD=D(G). |
Calculating the square A2α− and taking its trace, we find that
trA2α−=tr(α2D2+(1−α)2A2+α(α−1)DA+α(α−1)AD)=α2trD2+(1−α)2trA2−α(1−α)trDA+α(α−1)trAD=α2∑u∈V(G)d2(u)+2(1−α)2m. |
Pirzada et al. [16] introduced the sum of the βth powers of the Aα eigenvalues of G as
Sαβ(G)=n∑i=1ρβi. |
Now, we have the notation
Sα−β(G)=n∑i=1λβi |
for the sum of the βth powers of the Aα− eigenvalues of G. The following theorem is a generalization [10, Theorem 2].
Theorem 7.1. Let G be a graph of order n>1, α∈[12,1), and let β be a real number.
(i) If 0<β≤1 or 2≤β≤3, then
Sαβ(G)≥Sα−β(G). |
(ii) If 1≤β≤2, then
Sαβ(G)≤Sα−β(G). |
For β∈(0,1)∪(2,3), the equality holds in (i) if and only if G is a bipartite graph. Moreover, for β∈(1,2), the equality holds in (ii) if and only if G is a bipartite graph.
Proof. We recall that, for any real number r, the binomial series
∞∑k=0(rk)xk |
converges to (1+x)r if |x|<1. This also remains true for x=−1 if r>0 (see, e.g., [17, p. 419]). By Lemma 3.5, we find that,
λα−(G)≤λα−(Kn)=α(n−2)+1andρα(G)≤ρα(Kn)=n−1. |
Since Aα is positive semidefinite when α∈[12,1), then if ρi>0, we have
|ρin−1|≤|n−1n−1|=|−1n|<1 |
and if ρi=0, we get
ρin−1=−1. |
Therefore,
Sαβ(G)nβ=(ρ1n)β+⋯+(ρnn)β=∞∑k=0(βk)(ρ1n−1)k+⋯+∞∑k=0(βk)(ρnn−1)k=∞∑k=0(βk)tr(1n(αD+(1−α)A)−I)k. |
Also, since Aα− is positive semidefinite when α∈[12,1), then if λi>0,
|λin−1|≤|α(n−2)+1n−1|=|(α−1)+1−2αn|<1 |
and if λi=0, we have
λin−1=−1. |
Thus, in a similar manner as above, we obtain that
Sα−β(G)nβ=∞∑k=0(βk)tr(1n(αD+(α−1)A)−I)k. |
We claim that
ifkis even, tr(αD+(1−α)A−nI)k≤tr(αD+(α−1)A−nI)k; |
ifkis odd, tr(αD+(1−α)A−nI)k≥tr(αD+(α−1)A−nI)k. |
When ((αD−nI)+(1−α)A)k and ((αD−nI)+(α−1)A)k are expanded in terms of the powers of αD−nI and (1−α)A, respectively, the terms appearing in both expansions, regardless of their signs, are the same. To prove this claim, we identify the sign of each term in both expansions. Consider the terms in the expansion of ((αD−nI)+(1−α)A)k, where there are exactly j factors equal to αD−nI, for some j=0,1,⋯,k. The sign of the trace for each of these terms is (−1)j or 0 because all entries of αD−nI and (1−α)A are non-positive and non-negative, respectively. On the other hand, in each term in the expansion of ((αD−nI)+(α−1)A)k all factors are matrices with non-positive entries, hence the sign of the trace of each term is (−1)k or 0. Therefore, the claim has been proven.
Now, note that if 0<β<1 or 2<β<3, then the sign of (βk) is (−1)k−1, except that (β2)>0, for 2<β<3. According to this, for 0<β<1 and every k,
(βk)tr(αD+(1−α)A−nI)k≥(βk)tr(αD+(α−1)A−nI)k. |
This inequality remains true for 2≤β≤3 as
tr(αD+(1−α)A−nI)2=tr(αD+(α−1)A−nI)2, |
since trA2α=trA2α−. Thus, Part (i) is proved. For 1<β<2, the sign of (βk) is (−1)k−1, except that (β1)>0. Since trAα=trAα−, we have
tr(αD+(1−α)A−nI)=tr(αD+(α−1)A−nI), |
and so part (ii) is similarly proved.
Now, we examine the equality case. Since Aα and Aα− are similar if G is bipartite, it follows that the equality holds in both (i) and (ii). Since for any positive integer i, trAi equals the total number of closed walks of length i in G, then if G is not bipartite, there exists an odd integer r such that trAr>0 (see [18, Lemma 2.5]). Hence,
tr(αD+(1−α)A−nI)r>tr(αD+(α−1)A−nI)r; |
and so the inequalities in both (i) and (ii) are strict.
We know that the Aα-spectra and Aα−-spectra of Kn are {(n−1)[1],(αn−1)[n−1]} and {((2α−1)(n−1))[1],(α(n−2)+1)[n−1]}, respectively. Therefore,
Sαβ(Kn)−Sα−β(Kn)=(n−1)β+(n−1)(αn−1)β−((2α−1)(n−1))β−(n−1)(α(n−2)+1)β. |
By Theorem 7.1 and based on our numerical experiments, we propose the following conjecture:
Conjecture 7.2. For every α∈(12,1) and each integer n≥3, we have
Sαβ(Kn)−Sα−β(Kn)≥0 |
for any β∈[0,1]∪[2,∞) and
Sαβ(Kn)−Sα−β(Kn)≤0 |
for any β∈(−∞,0)∪(1,2).
In this paper, we have introduced the Aα−-matrix of a graph G, which is a generalization of the Laplacian matrix L(G), and we have studied the basic properties of Aα−; and derived some bounds for its spectral radius. Furthermore, we have determined the Aα−-spectra for the complete graph and the complete split graph. Building upon previous results, we have extended findings related to the spectral radius of L(G) and Q(G) matrices to the Aα- and Aα−-spectral radius. Specifically, in Theorem 6.1, we have generalized Merris' [15] observation that λ(L(G))≤ρ(Q(G)) with equality holding for bipartite graphs. Additionally, we have extended the known relation that G is bipartite if and only if Q(G) and L(G) share the same spectrum, as demonstrated in Corollary 6.4. Finally, in Theorem 7.1; we have generalized a relation established by S. Akbari et al. in [10]; which relates the sum of powers of the eigenvalues of Q(G) and L(G). In conclusion, these findings have implications for various applications involving graph analysis.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
The authors are grateful to the referees for their valuable suggestions and comments.
The authors declare no conflicts of interest.
[1] | K. Liu, S. Xu, G. Xu, M. Zhang, D. Sun, H. Liu, A review of android malware detection approaches based on machine learning, IEEE Access, 8 (2020). https://doi.org/10.1109/ACCESS.2020.3006143 |
[2] | B. Amos, H. Turner, J. White, Applying machine learning classifiers to dynamic Android malware detection at scale, In: 2013 9th International Wireless Communications and Mobile Computing Conference (IWCMC), IEEE, Italy, 2013, 1666–1671. https://doi.org/10.1109/IWCMC.2013.6583806 |
[3] | M. Egele, T. Scholte, E. Kirda, C. Kruegel, A survey on automated dynamic malware-analysis techniques and tools, ACM Comput. Surv., 44 (2012), 1–42. |
[4] | B. Amro, Malware detection techniques for mobile devices, Int. J. Mobile Netw. Commun. Telemat., 7 (2017). https://doi.org/10.1145/2089125.2089126 |
[5] | K. Kavitha, P. Salini, V. Ilamathy, Exploring the malicious Android applications and reducing risk using static analysis, In: 2016 International Conference on Electrical, Electronics, and Optimization Techniques (ICEEOT), IEEE, India, 2016, 1316–1319. https://doi.org/10.1109/ICEEOT.2016.7754896 |
[6] |
E. M. B. Karbab, M. Debbabi, MalDy: Portable, data-driven malware detection using natural language processing and machine learning techniques on behavioral analysis reports, Digit. Invest., 28 (2019), 77–87. https://doi.org/10.1016/j.diin.2019.01.017 doi: 10.1016/j.diin.2019.01.017
![]() |
[7] | R. Ito, M. Mimura, Detecting unknown malware from ASCII strings with natural language processing techniques, In: 2019 14th Asia Joint Conference on Information Security (AsiaJCIS), IEEE, Japan, 2019. https://doi.org/10.1109/AsiaJCIS.2019.00-12 |
[8] | P. Najafi, D. Koehler, F. Cheng, C. Meinel, NLP-based entity behavior analytics for malware detection, In: 2021 IEEE International Performance, Computing, and Communications Conference (IPCCC), IEEE, USA, 2021. https://doi.org/10.1109/IPCCC51483.2021.9679411 |
[9] | U. Raghav, E. Martinez-Marroquin, W. Ma, Static analysis for Android Malware detection with document vectors, In: 2021 International Conference on Data Mining Workshops (ICDMW), IEEE, New Zealand, 2021. https://doi.org/10.1109/ICDMW53433.2021.00104 |
[10] |
X. Xing, X. Jin, H. Elahi, H. Jiang, G. Wang, A malware detection approach using autoencoder in deep learning, IEEE Access, 10 (2022), 25696–25706. https://doi.org/10.1109/ACCESS.2022.3155695 doi: 10.1109/ACCESS.2022.3155695
![]() |
[11] |
Q. Le, O. Boydell, B. Mac, M. Scanlon, Deep learning at the shallow end: Malware classification for non-domain experts, Digit. Invest., 26 (2018), S118–S126. http://dx.doi.org/10.1016/j.diin.2018.04.024 doi: 10.1016/j.diin.2018.04.024
![]() |
[12] |
J. Y. Kim, S. J. Bu, S. B. Cho, Zeroday malware detection using transferred generative adversarial networks based on deep autoencoders, Inform. Sci., 460–461 (2018), 83–102. https://doi.org/10.1016/j.ins.2018.04.092 doi: 10.1016/j.ins.2018.04.092
![]() |
[13] | I. Goodfellow, NIPS 2016 Tutorial: Generative adversarial networks, arXiv preprint, 2014. https://doi.org/10.48550/arXiv.1701.00160 |
[14] | S. Kumar, B. Janet, DTMIC: Deep transfer learning for malware image classification, J. Inf. Secur. Appl., 64 (2022). https://doi.org/10.1016/j.jisa.2021.103063 |
[15] |
Ö. Aslan, A. A. Yilmaz, A new malware classification framework based on deep learning algorithms, IEEE Access, 9 (2021), 87936–87951. https://doi.org/10.1109/ACCESS.2021.3089586 doi: 10.1109/ACCESS.2021.3089586
![]() |
[16] |
F. Rustam, I. Ashraf, A. D. Jurcut, A. K. Bashir, Y. B. Zikria, Malware detection using image representation of malware data and transfer learning, J. Parallel Distr. Com., 172 (2023), 32–50. https://doi.org/10.1016/j.jpdc.2022.10.001 doi: 10.1016/j.jpdc.2022.10.001
![]() |
[17] | T. Li, Y. Luo, X. Wan, Q. Li, Q. Liu, R. Wang, et al., A malware detection model based on imbalanced heterogeneous graph embeddings, Expert Syst. Appl., 246 (2014), 123109. |
[18] | Google play store. Available from: https://https://play.google.com/store/apps. |
[19] | Virusshare. Available from: http://virusshare.com/. |
[20] | Virustotal. Available from: https://www.virustotal.com/gui/home/upload. |
[21] | L. Nataraj, S. Karthikeyan, G. Jacob, B. S. Manjunath, Malware images: Visualization and automatic classification, In: Proceedings of the 8th International Symposium on Visualization for Cyber Security, 2011, 1–7. https://doi.org/10.1145/2016904.2016908 |
[22] | A. S. Bozkir, A. O. Cankaya, M. Aydos, Utilization and comparison of convolutional neural networks in malware recognition, In: 2019 27th Signal Processing and Communications Applications Conference (SIU), IEEE, Turkey, 2019, 1–4. https://doi.org/10.1109/SIU.2019.8806511 |
[23] | MaleVis. Available from: https://web.cs.hacettepe.edu.tr/selman/malevis/. |
[24] | S. Venkatraman, M. Alazab, R. Vinayakumar, A hybrid deep learning image-based analysis for effective malware detection, J. Inf. Secur. Appl., 47 (2019), 377–389. |
[25] | A. Krizhevsky, I. Sutskever, G. E. Hinton, ImageNet classification with deep convolutional neural networks, Adv. Neural Inform. Proc. Syst., 2012. |
[26] | K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition, arXiv preprint, 2014. https://doi.org/10.48550/arXiv.1409.1556 |
[27] | K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016,770–778. |
[28] | C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, Z. Wojna, Rethinking the inception architecture for computer vision, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, 2818–2826. |
[29] | G. Huang, Z. Liu, L. Van Der Maaten, K. Q. Weinberger, Densely connected convolutional networks, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, 4700–4708. |
[30] | J. Yosinski, J. Clune, Y. Bengio, H. Lipson, How transferable are features in deep neural networks? In: Advances in Neural Information Processing Systems (NIPS), 2014. |
[31] |
S. J. Pan, Q. Yang, A survey on transfer learning, IEEE T. Knowl. Data Eng., 22 (2010), 1345–1359. https://doi.org/10.1109/TKDE.2009.191 doi: 10.1109/TKDE.2009.191
![]() |
[32] |
R. Vinayakumar, M. Alazab, K. P. Soman, P. Poornachandran, S. Venkatraman, Robust intelligent malware detection using deep learning, IEEE Access, 7 (2019), 46717–46738. https://doi.org/10.1109/ACCESS.2019.2906934 doi: 10.1109/ACCESS.2019.2906934
![]() |
[33] | J. S. Luo, D. C. T. Lo, Binary malware image classification using machine learning with local binary pattern, In: 2017 IEEE International Conference on Big Data (Big Data), IEEE, USA, 2017, 4664–4667. https://doi.org/10.1109/BigData.2017.8258512 |
[34] |
Z. Cui, F. Xue, X. Cai, Y. Cao, G. G. Wang, J. Chen, Detection of malicious code variants based on deep learning, IEEE T. Ind. Inform., 14 (2018), 3187–3196. https://doi.org/10.1109/tii.2018.2822680 doi: 10.1109/tii.2018.2822680
![]() |
[35] | D. Gibert, Convolutional neural networks for malware classification, M.S. thesis, Univ. Rovira Virgili, Tarragona, Spain, 2016. |
[36] | A. Singh, A. Handa, N. Kumar, S. K. Shukla, Malware classification using image representation, In: Proc. Int. Symp. Cyber Secur. Cryptogr. Mach. Learn. Cham, Switzerland: Springer, 2019, 75–92. https://doi.org/10.1007/978-3-030-20951-3_6 |