Processing math: 100%
Research article Special Issues

Leveraging deep learning and image conversion of executable files for effective malware detection: A static malware analysis approach

  • The escalating sophistication of malware poses a formidable security challenge, as it evades traditional protective measures. Static analysis, an initial step in malware investigation, involves code scrutiny without actual execution. One static analysis approach employs the conversion of executable files into image representations, harnessing the potency of deep learning models. Convolutional neural networks (CNNs), particularly adept at image classification, have potential for malware detection. However, their inclination towards structured data requires a preprocessing phase to convert software into image-like formats. This paper outlines a methodology for malware detection that involves applying deep learning models to image-converted executable files. Experimental evaluations have been performed by using CNN models, autoencoder-based models, and pre-trained counterparts, all of which have exhibited commendable performance. Consequently, employing deep learning for image-converted executable analysis emerges as a fitting strategy for the static analysis of software. This research is significant because it utilized the largest dataset to date and encompassed a wide range of deep learning models, many of which have not previously been tested together.

    Citation: Mesut GUVEN. Leveraging deep learning and image conversion of executable files for effective malware detection: A static malware analysis approach[J]. AIMS Mathematics, 2024, 9(6): 15223-15245. doi: 10.3934/math.2024739

    Related Papers:

    [1] Maha M. Althobaiti, José Escorcia-Gutierrez . Weighted salp swarm algorithm with deep learning-powered cyber-threat detection for robust network security. AIMS Mathematics, 2024, 9(7): 17676-17695. doi: 10.3934/math.2024859
    [2] Ibrahim R. Alzahrani, Randa Allafi . Integrating Ebola optimization search algorithm for enhanced deep learning-based ransomware detection in Internet of Things security. AIMS Mathematics, 2024, 9(3): 6784-6802. doi: 10.3934/math.2024331
    [3] Sultanah M. Alshammari, Nofe A. Alganmi, Mohammed H. Ba-Aoum, Sami Saeed Binyamin, Abdullah AL-Malaise AL-Ghamdi, Mahmoud Ragab . Hybrid arithmetic optimization algorithm with deep learning model for secure Unmanned Aerial Vehicle networks. AIMS Mathematics, 2024, 9(3): 7131-7151. doi: 10.3934/math.2024348
    [4] Alaa O. Khadidos . Advancements in remote sensing: Harnessing the power of artificial intelligence for scene image classification. AIMS Mathematics, 2024, 9(4): 10235-10254. doi: 10.3934/math.2024500
    [5] Khaled Tarmissi, Hanan Abdullah Mengash, Noha Negm, Yahia Said, Ali M. Al-Sharafi . Explainable artificial intelligence with fusion-based transfer learning on adverse weather conditions detection using complex data for autonomous vehicles. AIMS Mathematics, 2024, 9(12): 35678-35701. doi: 10.3934/math.20241693
    [6] Ebtesam Al-Mansor, Mohammed Al-Jabbar, Arwa Darwish Alzughaibi, Salem Alkhalaf . Dandelion optimization based feature selection with machine learning for digital transaction fraud detection. AIMS Mathematics, 2024, 9(2): 4241-4258. doi: 10.3934/math.2024209
    [7] Mashael M Asiri, Abdelwahed Motwakel, Suhanda Drar . Robust sign language detection for hearing disabled persons by Improved Coyote Optimization Algorithm with deep learning. AIMS Mathematics, 2024, 9(6): 15911-15927. doi: 10.3934/math.2024769
    [8] Thavavel Vaiyapuri, Prasanalakshmi Balaji, S. Shridevi, Santhi Muttipoll Dharmarajlu, Nourah Ali AlAseem . An attention-based bidirectional long short-term memory based optimal deep learning technique for bone cancer detection and classifications. AIMS Mathematics, 2024, 9(6): 16704-16720. doi: 10.3934/math.2024810
    [9] Mohammed Aljebreen, Hanan Abdullah Mengash, Khalid Mahmood, Asma A. Alhashmi, Ahmed S. Salama . Enhancing cybersecurity in cloud-assisted Internet of Things environments: A unified approach using evolutionary algorithms and ensemble learning. AIMS Mathematics, 2024, 9(6): 15796-15818. doi: 10.3934/math.2024763
    [10] Wahida Mansouri, Amal Alshardan, Nazir Ahmad, Nuha Alruwais . Deepfake image detection and classification model using Bayesian deep learning with coronavirus herd immunity optimizer. AIMS Mathematics, 2024, 9(10): 29107-29134. doi: 10.3934/math.20241412
  • The escalating sophistication of malware poses a formidable security challenge, as it evades traditional protective measures. Static analysis, an initial step in malware investigation, involves code scrutiny without actual execution. One static analysis approach employs the conversion of executable files into image representations, harnessing the potency of deep learning models. Convolutional neural networks (CNNs), particularly adept at image classification, have potential for malware detection. However, their inclination towards structured data requires a preprocessing phase to convert software into image-like formats. This paper outlines a methodology for malware detection that involves applying deep learning models to image-converted executable files. Experimental evaluations have been performed by using CNN models, autoencoder-based models, and pre-trained counterparts, all of which have exhibited commendable performance. Consequently, employing deep learning for image-converted executable analysis emerges as a fitting strategy for the static analysis of software. This research is significant because it utilized the largest dataset to date and encompassed a wide range of deep learning models, many of which have not previously been tested together.



    Let

    G=(V(G),E(G))

    be a simple undirected graph on n vertices such that V(G) and E(G) denote the vertex set and edge set of G, respectively. Let V(G)={v1,v2,,vn}, for viV(G), and dG(vi)=d(vi)=di denotes the degree of vertex vi and the set of vertices adjacent to vi, denoted by N(vi), refers to the neighborhood of vi. The maximum and the minimum degree of G are denoted by Δ=Δ(G) and δ=δ(G), respectively. λ(M), λk(M) and λmin(M) denote the largest eigenvalue, the k-th largest eigenvalue, and the smallest eigenvalue of a matrix M, respectively. The set of all eigenvalues for a matrix M together with their multiplicities is called the M-spectrum. A real symmetric matrix M is called positive semidefinite if λmin(M)0. Denote by Kn, Pn, Cn and Ks,ns the complete graph, path, cycle, and complete bipartite graph with n vertices, respectively. Let A(G) and D(G) be the adjacency matrix and the diagonal matrix of the degrees of G, respectively. The signless Laplacian Q(G) of G is defined as

    Q(G)=D(G)+A(G).

    The Laplacian L(G) of G is defined as

    L(G)=D(G)A(G).

    We denote the eigenvalues of Q(G) and L(G) by ρk(G) and λk(G), respectively, where 1kn (we drop G when it is clear from the context). In particular, the largest eigenvalue of Q(G) is called the signless Laplacian spectral radius of G, denoted by ρ(G), and the largest eigenvalue of L(G) is called the Laplacian spectral radius of G, denoted by λ(G).

    In [1], Nikiforov introduced the Aα-matrix of G, which is the convex linear combination of A(G) and D(G) defined by

    Aα(G)=αD(G)+(1α)A(G),0α1.

    Clearly,

    A0(G)=A(G),  A1(G)=D(G)and2A12(G)=Q(G).

    Thus, the matrix Aα(G) is a generalization of the adjacency matrix and the signless Laplacian matrix. The largest eigenvalue of Aα(G) is called the Aα-spectral radius of G, denoted by ρα(G). Nikiforov [1] had studied the matrix Aα(G), and he has investigated many properties on Aα(G), including bounds on the k-th largest (especially, the largest, the second largest, and the smallest) eigenvalue of Aα(G), the positive semidefiniteness of Aα(G), etc. From then on, the study of Aα-spectra and the Aα-spectral radius for the graph has attracted the attention of many researchers. In [2], Nikiforov et al. gave an upper bound of the spectral radius of Aα(TΔ), where TΔ is the tree of maximal degree Δ. Also, he obtained several bounds on the spectral radius of Aα of general graphs. The graphs with

    λk(Aα(G))=αn1,2kn

    had been characterized by Lin et al. in [3]. In [4], Guo and Zhou gave upper bounds for the Aα-spectral radius for some graphs under some conditions. Guo and Zhang [5]; obtained a sharp upper bound on the Aα-spectral radius for α[12,1), and proved that for two connected graphs G1 and G2, ρα(G1)>ρα(G2) under certain conditions. For further related studies, one may see [6,7,8].

    The research on L(G) has shown that it is a remarkable matrix in many respects. Since; L(G) is just the difference between D(G) and A(G), then to understand to what extent each of the summands A(G) and D(G) determines the properties of L(G), and motivated by the above works, we introduce the Aα-matrix defined by

    Aα(G)=αD(G)+(α1)A(G),0α1. (1.1)

    Through our study, several facts indicate that the study of the Aα(G) family is of a unique importance. First of all, obviously

    A(G)=A0(G),D(G)=A1(G)andL(G)=2A12(G).

    Since A12(G) is essentially equivalent to L(G), then the matrix Aα(G) can be regarded as a generalization of the Laplacian matrix L(G). Since

    A1(G)=A1(G)=D(G);

    unless otherwise specified, we only consider the case of 0α<1 in this paper.

    In this paper, we study some properties on Aα(G) and obtain some bounds of the largest eigenvalue of Aα(G). For the relation between the Q-spectrum and L-spectrum of a connected graph G, it is well known that G is bipartite if and only if Q(G) and L(G) has the same spectrum ([9, Proposition 1.3.10]), and we extend this relation on the Aα- and Aα-spectrum. In [10], Akbari et al. proved that, for a graph G of order n and any real number β, if 0<β1 or 2β3, then

    nk=1ρβknk=1λβk;

    and if 1β2, then

    nk=1ρβknk=1λβk.

    In this work we present the relation between the sum of powers of Aα and Aα eigenvalues, which is a generalization of the relation between the sum of powers of Q(G) and L(G) eigenvalues in [10].

    The rest of this paper is organized as follows. In Section 2, we give some preliminaries and lemmas used later. In Section 3, we derive new basic properties of Aα(G). In Section 4, we present the Aα-spectra of the complete graphs and the complete split graphs. Section 5 gives some bounds of the largest eigenvalue of Aα(G). In Section 6, we deduce the relation between the Aα- and Aα-spectral radius. Finally, in Section 7, we prove the relation between the sum of powers of Aα and Aα eigenvalues.

    Although Weyl's inequalities have been known for almost a century (see, e.g., [11]), their equality case was first established by So in [12], and for convenience we state Weyl and So's complete theorem as follows:

    Theorem 2.1. Let A and B be Hermitian matrices of order n, and let 1in and 1jn. Then,

    λi(A)+λj(B)λi+jn(A+B),ifi+jn+1, (2.1)
    λi(A)+λj(B)λi+j1(A+B),ifi+jn+1. (2.2)

    In either of these inequalities, the equality holds if and only if there exists a nonzero n-vector that is an eigenvector to each of the three eigenvalues involved.

    A simplified version of (2.1) and (2.2) gives

    λk(A)+λmin(B)λk(A+B)λk(A)+λ(B). (2.3)

    We shall need the following lemmas for our new results.

    Lemma 2.2. [3] Let G be a graph of order n. If eE(G) and α12, then,

    ρi(Aα(G))ρi(Aα(Ge)),for1in.

    Lemma 2.3. [13] Let B be a real symmetric nonnegative irreducible matrix and λ be the largest eigenvalue of B. ZRn. If ZtBZ=λ and Z=1, then BZ=λZ.

    Lemma 2.4. [5] Let G be a connected graph with n4 vertices and m edges. If α[12,1), then,

    ρα(G)max{αΔ(G),(1α)(mn12)}+2α.

    Equality holds if and only if α=n1n+1 and G=Kn.

    For a graph G of order n, suppose that λ is an eigenvalue of Aα(G) and x is an eigenvector of Aα(G) with respect to λ. We use x(v) to denote the entry of x corresponding to the vertex vV(G). It is clear that the system of eigenequations for the matrix Aα(G) is

    λx(v)=αdG(v)x(v)+(α1)uN(v)x(u). (3.1)

    If G is a graph of order n with Aα(G)=Aα, and x is a real vector, the quadratic form Aαx,x can be represented in several equivalent ways:

    Aαx,x=uvE(G)(αx(u)2+2(α1)x(u)x(v)+αx(v)2), (3.2)
    Aαx,x=(2α1)uV(G)x(u)2dG(u)+(1α)uvE(G)(x(u)x(v))2, (3.3)
    Aαx,x=αuV(G)x(u)2dG(u)+2(α1)uvE(G)x(u)x(v). (3.4)

    Each of these representations can be useful in proofs.

    Now, we give some of the spectral properties of the Aα-matrix. Let us call the largest eigenvalue of Aα(G) the Aα-spectral radius of G, and denote it as λα(G). Let us also denote the smallest eigenvalue of Aα(G) as μα(G). Since Aα(G) is a real symmetric matrix, and by using Rayleigh's principle, the following result holds:

    Proposition 3.1. If α[0,1) and G is a graph of order n, then

    λα(G)=maxx2=1Aα(G)x,xandμα(G)=minx2=1Aα(G)x,x. (3.5)

    Moreover, if x is a unit n-vector, then,

    λα(G)=Aα(G)x,x

    if and only if x is an eigenvector to λα(G), and

    μα(G)=Aα(G)x,x

    if and only if x is an eigenvector to μα(G).

    By using these relations, the following result is evident:

    Proposition 3.2. If α[0,1) and G is a graph, then

    λα(G)=max{λα(H):HisacomponentofG},
    μα(G)=min{μα(H):HisacomponentofG}.

    It is clear that if G is a d-regular graph of order n, then

    Aα(G)=αdIn+(α1)A(G),

    and so there is a linear correspondence between the spectra of Aα(G) and of A(G),

    λk(Aα(G))=αd+(α1)λnk+1(A(G)), 1kn. (3.6)

    In particular, if k=n, then

    μα(G)=(2α1)d

    for any α[0,1]. As a consequence of Weyl's inequality (2.3), the following result is immediate:

    Proposition 3.3. If α[0,1] and G is a graph with

    A(G)=AandAα(G)=Aα,

    then,

    αδ+(α1)λnk+1(A)λk(Aα)αΔ+(α1)λnk+1(A).

    An important property of the Laplacian L(G) is that it is positive semidefinite. This is certainly not true for Aα(G) if α<12 and G is a regular graph, but if α12, then Aα(G) is like L(G). We give this result in the following:

    Proposition 3.4. If α12, and G is a graph, then Aα(G) is positive semidefinite, and if α>12 and G has no isolated vertices, then Aα(G) is positive definite. Moreover, if α<12 and G is a regular graph, then Aα(G) is not positive semidefinite.

    Proof. Let x be a nonzero vector. If α12, then for any edge uvE(G), we see that

    Aα(G)x,x(2α1)(x(u)2+x(v)2)+(1α)(x(u)x(v))20. (3.7)

    Hence, Aα(G) is positive semidefinite. Now, assume that G has no isolated vertices. Choose a vertex u with x(u)0 and let uvE(G). Then, (3.7) becomes a strict inequality, and so Aα(G) is positive definite. Finally, suppose that G is a dregular graph, α<12, and let A be its adjacency matrix, then λ(A)=d. Thus

    μα(G)=(2α1)d<0.

    By the above proposition we get the following lemma which gives a relation between the Aα eigenvalues of G and the Aα eigenvalues of spanning subgraphs of G.

    Lemma 3.5. Let G be a graph of order n and let α[12,1). If G=Ge, where eE(G), then,

    λi(G)λi(G)for1in.

    Proof. Let e=uv such that u,vV(G). It is easy to see that

    Aα(G)=Aα(G)+M,

    where M is the matrix of order n indexed by the vertices of G having (u,v)th and (v,u)th entries both equal to α1, and the (u,u)th and (v,v)th entries both equal to α and all other entries equal to zero, hence M is an Aα-matrix of a graph containing only one edge. Since α[12,1), then Aα(G), Aα(G) and M are positive semidefinite and Weyl's inequalities (2.3) imply that λi(G)λi(G).

    Equality (3.6) and the fact the eigenvalues of A(Kn) are {n1,1,,1} give the spectrum of Aα(Kn) as follows:

    Proposition 4.1. The eigenvalues of Aα(Kn) are

    μα(Kn)=(2α1)(n1)andλk(Aα(Kn))=α(n2)+1for1kn1.

    If SV(G), then we use G[S] to denote the subgraph of G induced by S. Recall that G[S] is an independent set if no two vertices of S are adjacent and G[S] is a clique if it is a complete subgraph of G. The graph Kr(nr)K1 is called a complete split graph, denoted by CSr,nr. The work in the following proposition is motivated by the proof of [3, Proposition 2.4].

    Proposition 4.2. Let G be a graph with Aα(G), and α[0,1). Let SV(G) and |S|=k. Suppose that dG(u)=d for each vertex uS, and N(v){w}=N(w){v} for any two vertices v,wS. Then, we have the following statements:

    (i) If G[S] is a clique, then α(d1)+1 is an eigenvalue of Aα(G) with multiplicity at least k1.

    (ii) If G[S] is an independent set, then αd is an eigenvalue of Aα(G) with multiplicity at least k1.

    Proof. Let S={v1,v2,,vk}. Clearly, d1==dk=d. Let z1,z2,,zk1 be vectors such that

    {zi(v1)=1,zi(vi+1)=1,zi(v)=0,if vV(G){v1,vi+1},

    for i=1,,k1. Assume that G[S] is a clique. It is easy to obtain that

    Aα(G)zi=(α(d11)+1,0,,0,α(1di)1,0,,0)=(α(d1)+1)zi,

    for i=1,,k1. Hence, α(d1)+1 is an eigenvalue of Aα(G) and z1,z2,,zk1 are eigenvectors of Aα(G) corresponding to α(d1)+1. In addition, since z1,z2,,zk1 are linearly independent, the multiplicity of α(d1)+1 is at least k1. Now, supposing that G[S] is an independent set, we have

    Aα(G)zi=(αd1,0,,0,αdi,0,,0)=αdzi,

    for i=1,,k1. Since z1,z2,,zk1 are linearly independent, it follows that αd is an eigenvalue of Aα(G) with multiplicity at least k1. Thus, the proof is completed.

    Consider an n×n real symmetric matrix

    S=(S11S12S1tS21S22S2tSt1St2Stt),

    whose rows and columns are partitioned according to a partitioning P1, P2, , Pt of {1,2,,n}. The quotient matrix B of the matrix S is the t×t matrix whose entries are the average row sums of the blocks Sij of S. The partition is equitable if each block Sij of S has constant row sum.

    Lemma 4.3. [14] Let B be an equitable quotient matrix of a symmetric real matrix S. If λ is an eigenvalue of B, then λ is also an eigenvalue of S.

    Now, we can determine all Aα-eigenvalues of CSr,nr as follows:

    Proposition 4.4. The Aα-spectrum of CSr,nr contains α(n2)+1 with multiplicity r1, αr with multiplicity nr1, and the remaining two Aα-eigenvalues are

    α(n+2(r1))+1±(α(n+2(r1))+1)2+4r(12α)(nr+α(r1))2. (4.1)

    Proof. We can write Aα(CSr,nr) as

    Aα(CSr,nr)=((1α)Jr,r+(αn1)Ir(1α)Jr,nr(1α)Jnr,rαrInr,nr).

    Then, the quotient matrix of Aα(CSr,nr) is equitable and it can be written in the form

    B(Aα(CSr,nr))=((nr)α+r1(1α)(nr)(1α)rαr).

    Thus, by Lemma 4.3, the eigenvalues of B(Aα(CSr,nr)) are eigenvalues of Aα(CSr,nr), and according to Proposition 4.2, we get the result.

    In this section we give a few general bounds on λα(G).

    Proposition 5.1. Let G be a graph, with

    Δ(G)=Δ,  A(G)=AandD(G)=D.

    The following inequalities hold for λα(G):

    λα(G)αΔ+(α1)λ(A), (5.1)
    λα(G)(2α1)λ(A). (5.2)

    Proof. Inequality (5.1) follows by Weyl's inequalities (2.3) because

    αΔ+(α1)λ(A)=αλ(D)+(α1)λ(A)=λ(αD)+λmin((α1)A)λ(αD+(α1)A)=λα(G).

    To prove the inequality (5.2), let H be a component of G such that λ(A)=λ(A(H)). Let x be a positive unit vector to λ(A(H)). For every edge uvE(H), the AM-GM inequality implies that

    2x(u)x(v)=2αx(u)x(v)+2(1α)x(u)x(v)α(x(u)+x(v))22+2(1α)x(u)x(v)=12(αx(u)2+2(α1)x(u)x(v)+αx(v)2)+(32α)x(u)x(v).

    Summing this inequality over all edges uvE(H), and using (3.2), we get

    λ(A)=λ(A(H))=A(H)x,x12Aα(H)x,x+12(32α)λ(A);

    and then we get

    2λ(A)(32α)λ(A)Aα(H)x,xλα(G);

    hence

    (2α1)λ(A)λα(G).

    Having inequality (5.2) in hand, if α12, then every lower bound of λ(A) gives a lower bound on λα(G), but if α<12, then every upper bound of λ(A) gives a lower bound on λα(G). We mention just two such bounds.

    Corollary 5.2. Let G be a graph such that α12. If G is of order n and has m edges, then

    λα(G)(2α1)2mn.

    Corollary 5.3. Let G be a connected graph such that α<12. If G is of order n and has m edges, then

    λα(G)(2α1)2mn+1.

    Proposition 5.4. Let G be a graph of order n. If α(12,1], then,

    λα(G)α(n2)+1.

    Moreover,

    λα(G)=α(n2)+1

    with multiplicity k1 if G has k vertices of degree n1.

    Proof. Applying Lemma 3.5 leads to

    λα(G)λα(Kn)=α(n2)+1.

    If G has k vertices of degree n1, then it follows from Proposition 4.2 that α(n2)+1 is an eigenvalue of Aα(G) with multiplicity at least k1, and since

    λα(G)α(n2)+1,

    we get

    λα(G)=α(n2)+1

    with multiplicity k1.

    Let G be a connected graph. Merris [15] pointed out λ(L(G))ρ(Q(G)), and the equality holds if G is a bipartite graph. In the next result, we generalize this result to the Aα- and Aα-spectral radius of a connected graph.

    Theorem 6.1. Let G be a graph of order n, α(0,1), λα(G)=λα and ρα(G)=ρα. We have λαρα. Moreover, if G is connected, then the equality holds if and only if G is bipartite.

    Proof. Let

    V(G)={v1,v2,,vn},  x=(x1,x2,,xn)tRn

    be an arbitrary vector such that x=1. Let

    y=(y1,y2,,yn)tRn

    be a unit eigenvector of Aα(G) belonging to λα and

    y=(y1,y2,,yn)tRn

    be a unit eigenvector of Aα(G) belonging to ρα. Let

    |y|=(|y1|,|y2|,,|yn|)t.

    First, we prove that λαρα.

    λα=maxxtAαx=maxxt(αD+(α1)A)x=max(αni=1x2idi+2(α1)vivjE(G)xixj)=ytAαy=αni=1y2idi+2(α1)vivjE(G)yiyj

    and

    ρα=maxxtAαx=maxxt(αD+(1α)A)x=max(αni=1x2idi+2(1α)vivjE(G)xixj)=ytAαy=αni=1yi2di+2(1α)vivjE(G)yiyj.

    Thus,

    λα=αni=1y2idi+2(α1)vivjE(G)yiyjαni=1y2idi+2(1α)vivjE(G)|yiyj|=|yt|Aα|y|maxxtAαx=ρα. (6.1)

    Now, if G is bipartite, then the matrix Aα and the matrix Aα are similar by a diagonal matrix D with diagonal entries ±1, that is, Aα=DAαD1. Therefore, Aα and Aα have the same spectrum, and thus we get λα=ρα. Finally, when G is connected and λα=ρα, all inequalities (6.1) must be equalities. By Lemma 2.3 and the equality

    |yt|Aα|y|=ρα;

    we know that |y| is an eigenvector of Aα belonging to ρα. So, |y|=±y. Using the Perron-Frobenius' theorem for Aα(G), we have y>0, |y|=y, and |yi|>0, i=1,2,,n.

    Since

    αni=1y2idi+2(α1)vivjE(G)yiyj=αni=1y2idi+2(1α)vivjE(G)|yiyj|;

    we get

    vivjE(G)yiyj=vivjE(G)|yiyj|,

    hence, |yiyj|=yiyj when vivjE(G). Therefore, yiyj<0 if vivjE(G).

    Let

    U={vi:yi>0}andW={vj:yj<0}.

    For each edge e=vivj, we have yiyj<0, one of the vertices of edge e is in U, and the other is in W. So, G is a bipartite graph.

    Remark 6.2. If G is not bipartite and is not connected, and λα is the largest eigenvalue of Aα for a bipartite connected component of G, then the equality in Theorem 6.1 still holds.

    Example 6.3. Take

    G=C3C4.

    Then, G is not bipartite and not connected. Now, A14(G) has a spectrum 2[1],1.25[2],0.5[2],1[2] (where λ[i] means the eigenvalue λ is repeated i times in the spectrum). On the other hand, A14(G) has a spectrum 2[2],0.5[2],0.25[2],1[1]. Then, we have

    λ14(G)=ρ14(G)=2.

    Note that 2 is the largest eigenvalue of A14(C4) as well.

    Now, we introduce the relationship between the Aα- and Aα-spectra of bipartite graphs, which is a generalization [9, Proposition 1.3.10], and it follows from the proof of the above theorem.

    Corollary 6.4. Let G be a connected graph. Then, G is bipartite if and only if the Aα-spectrum and Aα-spectrum are equal.

    Remark 6.5. In fact, if G is bipartite and is not connected, the Aα-spectrum still equals the Aα-spectrum.

    Example 6.6. Take G=P3C4. Then G is bipartite and not connected. We have that A34(G) has a spectrum 2[1],1.64039[1],1.5[2],1[1],0.75[1],0.609612[1]. In contrast, A34(G) has the same spectrum as A34(G), although G is not connected.

    According to Corollary 6.4 and [1, Propositions 38 and 39], we get the following two results:

    Corollary 6.7. The Aα-spectrum and the Aα-spectrum of the complete bipartite graph Ka,b are equal, that is, if ab1 and α(0,1), the eigenvalues of Aα(Ka,b) are

    λα(Ka,b)=12(α(a+b)+α2(a+b)2+4ab(12α)),
    μα(Ka,b)=12(α(a+b)α2(a+b)2+4ab(12α)),
    λk(Aα(Ka,b))=αafor1<kb,
    λk(Aα(Ka,b))=αbforb<k<a+b.

    Corollary 6.8. The Aα-spectrum and the Aα-spectrum of the star K1,n1 are equal, that is, the eigenvalues of Aα(K1,n1) are

    λα(K1,n1)=12(αn+α2n2+4(n1)(12α)),
    μα(K1,n1)=12(αnα2n2+4(n1)(12α)),
    λk(Aα(K1,n1))=αfor1<k<n.

    Indeed, many practical results of Theorem 6.1 can be deduced, and here are some of them.

    Proposition 6.9. Let G be a connected graph with n4 vertices and m edges. If α[12,1), then

    λα(G)<max{αΔ(G),(1α)(mn12)}+2α. (6.2)

    Proof. Let

    λα(G)=λαandρα(G)=ρα.

    Then, Theorem 6.1 and Lemma 2.4 lead to

    λαmax{αΔ(G),(1α)(mn12)}+2α. (6.3)

    Suppose that the equality in (6.3) holds, thus ρα=λα, and so G is bipartite and

    ρα=max{αΔ(G),(1α)(mn12)}+2α.

    Therefore, by Lemma 2.4, G=Kn, and thus G is bipartite if and only if n=2, but n4, and hence the inequality is strict.

    Theorem 6.1 and [2, Theorem 2] lead directly to the next result.

    Proposition 6.10. If T is a tree of order n and α[0,1], then

    λα(T)nα+n2α2+4(n1)(12α)2.

    The equality holds if and only if T is the star K1,n1.

    By Theorem 6.1 and [1, Proposition 20], we get the next result.

    Proposition 6.11. If G is a graph with no isolated vertices, then

    λα(G)maxuV(G){αd(u)+1αd(u)uvE(G)d(v)}.

    If α(12,1) and G is connected, the equality holds if and only if G is a regular bipartite graph.

    In this part we give two explicit expressions for the sums and the sum of squares of the eigenvalues of Aα and Aα, considering [1, Propositions 34 and 35].

    Proposition 6.12. If G is a graph of order n and has m edges, then

    ni=1λi(Aα(G))=ni=1λi(Aα(G))=trAα(G)=trAα(G);

    where

    ni=1λi(Aα(G))=αuV(G)d(u)=2αm.

    A similar formula for the sum of the squares of the Aα and Aα-eigenvalues is given as follows:

    Proposition 6.13. If G is a graph of order n and has m edges, then

    ni=1λ2i(Aα(G))=ni=1λ2i(Aα(G))=trA2α(G)=trA2α(G);

    where

    ni=1λ2i(Aα(G))=α2uV(G)d2(u)+2(1α)2m.

    Proof. Let

    Aα=Aα(G),  A=A(G)andD=D(G).

    Calculating the square A2α and taking its trace, we find that

    trA2α=tr(α2D2+(1α)2A2+α(α1)DA+α(α1)AD)=α2trD2+(1α)2trA2α(1α)trDA+α(α1)trAD=α2uV(G)d2(u)+2(1α)2m.

    Pirzada et al. [16] introduced the sum of the βth powers of the Aα eigenvalues of G as

    Sαβ(G)=ni=1ρβi.

    Now, we have the notation

    Sαβ(G)=ni=1λβi

    for the sum of the βth powers of the Aα eigenvalues of G. The following theorem is a generalization [10, Theorem 2].

    Theorem 7.1. Let G be a graph of order n>1, α[12,1), and let β be a real number.

    (i) If 0<β1 or 2β3, then

    Sαβ(G)Sαβ(G).

    (ii) If 1β2, then

    Sαβ(G)Sαβ(G).

    For β(0,1)(2,3), the equality holds in (i) if and only if G is a bipartite graph. Moreover, for β(1,2), the equality holds in (ii) if and only if G is a bipartite graph.

    Proof. We recall that, for any real number r, the binomial series

    k=0(rk)xk

    converges to (1+x)r if |x|<1. This also remains true for x=1 if r>0 (see, e.g., [17, p. 419]). By Lemma 3.5, we find that,

    λα(G)λα(Kn)=α(n2)+1andρα(G)ρα(Kn)=n1.

    Since Aα is positive semidefinite when α[12,1), then if ρi>0, we have

    |ρin1||n1n1|=|1n|<1

    and if ρi=0, we get

    ρin1=1.

    Therefore,

    Sαβ(G)nβ=(ρ1n)β++(ρnn)β=k=0(βk)(ρ1n1)k++k=0(βk)(ρnn1)k=k=0(βk)tr(1n(αD+(1α)A)I)k.

    Also, since Aα is positive semidefinite when α[12,1), then if λi>0,

    |λin1||α(n2)+1n1|=|(α1)+12αn|<1

    and if λi=0, we have

    λin1=1.

    Thus, in a similar manner as above, we obtain that

    Sαβ(G)nβ=k=0(βk)tr(1n(αD+(α1)A)I)k.

    We claim that

    ifkis even, tr(αD+(1α)AnI)ktr(αD+(α1)AnI)k;
    ifkis odd, tr(αD+(1α)AnI)ktr(αD+(α1)AnI)k.

    When ((αDnI)+(1α)A)k and ((αDnI)+(α1)A)k are expanded in terms of the powers of αDnI and (1α)A, respectively, the terms appearing in both expansions, regardless of their signs, are the same. To prove this claim, we identify the sign of each term in both expansions. Consider the terms in the expansion of ((αDnI)+(1α)A)k, where there are exactly j factors equal to αDnI, for some j=0,1,,k. The sign of the trace for each of these terms is (1)j or 0 because all entries of αDnI and (1α)A are non-positive and non-negative, respectively. On the other hand, in each term in the expansion of ((αDnI)+(α1)A)k all factors are matrices with non-positive entries, hence the sign of the trace of each term is (1)k or 0. Therefore, the claim has been proven.

    Now, note that if 0<β<1 or 2<β<3, then the sign of (βk) is (1)k1, except that (β2)>0, for 2<β<3. According to this, for 0<β<1 and every k,

    (βk)tr(αD+(1α)AnI)k(βk)tr(αD+(α1)AnI)k.

    This inequality remains true for 2β3 as

    tr(αD+(1α)AnI)2=tr(αD+(α1)AnI)2,

    since trA2α=trA2α. Thus, Part (i) is proved. For 1<β<2, the sign of (βk) is (1)k1, except that (β1)>0. Since trAα=trAα, we have

    tr(αD+(1α)AnI)=tr(αD+(α1)AnI),

    and so part (ii) is similarly proved.

    Now, we examine the equality case. Since Aα and Aα are similar if G is bipartite, it follows that the equality holds in both (i) and (ii). Since for any positive integer i, trAi equals the total number of closed walks of length i in G, then if G is not bipartite, there exists an odd integer r such that trAr>0 (see [18, Lemma 2.5]). Hence,

    tr(αD+(1α)AnI)r>tr(αD+(α1)AnI)r;

    and so the inequalities in both (i) and (ii) are strict.

    We know that the Aα-spectra and Aα-spectra of Kn are {(n1)[1],(αn1)[n1]} and {((2α1)(n1))[1],(α(n2)+1)[n1]}, respectively. Therefore,

    Sαβ(Kn)Sαβ(Kn)=(n1)β+(n1)(αn1)β((2α1)(n1))β(n1)(α(n2)+1)β.

    By Theorem 7.1 and based on our numerical experiments, we propose the following conjecture:

    Conjecture 7.2. For every α(12,1) and each integer n3, we have

    Sαβ(Kn)Sαβ(Kn)0

    for any β[0,1][2,) and

    Sαβ(Kn)Sαβ(Kn)0

    for any β(,0)(1,2).

    In this paper, we have introduced the Aα-matrix of a graph G, which is a generalization of the Laplacian matrix L(G), and we have studied the basic properties of Aα; and derived some bounds for its spectral radius. Furthermore, we have determined the Aα-spectra for the complete graph and the complete split graph. Building upon previous results, we have extended findings related to the spectral radius of L(G) and Q(G) matrices to the Aα- and Aα-spectral radius. Specifically, in Theorem 6.1, we have generalized Merris' [15] observation that λ(L(G))ρ(Q(G)) with equality holding for bipartite graphs. Additionally, we have extended the known relation that G is bipartite if and only if Q(G) and L(G) share the same spectrum, as demonstrated in Corollary 6.4. Finally, in Theorem 7.1; we have generalized a relation established by S. Akbari et al. in [10]; which relates the sum of powers of the eigenvalues of Q(G) and L(G). In conclusion, these findings have implications for various applications involving graph analysis.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    The authors are grateful to the referees for their valuable suggestions and comments.

    The authors declare no conflicts of interest.



    [1] K. Liu, S. Xu, G. Xu, M. Zhang, D. Sun, H. Liu, A review of android malware detection approaches based on machine learning, IEEE Access, 8 (2020). https://doi.org/10.1109/ACCESS.2020.3006143
    [2] B. Amos, H. Turner, J. White, Applying machine learning classifiers to dynamic Android malware detection at scale, In: 2013 9th International Wireless Communications and Mobile Computing Conference (IWCMC), IEEE, Italy, 2013, 1666–1671. https://doi.org/10.1109/IWCMC.2013.6583806
    [3] M. Egele, T. Scholte, E. Kirda, C. Kruegel, A survey on automated dynamic malware-analysis techniques and tools, ACM Comput. Surv., 44 (2012), 1–42.
    [4] B. Amro, Malware detection techniques for mobile devices, Int. J. Mobile Netw. Commun. Telemat., 7 (2017). https://doi.org/10.1145/2089125.2089126
    [5] K. Kavitha, P. Salini, V. Ilamathy, Exploring the malicious Android applications and reducing risk using static analysis, In: 2016 International Conference on Electrical, Electronics, and Optimization Techniques (ICEEOT), IEEE, India, 2016, 1316–1319. https://doi.org/10.1109/ICEEOT.2016.7754896
    [6] E. M. B. Karbab, M. Debbabi, MalDy: Portable, data-driven malware detection using natural language processing and machine learning techniques on behavioral analysis reports, Digit. Invest., 28 (2019), 77–87. https://doi.org/10.1016/j.diin.2019.01.017 doi: 10.1016/j.diin.2019.01.017
    [7] R. Ito, M. Mimura, Detecting unknown malware from ASCII strings with natural language processing techniques, In: 2019 14th Asia Joint Conference on Information Security (AsiaJCIS), IEEE, Japan, 2019. https://doi.org/10.1109/AsiaJCIS.2019.00-12
    [8] P. Najafi, D. Koehler, F. Cheng, C. Meinel, NLP-based entity behavior analytics for malware detection, In: 2021 IEEE International Performance, Computing, and Communications Conference (IPCCC), IEEE, USA, 2021. https://doi.org/10.1109/IPCCC51483.2021.9679411
    [9] U. Raghav, E. Martinez-Marroquin, W. Ma, Static analysis for Android Malware detection with document vectors, In: 2021 International Conference on Data Mining Workshops (ICDMW), IEEE, New Zealand, 2021. https://doi.org/10.1109/ICDMW53433.2021.00104
    [10] X. Xing, X. Jin, H. Elahi, H. Jiang, G. Wang, A malware detection approach using autoencoder in deep learning, IEEE Access, 10 (2022), 25696–25706. https://doi.org/10.1109/ACCESS.2022.3155695 doi: 10.1109/ACCESS.2022.3155695
    [11] Q. Le, O. Boydell, B. Mac, M. Scanlon, Deep learning at the shallow end: Malware classification for non-domain experts, Digit. Invest., 26 (2018), S118–S126. http://dx.doi.org/10.1016/j.diin.2018.04.024 doi: 10.1016/j.diin.2018.04.024
    [12] J. Y. Kim, S. J. Bu, S. B. Cho, Zeroday malware detection using transferred generative adversarial networks based on deep autoencoders, Inform. Sci., 460–461 (2018), 83–102. https://doi.org/10.1016/j.ins.2018.04.092 doi: 10.1016/j.ins.2018.04.092
    [13] I. Goodfellow, NIPS 2016 Tutorial: Generative adversarial networks, arXiv preprint, 2014. https://doi.org/10.48550/arXiv.1701.00160
    [14] S. Kumar, B. Janet, DTMIC: Deep transfer learning for malware image classification, J. Inf. Secur. Appl., 64 (2022). https://doi.org/10.1016/j.jisa.2021.103063
    [15] Ö. Aslan, A. A. Yilmaz, A new malware classification framework based on deep learning algorithms, IEEE Access, 9 (2021), 87936–87951. https://doi.org/10.1109/ACCESS.2021.3089586 doi: 10.1109/ACCESS.2021.3089586
    [16] F. Rustam, I. Ashraf, A. D. Jurcut, A. K. Bashir, Y. B. Zikria, Malware detection using image representation of malware data and transfer learning, J. Parallel Distr. Com., 172 (2023), 32–50. https://doi.org/10.1016/j.jpdc.2022.10.001 doi: 10.1016/j.jpdc.2022.10.001
    [17] T. Li, Y. Luo, X. Wan, Q. Li, Q. Liu, R. Wang, et al., A malware detection model based on imbalanced heterogeneous graph embeddings, Expert Syst. Appl., 246 (2014), 123109.
    [18] Google play store. Available from: https://https://play.google.com/store/apps.
    [19] Virusshare. Available from: http://virusshare.com/.
    [20] Virustotal. Available from: https://www.virustotal.com/gui/home/upload.
    [21] L. Nataraj, S. Karthikeyan, G. Jacob, B. S. Manjunath, Malware images: Visualization and automatic classification, In: Proceedings of the 8th International Symposium on Visualization for Cyber Security, 2011, 1–7. https://doi.org/10.1145/2016904.2016908
    [22] A. S. Bozkir, A. O. Cankaya, M. Aydos, Utilization and comparison of convolutional neural networks in malware recognition, In: 2019 27th Signal Processing and Communications Applications Conference (SIU), IEEE, Turkey, 2019, 1–4. https://doi.org/10.1109/SIU.2019.8806511
    [23] MaleVis. Available from: https://web.cs.hacettepe.edu.tr/selman/malevis/.
    [24] S. Venkatraman, M. Alazab, R. Vinayakumar, A hybrid deep learning image-based analysis for effective malware detection, J. Inf. Secur. Appl., 47 (2019), 377–389.
    [25] A. Krizhevsky, I. Sutskever, G. E. Hinton, ImageNet classification with deep convolutional neural networks, Adv. Neural Inform. Proc. Syst., 2012.
    [26] K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition, arXiv preprint, 2014. https://doi.org/10.48550/arXiv.1409.1556
    [27] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016,770–778.
    [28] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, Z. Wojna, Rethinking the inception architecture for computer vision, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, 2818–2826.
    [29] G. Huang, Z. Liu, L. Van Der Maaten, K. Q. Weinberger, Densely connected convolutional networks, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, 4700–4708.
    [30] J. Yosinski, J. Clune, Y. Bengio, H. Lipson, How transferable are features in deep neural networks? In: Advances in Neural Information Processing Systems (NIPS), 2014.
    [31] S. J. Pan, Q. Yang, A survey on transfer learning, IEEE T. Knowl. Data Eng., 22 (2010), 1345–1359. https://doi.org/10.1109/TKDE.2009.191 doi: 10.1109/TKDE.2009.191
    [32] R. Vinayakumar, M. Alazab, K. P. Soman, P. Poornachandran, S. Venkatraman, Robust intelligent malware detection using deep learning, IEEE Access, 7 (2019), 46717–46738. https://doi.org/10.1109/ACCESS.2019.2906934 doi: 10.1109/ACCESS.2019.2906934
    [33] J. S. Luo, D. C. T. Lo, Binary malware image classification using machine learning with local binary pattern, In: 2017 IEEE International Conference on Big Data (Big Data), IEEE, USA, 2017, 4664–4667. https://doi.org/10.1109/BigData.2017.8258512
    [34] Z. Cui, F. Xue, X. Cai, Y. Cao, G. G. Wang, J. Chen, Detection of malicious code variants based on deep learning, IEEE T. Ind. Inform., 14 (2018), 3187–3196. https://doi.org/10.1109/tii.2018.2822680 doi: 10.1109/tii.2018.2822680
    [35] D. Gibert, Convolutional neural networks for malware classification, M.S. thesis, Univ. Rovira Virgili, Tarragona, Spain, 2016.
    [36] A. Singh, A. Handa, N. Kumar, S. K. Shukla, Malware classification using image representation, In: Proc. Int. Symp. Cyber Secur. Cryptogr. Mach. Learn. Cham, Switzerland: Springer, 2019, 75–92. https://doi.org/10.1007/978-3-030-20951-3_6
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1359) PDF downloads(89) Cited by(0)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog