Research article Special Issues

Evolving blocks by segmentation for neural architecture search

  • Received: 18 October 2023 Revised: 23 February 2024 Accepted: 27 February 2024 Published: 06 March 2024
  • Convolutional neural networks (CNNs) play a prominent role in solving problems in various domains such as pattern recognition, image tasks, and natural language processing. In recent years, neural architecture search (NAS), which is the automatic design of neural network architectures as an optimization algorithm, has become a popular method to design CNN architectures against some requirements associated with the network function. However, many NAS algorithms are characterised by a complex search space which can negatively affect the efficiency of the search process. In other words, the representation of the neural network architecture and thus the encoding of the resulting search space plays a fundamental role in the designed CNN performance. In this paper, to make the search process more effective, we propose a novel compact representation of the search space by identifying network blocks as elementary units. The study in this paper focuses on a popular CNN called DenseNet. To perform the NAS, we use an ad-hoc implementation of the particle swarm optimization indicated as PSO-CNN. In addition, to reduce size of the final model, we propose a segmentation method to cut the blocks. We also transfer the final model to different datasets, thus demonstrating that our proposed algorithm has good transferable performance. The proposed PSO-CNN is compared with 11 state-of-the-art algorithms on CIFAR10 and CIFAR100. Numerical results show the competitiveness of our proposed algorithm in the aspect of accuracy and the number of parameters.

    Citation: Xiaoping Zhao, Liwen Jiang, Adam Slowik, Zhenman Zhang, Yu Xue. Evolving blocks by segmentation for neural architecture search[J]. Electronic Research Archive, 2024, 32(3): 2016-2032. doi: 10.3934/era.2024092

    Related Papers:

    [1] Nina Huo, Bing Li, Yongkun Li . Global exponential stability and existence of almost periodic solutions in distribution for Clifford-valued stochastic high-order Hopfield neural networks with time-varying delays. AIMS Mathematics, 2022, 7(3): 3653-3679. doi: 10.3934/math.2022202
    [2] Yongkun Li, Xiaoli Huang, Xiaohui Wang . Weyl almost periodic solutions for quaternion-valued shunting inhibitory cellular neural networks with time-varying delays. AIMS Mathematics, 2022, 7(4): 4861-4886. doi: 10.3934/math.2022271
    [3] Ardak Kashkynbayev, Moldir Koptileuova, Alfarabi Issakhanov, Jinde Cao . Almost periodic solutions of fuzzy shunting inhibitory CNNs with delays. AIMS Mathematics, 2022, 7(7): 11813-11828. doi: 10.3934/math.2022659
    [4] Yuwei Cao, Bing Li . Existence and global exponential stability of compact almost automorphic solutions for Clifford-valued high-order Hopfield neutral neural networks with D operator. AIMS Mathematics, 2022, 7(4): 6182-6203. doi: 10.3934/math.2022344
    [5] Jin Gao, Lihua Dai . Weighted pseudo almost periodic solutions of octonion-valued neural networks with mixed time-varying delays and leakage delays. AIMS Mathematics, 2023, 8(6): 14867-14893. doi: 10.3934/math.2023760
    [6] Hedi Yang . Weighted pseudo almost periodicity on neutral type CNNs involving multi-proportional delays and D operator. AIMS Mathematics, 2021, 6(2): 1865-1879. doi: 10.3934/math.2021113
    [7] Yanshou Dong, Junfang Zhao, Xu Miao, Ming Kang . Piecewise pseudo almost periodic solutions of interval general BAM neural networks with mixed time-varying delays and impulsive perturbations. AIMS Mathematics, 2023, 8(9): 21828-21855. doi: 10.3934/math.20231113
    [8] Zhangir Nuriyev, Alfarabi Issakhanov, Jürgen Kurths, Ardak Kashkynbayev . Finite-time synchronization for fuzzy shunting inhibitory cellular neural networks. AIMS Mathematics, 2024, 9(5): 12751-12777. doi: 10.3934/math.2024623
    [9] Abdulaziz M. Alanazi, R. Sriraman, R. Gurusamy, S. Athithan, P. Vignesh, Zaid Bassfar, Adel R. Alharbi, Amer Aljaedi . System decomposition method-based global stability criteria for T-S fuzzy Clifford-valued delayed neural networks with impulses and leakage term. AIMS Mathematics, 2023, 8(7): 15166-15188. doi: 10.3934/math.2023774
    [10] Xiaofang Meng, Yongkun Li . Pseudo almost periodic solutions for quaternion-valued high-order Hopfield neural networks with time-varying delays and leakage delays on time scales. AIMS Mathematics, 2021, 6(9): 10070-10091. doi: 10.3934/math.2021585
  • Convolutional neural networks (CNNs) play a prominent role in solving problems in various domains such as pattern recognition, image tasks, and natural language processing. In recent years, neural architecture search (NAS), which is the automatic design of neural network architectures as an optimization algorithm, has become a popular method to design CNN architectures against some requirements associated with the network function. However, many NAS algorithms are characterised by a complex search space which can negatively affect the efficiency of the search process. In other words, the representation of the neural network architecture and thus the encoding of the resulting search space plays a fundamental role in the designed CNN performance. In this paper, to make the search process more effective, we propose a novel compact representation of the search space by identifying network blocks as elementary units. The study in this paper focuses on a popular CNN called DenseNet. To perform the NAS, we use an ad-hoc implementation of the particle swarm optimization indicated as PSO-CNN. In addition, to reduce size of the final model, we propose a segmentation method to cut the blocks. We also transfer the final model to different datasets, thus demonstrating that our proposed algorithm has good transferable performance. The proposed PSO-CNN is compared with 11 state-of-the-art algorithms on CIFAR10 and CIFAR100. Numerical results show the competitiveness of our proposed algorithm in the aspect of accuracy and the number of parameters.



    As stated in [1], a nervous system in the real world, synaptic transmission is a noisy process caused by random fluctuations in neurotransmitter release and other probabilistic factors. Therefore, it is necessary to consider stochastic neural networks (NNs) because random inputs may change the dynamics of the (NN) [2,3,4,5].

    SICNNs, which were proposed in [6], have attracted the interest of many scholars since their introduction due to their special roles in psychophysics, robotics, adaptive pattern recognition, vision, and image processing. In the above applications, their dynamics play an important role. Thereupon, their various dynamics have been extensively studied (see [7,8,9,10,11,12,13] and references therein). However, there is limited research on the dynamics of stochastic SICNNs. Therefore, it is necessary to further study the dynamics of such NNs.

    On the one hand, research on the dynamics of NNs that take values from a non commutative algebra, such as quaternion-valued NNs [14,15,16], octonion-valued NNs [17,18,19,20], and Clifford-valued NNs [21,22,23], has gained the interest of many researchers because such neural networks can include typical real-valued NNs as their special cases, and they have superior multi-dimensional signal processing and data storage capabilities compared to real-valued NNs. It is worth mentioning that in recent years, many authors have conducted extensive research on various dynamics of Clifford-valued NNs, such as the existence, multiplicity and stability of equilibrium points, and the existence, multiplicity and stability of almost periodic solutions as well as the synchronization problems [22,23,24,25,26,27,28,29,30]. However, most of the existing results for the dynamics of Clifford-valued NNs has been obtained through decomposition methods [24,25,26,27]. However, the results obtained by decomposition methods are generally not convenient for direct application, and there is little research on Clifford-valued NNs using non decomposition methods [28,29,30]. Therefore, further exploration of using non decomposition methods to study the dynamics of Clifford-valued NNs has important theoretical significance and application value.

    On the other hand, Bohr's almost periodicity is a special case of Stepanov's almost periodicity, but there is little research on the Stepanov periodic oscillations of NNs [19,31,32,33], especially the results of Stepanov's almost periodic solutions of stochastic SICNNs with discrete and infinitely distributed delays have not been published yet.

    Motivated by the discussion above, our purpose of this article is to establish the existence and global exponential stability of Stepanov almost periodic solutions in the distribution sense for a stochastic Clifford-valued SICNN with mixed delays via non decomposition methods.

    The subsequent sections of this article are organized as follows. Section 2 introduces some concepts, notations, and basic lemmas and gives a model description. Section 3 discusses the existence and stability of Stepanov almost periodic solutions in the distribution sense of the NN under consideration. An example is provided in Section 4. Finally, Section 5 provides a brief conclusion.

    Let A={ϑPxϑeϑ,xϑR} be a real Clifford-algebra with N generators e=e0=1, and eh,h=1,2,,N, where P={,0,1,2,,ϑ,,12N}, e2i=1,i=1,2,,r,e2i=1,i=r+1,r+2,,m,eiej+ejei=0,ij and i,j=1,2,,N. For x=ϑPxϑeϑA, we indicate x=maxϑP{|xϑ|},xc=ϑxϑeϑ,x0=xxc, and for x=(x11,x12,,x1n,x21,x22,,x2n,,xmn)TAm×n, we denote x0=max{xij,1im,1jn}. The derivative of x(t)=ϑPxϑ(t)eϑ is defined by ˙x(t)=ϑP˙xϑ(t)eϑ and the integral of x(t)=ϑPxϑ(t)eϑ over the interval [a,b] is defined by bax(t)dt=ϑP(baxϑ(t)dt)eϑ.

    Let (Y,ρ) be a separable metric space and P(Y) the collection of all probability measures defined on Borel σ-algebra of Y. Denote by Cb(Y) the set of continuous functions f:YR with g:=supxY{|g(x)|}<.

    For gCb(Y), μ,νP(Y), let us define

    gL=supxy|g(x)g(y)|ρ(x,y),gBL=max{g,gL},
    ρBL(μ,ν):=supgBL1|Ygd(μν)|.

    According to [34], (Y,ρBL(,)) is a Polish space.

    Definition 2.1. [35] A continuous function g:RY is called almost periodic if for every ε>0, there is an (ε)>0 such that each interval with length has a point τ meeting

    ρ(g(t+τ),g(t))<ε,foralltR.

    We indicate by AP(R,Y) the set of all such functions.

    Let (X,) signify a separable Banach space. Denote by μ(X):=PX1 and E(X) the distribution and the expectation of X:(Ω,F,P)X, respectively.

    Let Lp(Ω,X) indicate the family of all X-valued random variables satisfying E(Xp)=ΩXpdP<.

    Definition 2.2. [21] A process Z:RLp(Ω,X) is called Lp-continuous if for any t0R,

    limtt0EZ(t)Z(t0)p=0.

    It is Lp-bounded if suptREZ(t)p<.

    For 1<p<, we denote by Lploc(R,X) the space of all functions from R to X which are locally p-integrable. For gLploc(R,X), we consider the following Stepanov norm:

    gSp=suptR(t+1tg(s)pds)1p.

    Definition 2.3. [35] A function gLploc(R,X) is called p-th Stepanov almost periodic if for any ε>0, it is possible to find a number >0 such that every interval with length has a number τ such that

    g(t+τ)g(t)Sp<ε.

    Definition 2.4. [9] A stochastic process ZLploc(R,Lp(Ω,X)) is said to be Sp-bounded if

    ZSps:=suptR(t+1tEZ(s)pds)1p<.

    Definition 2.5. [9] A stochastic process ZLloc(R,Lp(Ω,H)) is called Stepanov almost periodic in p-th mean if for any ε>0, it is possible to find a number >0 such that every interval with length has a number τ such that

    Z(t+τ)Z(t)Sps<ε.

    Definition 2.6. [9] A stochastic process Z:RLp(Ω,X)) is said to be p-th Stepanov almost periodic in the distribution sense if for each ε>0, it is possible to find a number >0 such that any interval with length has a number τ such that

    supaR(a+1adpBL(P[Z(t+τ)]1,P[Z(t)]1)dt)1p<ε.

    Lemma 2.1. [36] (Burkholder-Davis-Gundy inequality) If fL2(J,R), p>2, B(t) is Brownian motion, then

    E[suptJ|tt0f(s)dB(s)|p]CpE[Tt0|f(s)|2ds]p2,

    where cp=(pp+12(p1)p1)p2.

    The model that we consider in this paper is the following stochastic Clifford-valued SICNN with mixed delays:

    dxij(t)=[aij(t)xij(t)+CklNh1(i,j)Cklij(t)f(xkl(tτkl(t)))xij(t)+CklNh2(i,j)Bklij(t)0Kij(u)g(xkl(tu))duxij(t)+Lij(t)]dt+CklNh3(i,j)Eklij(t)δij(xij(tσij(t)))dωij(t), (2.1)

    where i=1,2,,m,j=1,2,,n, Cij(t) represents the cell at the (i,j) position, the h1-neighborhood Nh1(i,j) of Cij is given as:

    Nh1(i,j)={Ckl:max(|ki|,|lj|)h1,1km,1ln},

    Nh2(i,j),Nh3(i,j) are similarly defined, xij denotes the activity of the cell Cij, Lij(t):RA corresponds to the external input to Cij, the function aij(t):RA represents the decay rate of the cell activity, Cklij(t):RA,Bklij(t):RA and Eklij(t):RA signify the connection or coupling strength of postsynaptic activity of the cell transmitted to the cell Cij, and the activity functions f():AA, and g():AA are continuous functions representing the output or firing rate of the cell Ckl, and τkl(t),σij(t):RR+ are the transmission delay, the kernel Kij(t):RR is an integrable function, ωij(t) represents the Brownian motion defined on a complete probability space, δij():AA is a Borel measurable function.

    Let (Ω, F, {Ft}t0, P) be a complete probability space in which {Ft}t0 is a natural filtration meeting the usual conditions. Denote by BF0([θ,0],An) the family of bounded, F0-measurable and An-valued random variables from [θ,0]An. The initial values of system (2.1) are depicted as

    xi(s)=ϕi(s),s[θ,0],

    where ϕiBF0([θ,0],A),θ=max1i,jn{suptRτij(t),suptRσij(t)}.

    For convenience, we introduce the following notations:

    a_0=minijΛa_0ij=minijΛinftRa0ij(t),ˉa0=maxijΛˉa0ij=maxijΛsuptRa0ij(t),Cklij+=suptRCklij(t),¯ac=maxijΛˉacij=maxijΛsuptRacij(t),Bklij+=suptRBklij(t),Eklij+=suptREklij(t),K+ij=suptRKij(t),τ+kl=suptRτkl(t),˙τ+kl=suptR˙τkl(t),σ+ij=suptRσij(t),˙σ+ij=suptR˙σij(t),ML=maxijΛL+ij=maxijΛsuptRLij(t),θ=maxijΛ{τ+ij,σ+ij},Λ={11,12,,1n,,mn}.

    Throughout this paper, we make the following assumptions:

    (A1) For ijΛ, f,g,δijC(A,A) satisfy the Lipschitz condition, and f,g are bounded, that is, there exist constants Lf>0,Lg>0,Lδij>0,Mf>0,Mg>0 such that for all x,yA,

    ||f(x)f(y)||Lf||xy||,||g(x)g(y)||Lg||xy||,||δij(x)δij(y)||Lδij||xy||,||f(x)||Mf,||g(x)||Mg;

    furthermore, f(0)=g(0)=δij(0)=0.

    (A2) For ijΛ, a0ijAP(R,R+),acijAP(R,A),τij,σijAP(R,R+)C1(R,R) satisfying 1˙τ+ij,1˙σ+ij>0, Cklij,Bklij,EklijAP(R,A), L=(L11,L12,,Lmn)Lploc(R,Lp(Ω,Am×n)) is almost periodic in the sense of Stepanov.

    (A3) For p>2,1p+1q=1,

    0<r1:=8p4maxijΛ{(pqa_0ij)pqqpa_0ij[(ˉacij)p+(CklNh1(i,j)(Cklij+)q)pq(2κLf+Mf)p+(CklNh2(i,j)(Bklij+)q)pq((2κLg+Mg)0|Kij(u)|du)p]+Cp(p22a_0ij)p22qpa_0ij(CklNh3(i,j)(Eklij+)q)pq(Lδij)p}<1,

    and for p=2,

    0<r2:=16maxijΛ{1(a_0ij)2[(ˉacij)2+CklNh1(i,j)(Cklij+)2(2κLf+Mf)2+CklNh2(i,j)(Bklij+)2×((2κLg+Mg)0|Kij(u)|du)2]+12a_0ijCklNh3(i,j)(Eklij+)2(Lδij)2}<1.

    (A4) For 1p+1q=1,

    0<qpa_0ρ1:=16p1qpa_0maxijΛ{(pqa_0ij)pq[(ˉacij)p+(CklNh1(i,j)(Cklij+)q)pq[2p1(Lf)p×CklNh1(i,j)epqa_0ijτkl+(2κ)p1˙τ+kl+(Mf)p]+(CklNh2(i,j)(Bklij+)q)pq[(2κLg×0|Kij(u)|du)p+(Mg0|Kij(u)|du)p]]+2p1Cp(p22a_0ij)p22×(CklNh3(i,j)(Eklij+)q)pq(Lδij)pepqa_0ijσ+ij1˙σ+ij}<1,(p>2),
    0<ρ2a_0:=32a_0maxijΛ{(1a_0ij)CklNh1(i,j)(Cklij+)2[(Lf)2CklNh1(i,j)ea_0ijτkl+(2κ)21˙τ+kl+(Mf)22]+CklNh3(i,j)(Eklij+)2(Lδij)2e2a_0ijσ+ij1˙σ+ij+12a_0ijCklNh2(i,j)(Bklij+)2(4κ2L2g+M2g)×(0|Kij(u)|du)2+(ˉacij)22a_0ij}<1,(p=2).

    (A5) The kernel Kij is almost periodic and there exist constants M>0 and u>0 such that |Kij(t)|Meut for all tR.

    Let X indicate the space of all Lp-bounded and Lp-uniformly continuous stochastic processes from R to Lp(Ω,Am×n), then with the norm ϕX=suptR{Eϕ(t)p0}1p, where ϕ=(ϕ11,ϕ12,,ϕmn)X, it is a Banach space.

    Set ϕ0=(ϕ011,ϕ012,,ϕ0mn)T, where ϕ0ij(t)=tetsa0ij(u)duLij(s)ds,tR,ijΛ. Then, ϕ0 is well defined under assumption (A2). Consequently, we can take a constant κ such that κϕ0X.

    Definition 3.1. [37] An Ft-progressively measurable stochastic process x(t)=(x11(t),x12(t),,xmn(t))T is called a solution of system (2.1), if x(t) solves the following integral equation:

    xij(t)=xij(t0)ett0a0ij(u)du+tt0etsa0ij(u)du[acij(s)xij(s)+CklNh1(i,j)Cklij(s)×f(xkl(sτkl(s)))xij(s)+CklNh2(i,j)Bklij(s)0Kij(u)g(x(su))duxij(s)+Lij(s)]ds+tt0etsa0ij(u)duCklNh3(i,j)Eklij(s)δij(xij(sσij(s)))dwij(s). (3.1)

    In (3.1), let t0, then one gets

    xij(t)=tetsa0ij(u)du[acij(s)xij(s)+CklNh1(i,j)Cklij(s)f(xkl(sτkl(s)))xij(s)+CklNh2(i,j)Bklij(s)0Kij(u)g(x(su))duxij(s)+Lij(s)]ds+tetsa0ij(u)du×CklNh3(i,j)Eklij(s)δij(xij(sσij(s)))dwij(s),tt0,ijΛ. (3.2)

    It is easy to see that if x(t) solves (3.2), then it also solves (2.1).

    Theorem 3.1. Assume that (A1)(A4) hold. Then the system (2.1) has a unique Lp-bounded and Lp-uniformly continuous solution in X={ϕX:ϕϕ0Xκ}, where κ is a constant satisfying κϕ0X.

    Proof. Define an operator ϕ:XX as follows:

    (Ψϕ)(t)=((Ψ11ϕ)(t),(Ψ12ϕ)(t),,(Ψmnϕ)(t))T,

    where (ϕ11,ϕ12,,ϕmn)TX, tR and

    (Ψijϕ)(t)=tetsa0ij(u)du[acij(s)ϕij(s)+CklNh1(i,j)Cklij(s)f(ϕkl(sτkl(s)))ϕij(s)+CklNh2(i,j)Bklij(s)0Kij(u)g(ϕkl(su))duϕij(s)+Lij(s)]ds+tetsa0ij(u)duCklNh3(i,j)Eklij(s)δij(ϕij(sσij(s)))dωij(s),ijΛ. (3.3)

    First of all, let us show that EΨϕ(t)ϕ0(t)p0κ for all ϕX.

    Noticing that for any ϕX, it holds

    ϕXϕ0X+ϕϕ0X2κ.

    Then, we deduce that

    EΨϕ(t)ϕ0(t)p04p1maxijΛ{Etetsa0ij(u)duacij(s)ϕij(s)p}+4p1maxijΛ{Etetsa0ij(u)du×CklNh1(i,j)Cklij(s)f(ϕkl(sτkl(s)))ϕij(s)dsp}+4p1maxijΛ{Etetsa0ij(u)du×CklNh2(i,j)Bklij(s)0Kij(u)g(ϕkl(su))duϕij(s)dsp}+4p1maxijΛ{Etetsa0ij(u)duCklNh3(i,j)Eklij(s)δij(ϕij(sσij(s)))dωij(s)p}:=F1+F2+F3+F4. (3.4)

    By the Hölder inequality, we have

    F24p1maxijΛ{E[teqptsa0ij(u)duds]pq[tepqtsa0ij(u)du×(CklNh1(i,j)Cklij(s)f(ϕkl(sτkl(s)))ϕij(s))pds]}4p1maxijΛ{(pqa_0ij)pqE[tepqtsa0ij(u)du(CklNh1(i,j)(Cklij(s))q)pq×ijΛ(2κLf)pϕij(s)pds]}4p1maxijΛ{(pqa_0ij)pqqpa_0ij(CklNh1(i,j)(Cklij+)q)pq(2κLf)p}ϕpX. (3.5)

    Similarly, one has

    F14p1maxijΛ{(pqa_0ij)pqqpa_0ij(ˉacij)p}ϕpX, (3.6)
    F34p1maxijΛ{(pqa_0ij)pqqpa_0ij(CklNh2(i,j)(Bklij+)q)pq(2κLg0|Kij(u)|du)p}ϕpX. (3.7)

    By the Burkolder-Davis-Gundy inequality and the Hölder inequality, when p>2, we infer that

    F44p1CpmaxijΛ{E[tetsa0ij(u)duCklNh3(i,j)Eklij(s)δij(ϕij(sσij(s)))2ds]p2}4p1CpmaxijΛ{E[e2tsa0ij(u)duCklNh3(i,j)Eklijδij(ϕij(sσij(s)))2ds]p2}4p1CpmaxijΛ{E[t(e2tsa0ij(u)du)pp2×1pds]p2p×p2×E[t(e2tsa0ij(u)du)1q×p2(CklNh3(i,j)Eklij(s)δijϕij(sσij(s))2)p2ds]}4p1CpmaxijΛ{(p22a_0ij)p22qpa_0ijECklNh3(i,j)Eklij(s)δij(ϕij(sσij(s)))p}4p1CpmaxijΛ{(p22a_0ij)p22qpa_0ij(CklNh3(i,j)(Eklij+)q)pq(Lδij)p}ϕpX. (3.8)

    When p=2, by the Itˆo isometry, it follows that

    F44maxijΛ{E[te2tsa0ij(u)duCklNh3(i,j)Eklij(s)δij(ϕij(sσij(s)))2Ads]}4maxijΛ{12a_0ijCklNh3(i,j)(Eklij+)2(Lδij)2}ϕ2X. (3.9)

    Putting (3.5)–(3.9) into (3.4), we obtain that

    Ψϕϕ0pX4p1maxijΛ{(pqa_0ij)pqqpa_0ij[(ˉacij)p+(CklNh1(i,j)(Cklij+)q)pq(2κLf)p+(CklNh2(i,j)(Bklij+)q)pq(2κLg0|Kij(u)|du)p]+Cp(p22a_0ij)p22qpa_0ij(CklNh3(i,j)(Eklij+)q)pq(Lδij)p}ϕpXκp,(p>2), (3.10)

    and

    Ψϕϕ02X4maxijΛ{1(aij)2[(ˉacij)2+CklNh1(i,j)(Cklij+)2(2κLf)2+CklNh2(i,j)(Bklij+)2(2κLg×0|Kij(u)|du)2]+12a_0ijCklNh3(i,j)(Eklij+)2(Lδij)2}ϕ2Xκ2,(p=2). (3.11)

    It follows from (3.10), (3.11) and (A3) that Ψϕϕ0Xκ.

    Then, using the same method as that in the proof of Theorem 3.2 in [21], we can show that Ψϕ is Lp-uniformly continuous. Therefore, we have Ψ(X)X.

    Last, we will show that Ψ is a contraction mapping. Indeed, for any ψ,φX, when p>2, we have

    E(Φφ)(t)(Φψ)(t)p04p1maxijΛ{Etetsa0ij(u)du(acij(s)φij(s)+acij(s)ψij(s))dsp}+4p1maxijΛ{Etetsa0ij(u)duCklNh1(i,j)Cklij(s)[f(φkl(sτkl(s)))φij(s)f(ψkl(sτkl(s)))ψij(s)]dsp}+4p1maxijΛ{Etetsa0ij(u)duCklNh2(i,j)Bklij(s)×[0Kij(u)g(φkl(su))duφij(s)0Kij(u)g(ψkl(su))duψij(u)]dsp}+4p1maxijΛ{Etetsa0ij(u)duCklNh3(i,j)Eklij(s)[δij(φij(sσij(s)))δij(ψij(sσij(s)))]dωij(s)p}4p1maxijΛ{(pqa_0ij)pqqpa_0ij[(ˉacij)p+(CklNh1(i,j)(Cklij+)q)pq(2κLf+Mf)p+(CklNh2(i,j)(Bklij+)q)pq((2κLg+Mg)0|Kij(u)|du)p]+Cp(p22a_0ij)p22qpa_0ij×(CklNh3(i,j)(Eklij+)q)pq(Lδij)p}φψpX. (3.12)

    Similarly, for p=2, we can get

    E(Φφ)(t)(Φψ)(t)204maxijΛ{1(a_0ij)2[(ˉacij)2+CklNh1(i,j)(Cklij+)2(2κLf+Mf)2+CklNh2(i,j)(Bklij+)2×((2κLg+Mg)0|Kij(u)|du)2]+12a_0ijCklNh3(i,j)(Eklij+)2(Lδij)2}φψ2X. (3.13)

    From (3.12) and (3.13) it follows that

    (Φφ)(t)(Φψ)(t)Xpr1φψX,(p>2),(Φφ)(t)(Φψ)(t)Xr2φψX,(p=2).

    Hence, by virtue of (A3), Ψ is a contraction mapping. So, Ψ has a unique fixed point x in X, i.e., (2.1) has a unique solution x in X.

    Theorem 3.2. Assume that (A1)(A5) hold. Then the system (2.1) has a unique p-th Stepanov-like almost periodic solution in the distribution sense in X={ϕX:ϕϕ0Xκ}, where κ is a constant satisfying κϕ0X.

    Proof. From Theorem 3.1, we know that (2.1) has a unique solution x in X. Now, let us show that x is Stepanov-like almost periodic in distribution. Since xX, it is Lp-uniformly continuous and satisfies x2κ. So, for any ε>0, there exists δ(0,ε), when |h|<δ, we have suptREx(t+h)x(t)p0<ε. Hence, we derive that

    supξRξ+1ξEx(t+h)x(t)p0dt<ε. (3.14)

    For the δ above, according to (A2), we have, for ijΛ,

    |a0ij(t+τ)a0ij(t)|<δ,acij(t+τ)acij(t)p<δ,Cklij(t+τ)Cklij(t)p<δ,|τij(t+τ)τij(t)|<δ,Bklij(t+τ)Bklij(t)p<δ,Eklij(t+τ)Eklij(t)p<δ,|σij(t+τ)σij(t)|<δ,supξRξ+1ξLij(t+τ)Lij(t)pdt<δ.

    As |τij(t+τ)τij(t)|<δ, by (3.14), there holds

    supξRξ+1ξEx(sτij(s+τ))x(sτij(s))p0ds<ε.

    Based on (3.2), we can infer that

    xij(t+τ)=tetsa0ij(u+τ)du[acij(s+τ)xij(s+τ)+CklNh1(i,j)Cklij(s+τ)×f(xkl(s+ττkl(s+τ)))xij(s+τ)+CklNh2(i,j)Bklij(s+τ)0Kij(u)×g(xkl(s+τu))duxij(s+τ)+Lij(s+τ)]ds+tetsaij(u+τ)du×CklNh3(i,j)Eklij(s+τ)δij(xij(s+τσij(s+τ)))d[ωij(s+τ)ωij(τ)],

    in which ijΛ,ωij(s+τ)ωij(τ) is a Brownian motion having the same distribution as ωij(s).

    Let us consider the process

    xij(t+τ)=tetsa0ij(u+τ)du[acij(s+τ)xij(s+τ)+CklNh1(i,j)Cklij(s+τ)×f(xkl(s+ττkl(s+τ)))xij(s+τ)+CklNh2(i,j)Bklij(s+τ)0Kij(u)×g(xkl(s+τu))duxij(s+τ)+Lij(s+τ)]ds+tetsaij(u+τ)du×CklNh3(i,j)Eklij(s+τ)δij(xij(s+τσij(s+τ)))dωij(s). (3.15)

    From (3.2) and (3.15), we deduce that

    ξ+1ξEx(t+τ)x(t)p0dt16p1maxijΛ{ξ+1ξEtetsa0ij(u+τ)duCklNh1(i,j)Cklij(s+τ)×[f(xkl(s+ττkl(s+τ)))xij(s+τ)f(xkl(sτkl(s)))xij(s+τ)]dspdt}+16p1maxijΛ{ξ+1ξEtetsa0ij(u+τ)duCklNh1(i,j)(Cklij(s+τ)Cklij(s))×f(xkl(sτkl(s)))xij(s+τ)dspAdt}+16p1maxijΛ{ξ+1ξEtetsa0ij(u+τ)duCklNh1(i,j)Cklij(s)×f(xkl(sτkl(s)))(xij(s+τ)xij(s))dspdt}+16p1maxijΛ{ξ+1ξEt(etsa0ij(u+τ)duetsa0ij(u)du)CklNh1(i,j)Cklij(t)×f(xkl(sτkl(s)))xij(s)dspdt}+16p1maxijΛ{ξ+1ξEtetsa0ij(u+τ)duCklNh2(i,j)Bklij(s+τ)[0Kij(u)g(xkl(s+τu))duxij(s+τ)0Kij(u)×g(xkl(su))duxij(s+τ)]dspdt}+16p1maxijΛ{ξ+1ξEtetsa0ij(u+τ)duCklNh2(i,j)(Bklij(s+τ)Bklij(s))0Kij(u)g(xkl(su))duxij(s+τ)dspdt}+16p1maxijΛ{ξ+1ξEtetsa0ij(u+τ)duCklNh2(i,j)Bklij(s)0Kij(u)g(xkl(su))du×(xij(s+τ)xij(s))dspdt}+16p1maxijΛ{ξ+1ξEt(etsa0ij(u+τ)duetsa0ij(u)du)CklNh2(i,j)Bklij(s)0Kij(u)g(xkl(su))duxij(s)dspdt}+16p1maxijΛ{ξ+1ξEtetsa0ij(u+τ)du(Lij(s+τ)Lij(s))dspdt}+16p1maxijΛ{ξ+1ξEt(etsa0ij(u+τ)duetsa0ij(u)du)Lij(s)dspdt}+16p1maxijΛ{ξ+1ξEtetsaij(u+τ)duCklNh3(i,j)Eklij(s+τ)×[δij(xij(s+τσij(s+τ)))δij(xij(sσij(s)))]dωij(s)pAdt}+16p1maxijΛ{ξ+1ξEtetsa0ij(u+τ)duCklNh3(i,j)(Eklij(s+τ)Eklij(s))×δij(xij(sσij(s)))dωij(s)pdt}+16p1maxijΛ{ξ+1ξEt(etsa0ij(u+τ)duetsa0ij(u)du)CklNh3(i,j)Eklij(t)δij(xij(sσij(s)))dωij(s)pdt}+16p1maxijΛ{ξ+1ξEtetsa0ij(u+τ)du(acij(s)acij(s+τ))xij(s+τ)dspdt}+16p1maxijΛ{ξ+1ξEtetsaij(u+τ)duacij(s)(xij(s)xij(s+τ))dspdt}+16p1maxijΛ{ξ+1ξEt(etsaij(u+τ)duetsaij(u)du)(acij(s))xij(s)dspdt}:=16i=1Hi. (3.16)

    Employing the Hölder inequality, we can obtain

    H132p1maxijΛ{(pqa_0ij)pq(CklNh1(i,j)(Cklij+)q)pq(Lf)pCklNh1(i,j)ξ+1ξ[tepq(ts)a_0ij×E[x(s+ττkl(s+τ))x(sτkl(s+τ))]x(t+τ)p0ds+tepq(ts)a_0ijE[x(sτkl(s+τ))x(sτkl(s))]x(t+τ)p0ds]dt}.

    By a change of variables and Fubini's theorem, we infer that

    (3.17)

    where

    Similarly, when , one can obtain

    (3.18)

    where

    (3.19)

    and when , we have

    (3.20)

    where

    (3.21)

    In the same way, we can get

    (3.22)
    (3.23)
    (3.24)
    (3.25)
    (3.26)
    (3.27)
    (3.28)
    (3.29)

    Noting that

    (3.30)

    We can gain

    (3.31)
    (3.32)
    (3.33)
    (3.34)

    when , we have

    (3.35)

    for , we get

    (3.36)

    Substituting (3.17)–(3.36) into (3.16), we have the following two cases:

    Case 1. When , we have

    where is the same as that in and

    By , we know . Hence, we derive that

    (3.37)

    Case 2. When , we can obtain

    where is defined in and

    Similar to the previous case, by , we know and hence, we can get that

    (3.38)

    Noting that

    Hence, we have

    (3.39)

    Combining (3.37)–(3.39), we can conclude that is -th Stepanov almost periodic in the distribution sense. The proof is complete.

    Similar to the proof of Theorem 3.7 in [21], one can easily show that.

    Theorem 3.3. Suppose that are fulfilled and let be the Stepanov almost periodic solution in the distribution sense of system (2.1) with initial value . Then there exist constants and such that for an arbitrary solution with initial value satisfies

    where , i.e., the solution is globally exponentially stable.

    The purpose of this section is to demonstrate the effectiveness of the results obtained in this paper through a numerical example.

    In neural network (2.1), choose , , and

    and let . Then we get

    Take , , then we have

    And when , we have

    Thus, all assumptions in Theorems 3.2 and 3.3 are fulfilled. So we can conclude that the system (2.1) has a unique -almost periodic solution in the distribution sense which is globally exponentially stable.

    The results are also verified by the numerical simulations in Figures 14.

    Figure 1.  Global exponential stability of states and of (2.1).
    Figure 2.  Global exponential stability of states and of (2.1).
    Figure 3.  Global exponential stability of states and of (2.1).
    Figure 4.  Global exponential stability of states and of (2.1).

    From these figures, we can observe that when the four primitive components of each solution of this system take different initial values, they eventually tend to stabilize. It can be seen that these solutions that meet the above conditions do exist and are exponentially stable.

    In this article, we establish the existence and global exponential stability of Stepanov almost periodic solutions in the distribution sense for a class of stochastic Clifford-valued SICNNs with mixed delays. Even when network (2.1) degenerates into a real-valued NN, the results of this paper are new. In fact, uncertainty, namely fuzziness, is also a problem that needs to be considered in real system modeling. However, we consider only the disturbance of random factors and do not consider the issue of fuzziness. In a NN, considering the effects of both random perturbations and fuzziness is our future direction of effort.

    The authors declare that they have not used Artificial Intelligence (AI) tools in the creation of this article.

    This work is supported by the National Natural Science Foundation of China under Grant No. 12261098.

    The authors declare that they have no conflicts of interest.



    [1] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, (2016), 770–778. https://doi.org/10.1109/CVPR.2016.90
    [2] S. Singaravel, J. Suykens, P. Geyer, Deep-learning neural-network architectures and methods: Using component-based models in building-design energy prediction, Adv. Eng. Inf., 38 (2018), 81–90. https://doi.org/10.1016/j.aei.2018.06.004 doi: 10.1016/j.aei.2018.06.004
    [3] H. Xu, J. Kong, M. Liang, H. Sun, M. Qi, Video behavior recognition based on actional-structural graph convolution and temporal extension module, Electron. Res. Arch., 30 (2022), 4157–4177. https://doi.org/10.3934/era.2022210 doi: 10.3934/era.2022210
    [4] D. Peng, Y. Lei, H. Munawar, Y. Guo, W. Li, Semantic-aware domain generalized segmentation, in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, (2022), 2584–2595. https://doi.org/10.1109/CVPR52688.2022.00262
    [5] T. Korbak, K. Shi, A Chen, R. V. Bhalerao, C. Buckley, J. Phang, et al., Pretraining language models with human preferences, in Proceedings of the 40th International Conference on Machine Learning, PMLR, 202 (2023), 17506–17533.
    [6] A. Krizhevsky, I. Sutskever, G. Hinton, Imagenet classification with deep convolutional neural networks, Commun. ACM, 60 (2017), 84–90. https://doi.org/10.1145/3065386 doi: 10.1145/3065386
    [7] K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition, preprint, arXiv: 1409.1556.
    [8] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, et al., Going deeper with convolutions, in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, (2015), 1–9. https://doi.org/10.1109/CVPR.2015.7298594
    [9] G. Huang, Z. Liu, L. Van Der Maaten, K. Q. Weinberger, Densely connected convolutional networks, in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, (2017), 2261–2269. https://doi.org/10.1109/CVPR.2017.243
    [10] J. Xi, Z. Xu, Z. Yan, W. Liu, Y. Liu, Portrait age recognition method based on improved ResNet and deformable convolution, Electron. Res. Arch., 31 (2023), 6585–6599. https://doi.org/10.3934/era.2023333 doi: 10.3934/era.2023333
    [11] C. Swarup, K. U. Singh, A. Kumar, S. K. Pandey, N. Varshney, T. Singh, Brain tumor detection using CNN, AlexNet & GoogLeNet ensembling learning approaches, Electron. Res. Arch., 31 (2023), 2900–2924. https://doi.org/10.3934/era.2023146 doi: 10.3934/era.2023146
    [12] B. Zoph, Q. V. Le, Neural architecture search with reinforcement learning, preprint, arXiv: 1611.01578.
    [13] T. Elsken, J. H. Metzen, F. Hutter, Neural architecture search: A survey, J. Mach. Learn. Res., 20 (2019), 1997–2017.
    [14] Y. Xue, W. Tong, F. Neri, P. Chen, T. Luo, L. Zhen, et al., Evolutionary architecture search for generative adversarial networks based on weight sharing, IEEE Trans. Evol. Comput., 2023 (2023), 1. https://doi.org/10.1109/TEVC.2023.3338371 doi: 10.1109/TEVC.2023.3338371
    [15] Y. Xue, X. Han, Z. Wang, Self-adaptive weight based on dual-attention for differentiable neural, IEEE Trans. Ind. Inf., 2024 (2024), 1–10. https://doi.org/10.1109/TII.2023.3348843 doi: 10.1109/TII.2023.3348843
    [16] Y. Xue, Z. Zhang, F. Neri, Similarity surrogate-assisted evolutionary neural architecture search with dual encoding strategy, Electron. Res. Arch., 32 (2024), 1017–1043. https://doi.org/10.3934/era.2024050 doi: 10.3934/era.2024050
    [17] H. Liu, K. Simonyan, Y. Yang, DARTS: Differentiable architecture search, preprint, arXiv: 1806.09055.
    [18] Y. Xue, J. Qin, Partial connection based on channel attention for differentiable neural architecture search, IEEE Trans. Ind. Inf., 19 (2023), 6804–6813. https://doi.org/10.1109/TII.2022.3184700 doi: 10.1109/TII.2022.3184700
    [19] Y. Xue, C. Lu, F. Neri, J. Qin, Improved differentiable architecture search with multi-stage progressive partial channel connections, IEEE Trans. Emerging Top. Comput. Intell., 8 (2024), 32–43. https://doi.org/10.1109/TETCI.2023.3301395 doi: 10.1109/TETCI.2023.3301395
    [20] Y. Liu, Y. Sun, B. Xue, M. Zhang, G. G. Yen, K. C. Tan, A survey on evolutionary neural architecture search, IEEE Trans. Neural Networks Learn. Syst., 34 (2023), 550–570. https://doi.org/10.1109/TNNLS.2021.3100554 doi: 10.1109/TNNLS.2021.3100554
    [21] Y. Xue, C. Chen, A. Slowik, Neural architecture search based on a multi-objective evolutionary algorithm with probability stack, IEEE Trans. Evol. Comput., 27 (2023), 778–786. https://doi.org/10.1109/TEVC.2023.3252612 doi: 10.1109/TEVC.2023.3252612
    [22] E. Real, S. Moore, A. Selle, S. Saxena, Y. L. Suematsu, J. Tan, et al., Large-scale evolution of image classifiers, in Proceedings of the 34th International Conference on Machine Learning (ICML), PMLR, 70 (2017), 2902–2911.
    [23] Y. Sun, B. Xue, M. Zhang, G. G. Yen, Completely automated CNN architecture design based on blocks, IEEE Trans. Neural Networks Learn. Syst., 31 (2020), 1242–1254. https://doi.org/10.1109/TNNLS.2019.2919608 doi: 10.1109/TNNLS.2019.2919608
    [24] Y. Xue, Y. Wang, J. Liang, A. Slowik, A self-adaptive mutation neural architecture search algorithm based on blocks, IEEE Comput. Intell. Mag., 16 (2021), 67–78. https://doi.org/10.1109/MCI.2021.3084435 doi: 10.1109/MCI.2021.3084435
    [25] D. Song, C. Xu, X. Jia, Y. Chen, C. Xu, Y. Wang, Efficient residual dense block search for image super-resolution, in Proceedings of the AAAI conference on artificial intelligence, AAAI Press, 34 (2020), 12007–12014. https://doi.org/10.1609/aaai.v34i07.6877
    [26] J. Fang, Y. Sun, Q. Zhang, Y. Li, W. Liu, X. Wang, Densely connected search space for more flexible neural architecture search, in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, (2020), 10625–10634. https://doi.org/10.1109/CVPR42600.2020.01064
    [27] J. Kennedy, R. C. Eberhart, Particle swarm optimization, in Proceedings of ICNN'95 - International Conference on Neural Networks, IEEE, 4 (1995), 1942–1948. https://doi.org/10.1109/ICNN.1995.488968
    [28] B. Wang, B. Xue, M. Zhang, Particle swarm optimisation for evolving deep neural networks for image classification by evolving and stacking transferable blocks, in 2020 IEEE Congress on Evolutionary Computation (CEC), IEEE, (2020), 1–8. https://doi.org/10.1109/CEC48606.2020.9185541
    [29] E. Real, A. Aggarwal, Y. Huang, Q. V. Le, Regularized evolution for image classifier architecture search, in Proceedings of the AAAI Conference on Artificial Intelligence, AAAI Press, 33 (2019), 4780–4789. https://doi.org/10.1609/aaai.v33i01.33014780
    [30] G. Huang, S. Liu, L. v. d. Maaten, K. Q. Weinberger, CondenseNet: An efficient denseNet using learned group convolutions, in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, (2018), 2752–2761. https://doi.org/10.1109/CVPR.2018.00291
    [31] H. Cai, T. Chen, W. Zhang, Y. Yu, J. Wang, Regularized evolution for image classifier architecture search, in Proceedings of the AAAI Conference on Artificial Intelligence, AAAI Press, 32 (2018), 4780–4789. https://doi.org/10.1609/aaai.v32i1.11709
    [32] B. Baker, O. Gupta, N. Naik, R. Raskar, Designing neural network architectures using reinforcement learning, preprint, arXiv: 1611.02167.
    [33] L. Xie, A. Yuille, Genetic CNN, in 2017 IEEE International Conference on Computer Vision (ICCV), IEEE, (2017), 1388–1397. https://doi.org/10.1109/ICCV.2017.154
    [34] H. Liu, K. Simonyan, O. Vinyals, C. Fernando, K. Kavukcuoglu, Hierarchical representations for efficient architecture search, preprint, arXiv: 1711.00436.
    [35] A. I. Sharaf, E. S. F. Radwan, An automated approach for developing a convolutional neural network using a modified firefly algorithm for image classification, in Applications of Firefly Algorithm and its Variants, Springer, (2020), 99–118. https://doi.org/10.1007/978-981-15-0306-1_5
    [36] G. Cuccu, J. Togelius, P. Cudre-Mauroux, Playing atari with six neurons (Extended abstract), in Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20), International Joint Conferences on Artificial Intelligence Organization, (2020), 4711–4715. https://doi.org/10.24963/ijcai.2020/651
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1251) PDF downloads(46) Cited by(1)

Figures and Tables

Figures(8)  /  Tables(2)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog