Processing math: 100%
Research article

Distributed convex optimization of bipartite containment control for high-order nonlinear uncertain multi-agent systems with state constraints


  • Received: 14 July 2023 Revised: 16 August 2023 Accepted: 28 August 2023 Published: 07 September 2023
  • This article investigates a penalty-based distributed optimization algorithm of bipartite containment control for high-order nonlinear uncertain multi-agent systems with state constraints. The proposed method addresses the distributed optimization problem by designing a penalty function in the form of a quadratic function, which is the sum of the global objective function and the consensus constraint. Moreover, the observer is presented to address the unmeasurable state of each agent. Radial basis function neural networks (RBFNN) are employed to approximate the unknown nonlinear functions. Then, by integrating RBFNN and dynamic surface control (DSC) techniques, an adaptive backstepping controller based on the barrier Lyapunov function (BLF) is proposed. Finally, the effectiveness of the suggested control strategy is verified under the condition that the state constraints are not broken. Simulation results indicate that the output trajectories of all agents remain within the upper and lower boundaries, converging asymptotically to the global optimal signal.

    Citation: Yuhang Yao, Jiaxin Yuan, Tao Chen, Xiaole Yang, Hui Yang. Distributed convex optimization of bipartite containment control for high-order nonlinear uncertain multi-agent systems with state constraints[J]. Mathematical Biosciences and Engineering, 2023, 20(9): 17296-17323. doi: 10.3934/mbe.2023770

    Related Papers:

    [1] Jinxin Du, Lei Liu . Adaptive fuzzy fixed time formation control of state constrained nonlinear multi-agent systems against FDI attacks. Mathematical Biosciences and Engineering, 2024, 21(3): 4724-4741. doi: 10.3934/mbe.2024207
    [2] Qiushi Wang, Hongwei Ren, Zhiping Peng, Junlin Huang . Dynamic event-triggered consensus control for nonlinear multi-agent systems under DoS attacks. Mathematical Biosciences and Engineering, 2024, 21(2): 3304-3318. doi: 10.3934/mbe.2024146
    [3] Run Tang, Wei Zhu, Huizhu Pu . Event-triggered distributed optimization of multi-agent systems with time delay. Mathematical Biosciences and Engineering, 2023, 20(12): 20712-20726. doi: 10.3934/mbe.2023916
    [4] Xinxin Zhang, Huaiqin Wu . Bipartite consensus for multi-agent networks of fractional diffusion PDEs via aperiodically intermittent boundary control. Mathematical Biosciences and Engineering, 2023, 20(7): 12649-12665. doi: 10.3934/mbe.2023563
    [5] Meiqiao Wang, Wuquan Li . Distributed adaptive control for nonlinear multi-agent systems with nonlinear parametric uncertainties. Mathematical Biosciences and Engineering, 2023, 20(7): 12908-12922. doi: 10.3934/mbe.2023576
    [6] Tong Guo, Jing Han, Cancan Zhou, Jianping Zhou . Multi-leader-follower group consensus of stochastic time-delay multi-agent systems subject to Markov switching topology. Mathematical Biosciences and Engineering, 2022, 19(8): 7504-7520. doi: 10.3934/mbe.2022353
    [7] Xiaofeng Chen . Double-integrator consensus for a switching network without dwell time. Mathematical Biosciences and Engineering, 2023, 20(7): 11627-11643. doi: 10.3934/mbe.2023516
    [8] Hebing Zhang, Xiaojing Zheng . Invariable distribution of co-evolutionary complex adaptive systems with agent's behavior and local topological configuration. Mathematical Biosciences and Engineering, 2024, 21(2): 3229-3261. doi: 10.3934/mbe.2024143
    [9] Siyu Li, Shu Li, Lei Liu . Fuzzy adaptive event-triggered distributed control for a class of nonlinear multi-agent systems. Mathematical Biosciences and Engineering, 2024, 21(1): 474-493. doi: 10.3934/mbe.2024021
    [10] Ryoto Himo, Masaki Ogura, Naoki Wakamiya . Iterative shepherding control for agents with heterogeneous responsivity. Mathematical Biosciences and Engineering, 2022, 19(4): 3509-3525. doi: 10.3934/mbe.2022162
  • This article investigates a penalty-based distributed optimization algorithm of bipartite containment control for high-order nonlinear uncertain multi-agent systems with state constraints. The proposed method addresses the distributed optimization problem by designing a penalty function in the form of a quadratic function, which is the sum of the global objective function and the consensus constraint. Moreover, the observer is presented to address the unmeasurable state of each agent. Radial basis function neural networks (RBFNN) are employed to approximate the unknown nonlinear functions. Then, by integrating RBFNN and dynamic surface control (DSC) techniques, an adaptive backstepping controller based on the barrier Lyapunov function (BLF) is proposed. Finally, the effectiveness of the suggested control strategy is verified under the condition that the state constraints are not broken. Simulation results indicate that the output trajectories of all agents remain within the upper and lower boundaries, converging asymptotically to the global optimal signal.



    Recently, due to its remarkable computational performance and scalability, distributed convex optimization has garnered significant attention from researchers [1,2]. Distributed convex optimization algorithms effectively address optimization problems in complex and large-scale network environments by integrating traditional optimization theories with recently developed theoretical techniques for coordinated control of multi-agent systems (MASs) [3,4]. Compared to centralized optimization, distributed convex optimization offers improved scalability, robustness, privacy and flexibility, making it widely applicable in various fields, such as economic dispatching of smart grid systems [5,6], resource assignment for sensor networks [7], large-scale machine learning [8] and distributed source location and estimation [9].

    In distributed convex optimization, agents are assigned local objective functions, and the global objective function is constructed by aggregating these local objective functions. Through continuous information exchange in the communication network, the optimal solution to the global objective function is eventually obtained. As an extension of the consensus problem in MASs, the consensus problem in distributed optimization has garnered increased attention from researchers. In engineering practices, designing distributed protocols to achieve both consensus and performance optimization in MASs is highly reasonable. In UAV formation control, aircraft should consume a small amount of fuel when the speed and attitude are consistent. Discrete-time distributed consensus optimization algorithms for MASs have been studied in [10,11]. However, to meet the requirements of practical applications, research on a distributed consensus in continuous-time MASs [12,13] is more prevalent.

    Actually, most distributed optimization problems can be reformulated as optimization problems with consensus constraints. For example, in [14], a distributed consensus optimization algorithm based on a penalty function was proposed for a first-order system. For first-order [15,16] and second-order [17,18,19] MASs, several distributed optimization algorithms have been presented. For instance, a non-uniform gradient gain was proposed in [20], and a continuous-time zero-gradient-sum algorithm was designed in [21]. However, in comparison to the articles [20,21,22], a more concise distributed optimization algorithm based on the penalty function was proposed in [23], which avoided the involvement of distributed estimations.

    The objective of this paper is to devise a distributed controller that facilitates the collaborative convergence of all agents towards the optimal solution. For high-order MASs, there are various control methods for solving distributed optimization. For example, a distributed output feedback integral controller is proposed to address the output consensus problem for the distributed optimization in [24]. A projection-based second-order control algorithm was proposed in [19]. A Non-smooth embedded control framework was designed in [25]. In [26], the penalty function method and the additive power integrator technique were combined. However, when the states are unmeasurable, the aforementioned distributed optimization problem cannot be solved. Therefore, the unmeasurable system states highly affect the running of high-order nonlinear MASs. Compared to other control methods, adaptive control is widely used and exhibits a robustness to system uncertainties.

    As suggested by the general approximation theory, radial basis function neural networks (RBFNN) and fuzzy logic systems are applicable in addressing the uncertainty of nonlinear systems. In [27], a fuzzy state observer based on radial basis functions was presented to estimate unavailability states. In [28], for a class of nonlinear MASs distributed control with strict feedback, adaptive neural networks were employed to approximate uncertain states. Moreover, RBFNN and backstepping techniques were combined to} construct an adaptive controller in [29]. Dynamic surface control (DSC) was employed to avoid the complexity explosion in the backstepping method [29,30,31]. For high-order nonlinear uncertain MASs, the utilization of the adaptive neural network backstepping control method to solve distributed optimization problems is relatively rare. Therefore, designing an adaptive controller to address the distributed optimization problem for high-order nonlinear MASs with unknown states is a meaningful task.

    Furthermore, these above protocols are designed with little regard to the control performance and state constraints, which is impractical in systems with either limited resources or actuators. As a matter of fact, the problem of being state-constrained is common in practical systems. The barrier Lyapunov function (BLF) method has been presented by domestic and foreign scholars [32], keeping the system state within constraints. In [33], the employment of BLF ensured the boundedness of the entire system state and constrained all closed-loop signals within a compact set. In [34], the log-asymmetric BLF was proposed, combined with the DSC technique and an adaptive backstepping controller is designed to solve the constraints. With the extension of the nonlinear system with state constraints, the BLF is applied more widely. However, distributed optimization for high-order nonlinear uncertain MASs with state constraints have not been studied, which provides the research motivation of this paper.

    Based on previous observations and discussions, this paper focuses on investigating the distributed optimization problem of high-order MASs with state constraints by using the penalty function method. First, by integrating RBFNN and DSC techniques, an adaptive backstepping controller is proposed. From the second step, the introduction of the BLF guarantees the preservation of state constraints. The unknown state of the system is observed by a state observer. The key research contributions of the article are outlined below in comparison to earlier works.

    1) Differing from the adaptive controller designed only for the consensus problem [35,36] and the containment problem [37], we take the distributed optimization problem into account to enhance the system performance. Besides, compared to the bipartite consensus control studied in [38,39], we introduce two virtual leaders to achieve a bipartite containment effect. The trajectory of the optimal solution in the distributed optimization remains within the convex hulls delineated by the upper and lower reference signals.

    2) Unlike [40] and [26], in which the distributed optimization for high-order nonlinear MASs is studied, in this paper, for purpose of meeting the practical requests of the system, we consider state constraints in distributed optimization problems. The introduction of the BLF guarantees the preservation of state constraints. Distributed optimization for high-order nonlinear uncertain MASs with state constraints have not been studied before.

    3) In contrast to [14,15,41], in which the distributed optimization algorithm is proposed for the fist-order nonlinear MASs, in this paper, we focus on the high-order MASs with unknown variables, which have an increased engineering application value. Combined with RBFNN technology, the observer is designed to estimate the states of each agent.

    The structure of the remainder of this paper is presented below. The second part introduces some reserve knowledge, such as the basic theory of the multi-agent and distributed optimization principle. In Section 3, an adaptive observer is prepared and we create a BLF-based adaptive backstepping controller, and a DSC technique is utilized to update virtual control law constantly. In Section 4, simulations are carried out to verify the validity of the controller. The final section summarizes the whole text and draws some conclusions.

    Based on the information interaction between multiple agents, we use digraph G=(w,χ,ˉA) to represent the relationship between agents. w={n1,...,nN} represents a set of node and χ={(ni,nj)}w×w is defined as a set of edges. Information can be transferred between the agent i and agent j. There exists Ni={j|(ni,nj)χ}, which expresses the set of neighboring agents for agents i. ˉA={aij}RN×N is the adjacency matrix and the element aij of it has two possible values. When (ni,nj)χ, aij=1; in other cases, aij=0. It is presumed that aij=0. Then, we design D=diag(d1,...,dN) as the diagonal matrix, di=jn=Niaij. The Laplacian matrix can be defined as L=DˉA.

    Consider the following nonlinear high-order multi-agent system:

    {˙xi,p=xi,p+1+hi,p(xi,1,,xi,p)˙xi,n=ui(t)+hi,n(xi,1,,xi,n)yi=xi,1 (2.1)

    where yi represents the system output and hi,p(),(p=1,2,...,n1) are unknown nonlinear functions. The control input is defined as ui. Xi,p=(xi,1,,xi,p)TRp is defined as the system state vectors. Rewrite the system (2.1) into the form below:

    ˙Xi,n=AiXi,n+Tiyi+np=1Bi,p[hi,p(Xi,p)]+Biui (2.2)

    where Hurwitz matrix Ai=[εi,1In1εi,n00], Ti=[Ti,1Ti,n], Bi=[01], Bi,p=[010]. QTi=Qi and PTi=Pi are predetermined positive matrix,

    ATiPi+PiAi=2Qi. (2.3)

    In this paper, consider the BLF of the ith agent as follows:

    Vi,1=12logki,b12ki,b12si,12 (2.4)

    where ki,b1 is the constraint on si,1 and |si,1|ki,b1.

    Lemma 1. ([42,43]) For any given positive constant ki,b1 and si,1 satisfying |si,1|ki,b1, we have

    logki,b12ki,b12si,12<si,12ki,b12si,12. (2.5)

    To facilitate subsequent calculations, we refer to the following lemmas.

    Lemma 2. ([44]) For an undirected connected communication topology, distributed optimization problems can be

    minxRNmNi=1fi(xi),s.t.(LIm)x=0Nm (2.6)

    where x=[x1T,,xNT]T. Utilizing the principles of the penalty function theory, the approximate optimization problem is formulated as follows:

    minxRNmNi=1fi(xi)+12ϑxT(LIm)x (2.7)

    where ϑ>0 is a constant penalty parameter and 12ϑxT(LIm)x is the penalty term for violating the consensus constraint LImx=0Nm.

    Lemma 3. ([45]) The inequality relationship shown below is valid

    xTynaaxa+1bnbyb (2.8)

    where x,yRn,a,b>1,n>0, and (a1)(b1)=1.

    The distributed optimization problem is the subject of this paper, so for N agents, the global objective function f is the sum of strictly convex local objective functions fi:

    f(xi,1)=Ni=1fi(xi,1). (2.9)

    Define x1=[x1,1 x2,1  xN,1]T. According to [46], 1N is eigenvector for eigenvalue 0 of Laplacian matrix; for some αR, if x1=α1N, we obtain

    Lx1=0 (2.10)
    xT1Lx1=0. (2.11)

    It should be noted that in this paper xRN, therefore, based on Lemma 2, we construct the following penalty function:

    P(x1)=Ni=1fi(xi,1)+xT1Lx1. (2.12)

    Our target is as follows

    (x1,1,,xN,1)=argmin(x1,1,,xN,1)P(x1). (2.13)

    The local cost function of the ith agent fi(xi,1) can be designed as:

    fi(xi,1)=ai,1(xi,1xd1)2+ai,2(xi,1xd2)2+c=aix2i,1+bixi,1+ci (2.14)

    where xd1 and xd2 are the upper and lower bound of the trajectory of motion, ai=ai,1+ai,2, ai>0, bi=2ai,1xd12ai,2xd2, ci=ai,1xd12+ai,2xd22+c and ai,c are scalars.

    Remark 1. Based on the strong convexity of the quadratic function, the global objective function designed in this paper is composed of quadratic functions. To solve the optimization problem of the bipartite containment control, it involves multiple virtual leaders. By constructing a penalty function, all followers are converged to the optimal solution of the distributed optimization problem, which lies within the convex hull of the trajectories of each virtual leader and their opposite trajectories.

    Control objectives: The purpose of this paper is to design a distributed optimization controller, so that the optimal solution for the trajectory moving between the upper and lower bounds of the virtual leaders under state constraints can be found.

    Assumption 1. The unknown functions hi,p(Xi,p), (i=1, , n) can be shown as follows:

    hi,p(Xi,p|θi,p)=θTi,pΨi,p(Xi,p),1in (3.1)

    in which Ψi,p(Xi,p) delegates the Gaussian basis function vector, and θi,p represents the ideal constant vector.

    In this paper, an observer is placed to estimate the agent's unmeasurable states, since the assumption that the state variables given in (2.2) are unavailable. The observer is presented as follows:

    ˙ˆXi,n=AiˆXi,n+Tiyi+np=1Bi,p[ˆhi,p(ˆXi,p|θi,p)]+Biui^yi=CiˆXi,n (3.2)

    where Ci=[100], ˆXi,p=(ˆxi,1,ˆxi,2,...,ˆxi,p)T are the estimated values of Xi,p.

    Let ei=Xi,nˆXi,n be state observation errors of system (2.1). Combining Eqs (2.2) and (3.2), we get

    ˙ei=Aiei+np=1Bi,p[hi,p(ˆXi,p)ˆhi,p(ˆXi,p|θi,p)+Δhi,p] (3.3)

    where Δhi,p=hi,p(Xi,p)hi,p(ˆXi,p).

    By Assumption 1, we can obtain

    ˆhi,p(ˆXi,p|θi,p)=θTi,pΨi,p(ˆXi,p). (3.4)

    The optimal parameters are set as

    θi,p=argminθi,pΩi,p[supˆXi,pUi,p|ˆhi,l(ˆXi,p|θi,p)hi,p(ˆXi,p)|] (3.5)

    where 1pn, Ωi,p and Ui,p are tight regions for θi,p, Xi,p and ˆXi,p.

    Define errors of the optimal approximation ξi,p and parameter estimation ˜θi,p as

    ξi,p=hi,p(ˆXi,p)ˆhi,p(ˆXi,p|θi,p)˜θi,p=θi,pθi,p. (3.6)

    Assumption 2. ([47,48]) The boundedness of the optimal approximation errors is ensured by the existence of positive constants ξi0, such that |ξi,p|ξi0.

    Assumption 3. The set of constants γi satisfies the following relationship

    |hi,p(Xi,p)hi,p(ˆXi,p)|γi,pXi,pˆXi,p. (3.7)

    From Eqs (3.2) and (3.3), we obtain

    ˙ei=Aiei+np=1Bi,p[ξi,p+Δhi,p+˜θTi,pΨi,p(ˆXi,p)]=Aiei+Δhi+ξi+np=1Bi,p[˜θTi,pΨi,p(ˆXi,p)] (3.8)

    where ξi=[ξi,1,...,ξi,n]T, Δhi=[Δhi,1,...,Δhi,n]T.

    Constructing the Lyapunov function:

    V0=Ni=1Vi,0=Ni=112eTiPiei.

    Then, by derivation, we have

    ˙V0Ni=1{12eTi(PiATi+AiPi)ei+eTiPi(ξi+Δhi)+np=1eTiPiBi,p[˜θTi,pΨi,p(ˆXi,p)]}Ni=1{eTiQiei+eTiPi(ξi+Δhi)+eTiPinp=1Bi,p˜θTi,pΨi,p(ˆXi,p)}. (3.9)

    According to Lemma 3 and Assumption 3, we obtain

    eTiPi(ξi+Δhi)|eTiPiξi|+|eTiPiΔhi|ei2+12Piξi2+12Pi2np=1|Δhi,p|2ei2+12ei2Pi2np=1γ2i,p+12Piξi2ei2(1+12Pi2np=1γi,p2)+12Piξi2 (3.10)

    and

    eTiPinp=1Bi,p˜θTi,pΨi,p(ˆXi,p)12eTiPTiPiei+12np=1˜θTi,pΨi,p(ˆXi,p)ΨTi,p(ˆXi,p)˜θi,p12λ2i,max(Pi)ei2+12np=1˜θTi,p˜θi,p. (3.11)

    Due to the positive definite matrix Pi, the maximum eigenvalue is proposed as λi,max(Pi). Based on Eqs (3.9), (3.10) and (3.11), we have

    ˙V0Ni=1(qi,0ei2+12Piξi2+12np=1˜θTi,p˜θi,p) (3.12)

    where 0<Ψi,p()ΨTi,p()1 and qi,0=λi,min(Qi)(1+12Pi2np=1γ2i,p+12λ2i,max(Pi)).

    Then, (3.12) turns to

    ˙V0q0e2+12Pξ2+Ni=1np=112˜θTi,p˜θi,p (3.13)

    where q0=Ni=1qi,0.

    Theorem 1. For systems (2.1) where Assumptions 1–3 hold, combining observer (3.2), virtual control laws (3.36), (3.48) and (3.60), adaptive laws (3.37), (3.49), (3.61), (3.70) and control input (3.69) together, signals xi,1 which converge to the distributed optimization problem's optimal solution x1, remain semi-global uniformly ultimately bounded (SGUUB) in the closed-loop system.

    Define the virtual control laws

    {xi,2=ci,1[2ai,1(xi,1xd1)+2ai,2(xi,1xd2)+jNiaij(xi,1xj,1)]θTi,1Ψi,1(ˆXi,1)xi,3=ci,2si,2(12δ1+52)si,2θTi,2Ψi,2(ˆXi,2)+xi,2vi,2λi,2xi,m+1=ci,msi,m(12δm+52)si,mθTi,mΨi,m(ˆXi,m)+xi,mvi,mλi,m (3.14)

    adaptive laws

    {˙θi,1=σi,1Ψi,1(ˆXi,1)[2ai,1(xi,1xd1)+2ai,2(xi,1xd2)+jNiaij(xi,1xj,1)]ρi,1θi,1˙θi,2=σi,2δ1Ψi,2(ˆXi,2)si,2ρi,2θi,2˙θi,m=σi,mδmΨi,m(ˆXi,m)si,mρi,mθi,m˙θi,n=σi,nδnΨi,n(ˆXi,n)si,nρi,nθi,n (3.15)

    control input

    ui=ci,nsi,n(12δn+32)si,nθTi,nΨi,n(ˆXi,n)+xi,nvi,nλi,n (3.16)

    where ci,1=3+γ 2i,12, c>0, ρ>0, σ>0, δ=1k2i,bs2i, |si|ki,b.

    Proof. Specify the error variables in the following manner:

    si,1=xi,1xi,1si,p=ˆxi,pvi,pwi,p=vi,pxi,pp=2,,n (3.17)

    where si,p represents the tracking error, vi,p is a state variable that can be obtained using a filter with the virtual controller xi,p, wi,p denotes the error between vi,p and xi,p, and ˆxi,p is the estimation of xi,p.

    Through the DSC technique, this paper constructs the following filter

    λi,p˙vi,p+vi,p=xi,p,vi,p(0)=xi,p(0) (3.18)

    where p is the order of the multi-agent model and 2pn. Combining (3.17) with (3.18), we have

    ˙wi,p=˙vi,p˙xi,p=vi,pxi,pλi,p˙xi,p=wi,pλi,p+Bi,p (3.19)

    where λi,p is the parameter we set. Bi,p=˙xi,p. According to [49] and [50], there exist constants Mi,p>0, |Bi,p|Mi,p.

    Step 1. First, the gradient of the penalty function (2.12) is calculated as follows

    P(x1)x1=vec(fi(xi,1(t))xi,1)+Lx1 (3.20)

    where vec(fi(xi,1(t))xi,1) is a column vector. To find the extremum of the penalty function, we need to locate the point where its derivative is zero

    P(x1)x1=0.

    Combining (2.12) and (3.20), we obtain:

    fi(xi,1(t))xi,1+jNiaij(xi,1xj,1)=0. (3.21)

    According to (2.14) and (3.21), we have

    2ai,1(xixd1)+2ai,2(xixd2)+jNiaij(xi,1xj,1)=0. (3.22)

    Then according to (3.17) and (3.22), we have

    P(x1)xi,1=fi(xi,1(t))xi,1+jNiaij(xi,1xj,1)=2ai,1(xi,1xd1)+2ai,2(xi,1xd2)+jNiaij(xi,1xj,1)2ai,1(xixd1)2ai,2(xixd2)+jNiaij(xi,1xj,1)=2aisi,1+jNiaij(si,1sj,1). (3.23)

    Let s1=[s1,1  sN,1]T, A=diag{2ai}, H=A+L. According to (3.23), we have

    P(x1)x1=Hs1.

    Construct the following Lyapunov function:

    V1=V0+12(P(x1)x1)TH1(P(x1)x1)+Ni=11σi,1˜θTi,1˜θi,1=V0+12sT1Hs1+Ni=11σi,1˜θTi,1˜θi,1 (3.24)

    where s1=[s1,1  sN,1]T and σi,1 are the parameter we set. According to (2.1), (3.2) and (3.17), we have

    ˙si,1=ˆxi,2+θTi,1Ψi,1+˜θTi,1Ψi,1+Δhi,1+ξi,1+ei,2. (3.25)

    Then, according to (3.24) and (3.25), we can obtain

    ˙V1=˙V0+sT1H˙s1+Ni=11σi,1˜θTi,1˙˜θi,1=˙V0+sT1H(ˆx2+vec(θTi,1Ψi,1)+vec(˜θTi,1Ψi,1)+Δh1+ξ1+e2)+Ni=11σi,1˜θTi,1˙˜θi,1=˙V0+sT1H(s2+w2+x2+vec(θTi,1Ψi,1)+vec(˜θTi,1Ψi,1)+Δh1+ξ1+e2)+Ni=11σi,1˜θTi,1˙˜θi,1=˙V0+sT1Hs2+sT1Hw2+sT1H(x2+vec(θTi,1Ψi,1)+vec(˜θTi,1Ψi,1))+sT1HΔh1+sT1Hξ1+sT1He2Ni=11σi,1˜θTi,1˙θi,1 (3.26)

    where s2=[s1,2  sN,2]T, w2=[w1,2  wN,2]T, x2=[x1,2  xN,2]T, Δh1=[Δh1,1 Δh2,1  ΔhN,1]T, ξ1=[ξ1,1 ξ2,1  ξN,1]T, e2=[e1,2 e2,2  eN,2]T, vec(θTi,1Ψi,1) and vec(˜θTi,1Ψi,1) are column vectors. According to Lemma 3, we have

    sT1Hs212sT1HHTs1+12sT2s2 (3.27)
    sT1He212sT1HHTs1+12eT2e2 (3.28)
    sT1Hw212sT1HHTs1+12wT2w2 (3.29)
    sT1HΔh112sT1Hγ1γT1HTs1+12eT1e1 (3.30)
    sT1Hξ112sT1HHTs1+12ξT1ξ1 (3.31)

    where γ1=diag[γi,1],e1=[e1,1 e2,1  eN,1]T. Substituting (3.27)–(3.28) into (3.26), ˙V1 turns to

    ˙V1˙V0+sT1H(x2+vec(θTi,1Ψi,1)+vec(˜θTi,1Ψi,1))+12sT1HHTs1+12wT2w2+12sT1HHTs1+12sT2s2+12sT1Hγ1γT1HTs1+12eT1e1+12sT1HHTs1+12ξT1ξ1+12sT1HHTs1+12eT2e2Ni=11σi,1˜θTi,1˙θi,1. (3.32)

    Due to H=A+L, we can obtain

    sT1H=[2a1s1,1+jNia1j(s1,1sj,1), ,2aNsN,1+jNiaNj(sN,1sj,1)]. (3.33)

    Then, we have

    sT1HHTs1=(P(x1)x1)T(P(x1)x1)=Ni=1[2ai,1(xi,1xd1)+2ai,2(xi,1xd2)+jNiaij(si,1sj,1)]2 (3.34)

    and

    sT1Hγ1γT1HTs1=Ni=1γ 2i,1[2ai,1(xi,1xd1)+2ai,2(xi,1xd2)+jNiaij(si,1sj,1)]2. (3.35)

    According to Theorem1,

    xi,2=ci,1[2ai,1(xi,1xd1)+2ai,2(xi,1xd2)+jNiaij(xi,1xj,1)]θTi,1Ψi,1(ˆXi,1) (3.36)
    ˙θi,1=σi,1Ψi,1(ˆXi,1)[2ai,1(xi,1xd1)+2ai,2(xi,1xd2)+jNiaij(xi,1xj,1)]ρi,1θi,1. (3.37)

    Substituting (3.34)–(3.37) into (3.32), after (3.13), we have

    ˙V1q0e2+12Pξ2+Ni=1np=112˜θTi,p˜θi,p+12eT2e2+12eT1e1+12ξT1ξ1+Ni=1ρi,1σi,1˜θTi,1θi,1+12sT2s2+12wT2w2(P(x1)x1)T(P(x1)x1)q1e2+η1+Ni=1np=112˜θTi,p˜θi,p+Ni=1ρi,1σi,1˜θTi,1θi,1+Ni=112s2i,2+Ni=112w2i,22λmax(H1)(P(x1)x1)TH1(P(x1)x1) (3.38)

    where q1=q0N, η1=12Pξ2+12ξT1ξ1, λmax(H1) is the maximum eigenvalue of the positive matrix H1.

    Step 2. In accordance with (3.17), design si,2=ˆxi,2vi,2. By (3.2) and (3.4), we obtain

    ˙si,2=˙ˆxi,2˙vi,2=si,3+wi,3+xi,3+εi,2ei,1+˜θTi,2Ψi,2+θTi,2Ψi,2+ξi,2+Δhi,2˙vi,2. (3.39)

    Starting from this step, according to Lemma 1, we add the state constraints condition and define δ1=1k2i,b1s2i,2. Construct the BLF

    V2=V1+Ni=1Vi,2=V1+12Ni=1{δ1s2i,2+1σi,2˜θTi,2˜θi,2+w2i,2} (3.40)

    where ki,b1 and σi,2 are parameters we set.

    Then, we have

    ˙V2=˙V1+Ni=1{δ1si,2˙si,2+1σi,2˜θTi,2˙˜θi,2+wi,2˙wi,2}. (3.41)

    Substituting (3.39) into (3.41), ˙V2 turns to

    ˙V2=˙V1+Ni=1[δ1si,2(wi,3+si,3+xi,3+εi,2ei,1+θTi,2Ψi,2+˜θTi,2Ψi,2+ξi,2+Δhi,2˙vi,2)+1σi,2˜θTi,2˙˜θi,2+wi,2˙wi,2]. (3.42)

    According to Lemma 3, we obtain

    si,2εi,2ei,112s2i,2+12ε2i,2ei,12 (3.43)
    si,2si,3+si,2wi,3s2i,2+12(s2i,3+w2i,3) (3.44)
    si,2ξi,212s2i,2+12ξi,22 (3.45)
    si,2Δhi,212s2i,2+12γ2i,2ei,22. (3.46)

    Substituting (3.43)–(3.46) into (3.42), we have

    ˙V2˙V1+Ni=1[δ1si,2(xi,3+θTi,2Ψi,2+˜θTi,32Ψi,2˙vi,2)+5δ12s2i,2+12(s2i,3+w2i,3)+12ε2i,2ei,12+12ξi,22+12γ2i,2ei,221σi,2˜θTi,2˙θi,2+wi,2˙wi,2]. (3.47)

    According to Theorem1,

    xi,3=ci,2si,2(12δ1+52)si,2θTi,2Ψi,2(ˆXi,2)+xi,2vi,2λi,2 (3.48)
    ˙θi,2=σi,2δ1Ψi,2(ˆXi,2)si,2ρi,2θi,2. (3.49)

    Substitute (3.19), (3.48) and (3.49) into (3.47). According to Lemma 3, we have wi,2Bi,212w2i,2+12M2i,2; after (3.38), we can obtain

    ˙V2q2e2+η2+Ni=1np=112˜θTi,p˜θi,p+Ni=1ρi,1σi,1˜θTi,1θi,1+Ni=1ρi,2σi,2˜θTi,2θi,2Ni=1ci,2s2i,2Ni=1(1λi,21)w2i,22λmax(H1)(P(x1)x1)TH1(P(x1)x1)+Ni=1[12M2i,2+12(s2i,3+w2i,3)] (3.50)

    where q2=q112Ni=1(ε2i,2+γ2i,2),η2=η1+12Ni=1ξi,22.

    Step m. Design si,m=ˆxi,mvi,m, by (3.2) and (3.4), we obtain

    ˙si,m=ˆxi,m+1+εi,mei,1+θTi,mΨi,m+˜θTi,mΨi,m+ξi,m+Δhi,m˙vi,m. (3.51)

    According to Lemma 1, define δm=1k2i,bms2i,m. Put forward the BLF

    Vm=Vm1+12Ni=1{δms2i,m+1σi,m˜θTi,m˜θi,m+w2i,m} (3.52)

    where ki,bm and σi,m are designed parameters.

    After derivation,

    ˙Vm=Ni=1{δmsi,m˙si,m+1σi,m˜θTi,m˙˜θi,m+wi,m˙wi,m}+˙Vm1. (3.53)

    Substituting (3.51) into (3.53), refer to (3.17), we have

    ˙Vm=˙Vm1+Ni=1[δmsi,m(si,m+1+wi,m+1+xi,m+1+εi,mei,1+θTi,mΨi,m+˜θTi,mΨi,m+ξi,m+Δhi,m˙vi,m)+1σi,m˜θTi,m˙˜θi,m+wi,m˙wi,m]. (3.54)

    According to Lemma 3, we obtain

    si,mεi,mei,112s2i,m+12ε2i,mei,12 (3.55)
    si,msi,m+1+si,mwi,m+1s2i,m+12(s2i,m+1+w2i,m+1) (3.56)
    si,mξi,m12s2i,m+12ξi,m2 (3.57)
    si,mΔhi,m12s2i,m+12γ2i,mei,m2. (3.58)

    Then, ˙Vm turns to

    ˙Vm˙Vm1+Ni=1[δmsi,m(xi,m+1+θTi,mΨi,m+˜θTi,mΨi,m˙vi,m)+5δm2s2i,m+12(s2i,m+1+w2i,m+1)+12ε2i,mei,12+12ξi,m2+12γ2i,mei,m21σi,m˜θTi,m˙θi,m+wi,m˙wi,m]. (3.59)

    According to Theorem1,

    xi,m+1=ci,msi,m(12δm+52)si,mθTi,mΨi,m(ˆXi,m)+xi,mvi,mλi,m (3.60)
    ˙θi,m=σi,mδmΨi,m(ˆXi,m)si,mρi,mθi,m. (3.61)

    According to Eqs (3.60), (3.61) and (3.19), by Lemma 3, we have wi,mBi,m12w2i,m+12M2i,m. (3.59) can be rewritten as

    ˙Vm˙Vm1+Ni=1[δmsi,m(ci,msi,m(12δm+52)si,mθTi,mΨi,m(ˆXi,m)+xi,mvi,mλi,m+θTi,mΨi,m+˜θTi,mΨi,m˙vi,m)+5δm2s2i,m+12(s2i,m+1+w2i,m+1)+12ε2i,mei,12+12ξi,m2+12γ2i,mei,m21σi,m˜θTi,m(σi,mΨi,msi,mρi,mθi,m)w2i,mλi,m+12w2i,m+12M2i,m]. (3.62)

    Combining (3.13), (3.38) and (3.50), we have

    ˙Vmqme2+ηm+Ni=1np=112˜θTi,p˜θi,p2λmax(H1)(P(x1)x1)TH1(P(x1)x1)+Ni=1[mp=1ρi,pσi,p˜θTi,pθi,pmp=2ci,ps2i,pmp=2(1λi,p1)w2i,p+12mp=2M2i,m+12(s2i,m+1+w2i,m+1)] (3.63)

    where qm=qm112Ni=1(ε2i,m+γ2i,m),ηm=ηm1+12Ni=1ξi,m2.

    Step n. Design si,n=ˆxi,nvi,n, by (3.2) and (3.4), we obtain

    ˙si,n=˙ˆxi,n˙vi,n=ui+εi,nei,1+θTi,nΨi,n+˜θTi,nΨi,n+ξi,n+Δhi,n˙vi,n. (3.64)

    According to Lemma 1, define δn=1k2i,bns2i,n. Put forward the BLF,

    Vn=Vn1+12Ni=1{δns2i,n+1σi,n˜θTi,n˜θi,n+w2i,n} (3.65)

    where ki,bn and σi,n are the parameter we set.

    Then, we have

    ˙Vn=˙Vn1+Ni=1{δnsi,n˙si,n+1σi,n˜θTi,n˙˜θi,n+wi,n˙wi,n}. (3.66)

    Substituting (3.64) into (3.66), we obtain

    ˙Vn=˙Vn1+Ni=1[δnsi,n(ui+εi,mei,1+θTi,nΨi,n+˜θTi,nΨi,n+ξi,n+Δhi,n˙vi,n)+1σi,n˜θTi,n˙˜θi,n+wi,n˙wi,n]. (3.67)

    According to Lemma 3 and the derivation principle of the previous steps, Eq (3.67) is formulated as

    ˙Vn˙Vn1+Ni=1[δnsi,n(ui+θTi,nΨi,n+˜θTi,nΨi,n˙vi,n)+3δn2s2i,n+12ε2i,nei,12+12ξi,n2+12γ2i,nei,n21σi,n˜θTi,n˙θi,n+wi,n˙wi,n]. (3.68)

    According to Theorem1,

    ui=ci,nsi,n(12δn+32)si,nθTi,nΨi,n(ˆXi,n)+xi,nvi,nλi,n (3.69)
    ˙θi,n=σi,nδnΨi,n(ˆXi,n)si,nρi,nθi,n. (3.70)

    Substitute Eqs (3.69), (3.70) and (3.19) into (3.68). By Lemma 3, we have wi,nBi,n12w2i,n+12M2i,n, then we can obtain

    ˙Vnqne2+ηn+Ni=1np=112˜θTi,p˜θi,p2λmax(H1)(P(x1)x1)TH1(P(x1)x1)+Ni=1[np=1ρi,pσi,p˜θTi,pθi,pnp=2ci,ps2i,pnp=2(1λi,p1)w2i,p+12np=2M2i,p] (3.71)

    where qn=qn112Ni=1(ε2i,n+γ2i,n),ηn=ηn1+12Ni=1ξi,n2.

    Through Lemma 3, we have

    ˜θT,pθ,p12˜θT,p˜θ,p+12θT,pθ,p. (3.72)

    Define

    ζ=ηn+Ni=1np=1ρi,p2σi,pθTi,pθi,p+12np=2M2i,p. (3.73)

    Thus, we rewrite (3.71) as follows

    ˙Vnqne22λmax(H1)(P(x1)x1)TH1(P(x1)x1)+Ni=1[np=2ci,ps2i,pnp=1(ρi,p2σi,p12)˜θTi,p˜θi,pnp=2(1λi,p1)w2i,p]+ζ (3.74)

    where ci,p>0,(ρi,p2σi,p12)>0,(1λi,p1)>0,(p=2,,n),2λmax(H1)>0.

    Define

    C=min{2qnλmin(P),2ci,p,2(ρi,p2σi,p12),2(1λi,p1),4λmin(H1)}. (3.75)

    Thus, we obtain

    ˙VnCV(x(t))+ζ. (3.76)

    According to the study conducted by [51], it is easily verified that in the closed-loop system, all of the signals from system (2.1) stay SGUUB.

    Remark 2. As is shown in Theorem1, compared to the papers that study the bipartite consensus control of MASs [38,39], Eq (3.22) introduces two reference signals xd1 and xd2 in the design of a distributed optimization controller. Serving as virtual leaders, xd1 and xd2 achieve a bipartite containment effect, which aims to ensure that the trajectory of the optimal solution in the distributed optimization lies within the convex hulls defined by the upper and lower reference signals. These signals generate upper and lower convex hulls due to the negative signals emitted by agents.

    Remark 3. For the distributed optimization problem of bipartite containment control in high-order nonlinear MASs with state constraints, by integrating RBFNN and DSC techniques, a BLF-based adaptive backstepping controller is designed. Moreover, a more concise optimization algorithm based on penalty functions is presented to minimize the global objective function and achieve optimal output.

    In this session, we use two examples to show how successful the control mechanism is. The results prove that the distributed optimal control algorithm in this paper has practical applications.

    For a second-order system, the model is as follows

    {˙xi,1=xi,2+hi,1(xi,1)˙xi,2=ui+hi,2(xi,1,xi,2)yi=xi,1. (4.1)

    The unknown functions in system (4.1) are

    hi,1=0h1,2=x1,10.25x1,2x31,1+0.3cos(t)h2,2=x2,10.25x2,2x32,1+0.1(x22,1+x22,2)1/122+0.3cos(t)h3,2=x3,10.25x3,2x33,1+0.1sin(t)(x23,1+2x23,2)1/122+0.3cos(t)h4,2=x4,10.25x4,2x34,1+0.1sin(t)(2x24,1+2x24,2)1/122+0.3cos(t)h5,2=x5,10.1x5,2x35,1+0.2sin(t)(x25,1+x25,2)1/122+0.3cos(t).

    The topology of Figure 1 shows the information exchange between five agents, with agent 4 receiving the opposite information. The initial states of five agents are set as x1(0)=[0.05,0.05], x2(0)=[0.1,0.1], x3(0)=[0.15,0.15], x4(0)=[0.2,0.2], x5(0)=[0.25,0.25]. Define xd1=0.2sin(t)+0.2, xd2=0.2sin(t)+0.4 as the reference signals. Thus, the trajectory of the optimal signal is x1=0.2sin(t)+0.3. The local objective functions of agents are given as follows

    f1=8.5x21(8xd1+9xd2)x1+4xd12+4.5xd22+1f2=16.5x22(16xd1+17xd2)x2+8xd12+8.5xd22+2f3=13x23(12xd1+14xd2)x3+12xd12+14xd22+1f4=15.2x24+(14.4xd3+16xd4)x4+7.2xd32+8xd42+2f5=9.5x25(9xd1+10xd2)x5+4.5xd12+5xd22+2.
    Figure 1.  Communication graph in simulation.

    Regarding the observer, the parameters are chosen as εi,1=500, εi,2=5,000 and the initial states are selected as ˆx1=[0.05,0.05], ˆx2=[0.1,0.1], ˆx3=[0.15,0.15], ˆx4=[0.2,0.2], ˆx5=[0.25,0.25].

    According to the virtual control law xi,2 (3.36), adaptive law θi,1 (3.37), θi,n (3.70) and control input ui (3.69) in Theorem 3.1, we chose control parameters as} ci,1=3.5, ci,2=30, σi,1=σi,2=1, ρi,1=ρi,2=80, λi,2=0.05. To guarantee the state constraints are not violated, kb1=0.3 is given.

    In this simulation, Figures 27 display the simulation outcomes. Figure 2 describes the trajectories of x1 and xi,1, verifying that each agent can track the optimal signal x1. Besides, xd1 and xd2 act as virtual leaders to achieve a bipartite containment effect. In Figure 3, we use xi,1 as an example via comparing the true value and the estimated value. It can be seen that the designed observer performs well and can approximately observe the unknown states. Figure 4 shows the trajectories of si,1, which clearly demonstrates how the tracking error can rapidly approach zero by the designed controller. Together, Figures 2 and 4 illustrate that there is a good distributed optimization consensus tracking effect and the tracking error is within 0.05. Figure 5 displays the trajectories of xi,2, the state in this paper is constrained. Figure 6 gives the trajectories of control input ui. Figure 7 shows that si,2 are all in the range of 0.3 to 0.3, satisfying |si,2|kb1. From Figures 5 and 7, we can draw the conclusion that based on the BLF, si,2 and xi,2 can be limited successfully under the designed parameters. The state constraints are not violated, and the tracking error can converge to the compact sets.

    Figure 2.  Tracking performance of xi,1(i=1,,5).
    Figure 3.  Trajectories of xi,1(i=1,,5) estimation.
    Figure 4.  The curves of error si,1(i=1,,5).
    Figure 5.  Constraint trajectories of xi,2(i=1,,5).
    Figure 6.  Control input ui.
    Figure 7.  Constraint error si,2(i=1,,5).

    Remark 4. Compared to the simulation example in our previous work [37], under the unchanged initial conditions and parameters, Figure 8 is the trajectory tracking graph of all agents, which confirms that the agent outputs converge inside both the positive and negative convex hulls defined by the reference leader signals. Based on this, the distributed optimization problem is considered in our paper. We design a distributed optimization algorithm based on a penalty function. It can be clearly seen that the control protocol proposed in this paper enables all agents to track the optimal solution. The controller we designed can ensure that all agents have a good distributed optimization consensus tracking effect for high-order nonlinear uncertain MASs.

    Figure 8.  Trajectories of xi,1(i=1,,4) estimation in the previous work.

    Given a single-link manipulator that includes motor dynamics [52], the dynamic equation of the system is as follows:

    {Di¨qi+Bi˙qi+Nisinqi=τiMi˙τi+Hiτi=uiKi˙qi (4.2)

    where qi, ˙q and ¨q represent the link position, velocity and acceleration for the ith mechanical system. τi is the torque produced by the electrical subsystem, respectively. Di=1kgm2 is the inertia, Bi=1Nms/rad represents the viscous friction coefficient at the joint and Ni=2 is a constant that is related to the coefficient of gravity and the mass of the load. Mi=1H denotes the armature inductance. Hi=1Ω represents armature resistance. Ki=2Nm/A signifies the back electromotive force coefficient. ui is the input signal. Letting xi,1=qi, xi,2=˙qi, xi,3=τi, system (4.2) can be expressed as:

    {˙xi,1=xi,2˙xi,2=xi,3Nisinxi,1Bixi,2˙xi,3=uiKixi,2Hixi,3. (4.3)

    Consider the topology of Figure 1 including five agents, whose initial states are set as x1(0)=[0.3,0.3,0.3], x2(0)=[0.15,0.15,0.15], x3(0)=[0.15,0.15,0.15], x4(0)=[0.2,0.2,0.2], x5(0)=[0.25,0.25,0.25]. Define xd1=0.2sin(t)+0.2, xd2=0.2sin(t)+0.4 as the reference signals. The optimal signal is x1=0.2sin(t)+0.3.

    The local objective functions of agents are given as follows

    f1=8.5x21(8xd1+9xd2)x1+4xd12+4.5xd22+1f2=13.5x22(12xd1+15xd2)x2+6xd12+7.5xd22+2f3=13x23(12xd1+14xd2)x3+6xd12+7xd22+1f4=14.2x24+(14xd3+14.4xd4)x4+7xd32+7.2xd42+2f5=9.5x25(9xd1+10xd2)x5+4.5xd12+5xd22+2.

    According to the virtual control laws xi,2 (3.36), xi,3 (3.48), adaptive laws θi,1 (3.37), θi,2 (3.49), θi,n (3.70)} and control input ui (3.69) in Theorem 3.1, we chose control parameters as ci,1=3.5, ci,2=2, ci,3=60, σi,1=σi,2=σi,3=1, ρi,1=ρi,2=80, ρi,3=8 λi,2=λi,3=0.01. To guarantee the state constraints are not violated, kb1=4 and kb2=4 are given.

    In this simulation, Figures 913 display the simulation outcomes. Figure 9 shows the trajectories of xd and xi,1, each agent can track the optimal signal xi,1. Meanwhile, xd1 and xd2 act as virtual leaders to achieve a bipartite containment} effect. Figure 10 shows the trajectories of si,1 which clearly demonstrates how the tracking error can rapidly approach to near zero, proving that there is a good distributed optimization consensus tracking effect and the tracking error is within 0.05. Figure 11 gives the trajectories of the control input ui. The common barrier Lyapunov method mainly forms the error constraint, Figures 12 and 13 show that si,2 and si,3 are all in the range of –4 to 4, respectively, which indicates that the BLF can solve the state constraints problem of high-order nonlinear uncertain MASs well.

    Figure 9.  Tracking performance of xi,1(i=1,,5).
    Figure 10.  The curves of error si,1(i=1,,5).
    Figure 11.  Control input ui.
    Figure 12.  Constraint error si,2(i=1,,5).
    Figure 13.  Constraint error si,3(i=1,,5).

    According to the simulation results, the distributed optimization controller proposed in this research can make agents converge to the optimal solution, which lies within the convex hull of the trajectories of each virtual leader and their opposite trajectories. Meanwhile, the tracking error gradually converges to a small range of the origin, which indicates the control performance is good. In addition, the barrier Lyapunov method is generally used to impose constraints on errors. From step 2, the BLF-based control scheme transforms the original state constraints into a new bound on the tracking error and achieves the state constraints by constraining the error surfaces.

    This paper researches the bipartite containment distributed optimization problem of high-order MASs with uncertain nonlinear functions. Combining the consensus condition of MASs with the global objective function, we designed a penalty function, which is constructed by combining the bipartite containment definition. By integrating RBFNN and DSC techniques, an adaptive backstepping controller is proposed to avoid the complexity explosion and a distributed optimal consensus is accurately achieved. In addition, introducing state constraints in the distributed optimization holds significant practical importance in engineering applications. Besides, we will consider addressing the problem of prescribed-time distributed optimization of high-order nonlinear stochastic MASs with disturbances.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    This work was supported in part by the National Natural Science Foundation of China under Grant Number:5217052158.

    The authors declare there is no conflict of interest.



    [1] T. Yang, X. Yi, J. Wu, Y. Yuan, D. Wu, Z. Meng, et al., A survey of distributed optimization, Annu. Rev. Control, 47 (2019), 278–305. https://doi.org/10.1016/j.arcontrol.2019.05.006 doi: 10.1016/j.arcontrol.2019.05.006
    [2] X. Li, L. Xie, Y. Hong, Distributed aggregative optimization over multi-agent networks, IEEE Trans. Autom. Control, 67 (2021), 3165–3171. https://doi.org/10.1109/TAC.2021.3095456 doi: 10.1109/TAC.2021.3095456
    [3] R. Yang, L. Liu, G. Feng, An overview of recent advances in distributed coordination of multi-agent systems, Unmanned Syst., 10 (2022), 307–325. https://doi.org/10.1142/S2301385021500199 doi: 10.1142/S2301385021500199
    [4] Y. Li, C. Tan, A survey of the consensus for multi-agent systems, Syst. Sci. Control. Eng., 7 (2019), 468–482. https://doi.org/10.1080/21642583.2019.1695689 doi: 10.1080/21642583.2019.1695689
    [5] C. Zhao, X. Duan, Y. Shi, Analysis of consensus-based economic dispatch algorithm under time delays, IEEE Trans. Syst. Man Cybern., 50 (2018), 2978–2988. https://doi.org/10.1109/TSMC.2018.2840821 doi: 10.1109/TSMC.2018.2840821
    [6] Y. Wan, J. Qin, Q. Ma, W. Fu, S. Wang, Multi-agent drl-based data-driven approach for pevs charging/discharging scheduling in smart grid, J. Franklin Inst., 359 (2022), 1747–1767. https://doi.org/10.1016/j.jfranklin.2022.01.016 doi: 10.1016/j.jfranklin.2022.01.016
    [7] X. Zeng, P. Yi, Y. Hong, Distributed continuous-time algorithm for robust resource allocation problems using output feedback, in 2017 American Control Conference (ACC), (2017), 4643–4648. https://doi.org/10.23919/ACC.2017.7963672
    [8] K. I. Tsianos, S. Lawlor, M. G. Rabbat, Consensus-based distributed optimization: Practical issues and applications in large-scale machine learning, in 2012 50th Annual Allerton Conference on Communication, Control, and Computing (Allerton), (2012), 1543–1550. https://doi.org/10.1109/Allerton.2012.6483403
    [9] T. M. D. Tran, A. Y. Kibangou, Distributed estimation of laplacian eigenvalues via constrained consensus optimization problems, Syst. Control. Lett., 80 (2015), 56–62. https://doi.org/10.1016/j.sysconle.2015.04.001 doi: 10.1016/j.sysconle.2015.04.001
    [10] Q. Lü, H. Li, Event-triggered discrete-time distributed consensus optimization over time-varying graphs, Complexity, 2017 (2017), 1–12. https://doi.org/10.1155/2017/5385708 doi: 10.1155/2017/5385708
    [11] X. Shi, J. Cao, W. Huang, Distributed parametric consensus optimization with an application to model predictive consensus problem, IEEE Trans. Cybern., 48 (2017), 2024–2035. https://doi.org/10.1109/TCYB.2017.2726102 doi: 10.1109/TCYB.2017.2726102
    [12] G. Wang, Distributed control of higher-order nonlinear multi-agent systems with unknown non-identical control directions under general directed graphs, Automatica, 110 (2019), 108559. https://doi.org/10.1016/j.automatica.2019.108559 doi: 10.1016/j.automatica.2019.108559
    [13] T. Guo, J. Han, C. Zhou, J. Zhou, Multi-leader-follower group consensus of stochastic time-delay multi-agent systems subject to markov switching topology, Math. Biosci. Eng., 19 (2022), 7504–7520. https://doi.org/10.3934/mbe.2022353 doi: 10.3934/mbe.2022353
    [14] C. Sun, M. Ye, G. Hu, Distributed time-varying quadratic optimization for multiple agents under undirected graphs, IEEE Trans. Autom. Control, 62 (2017), 3687–3694. https://doi.org/10.1109/TAC.2017.2673240 doi: 10.1109/TAC.2017.2673240
    [15] Z. Li, Z. Ding, J. Sun, Z. Li, Distributed adaptive convex optimization on directed graphs via continuous-time algorithms, IEEE Trans. Autom. Control, 63 (2017), 1434–1441. https://doi.org/10.1109/TAC.2017.2750103 doi: 10.1109/TAC.2017.2750103
    [16] S. Yang, Q. Liu, J. Wang, A multi-agent system with a proportional-integral protocol for distributed constrained optimization, IEEE Trans. Autom. Control, 62 (2016), 3461–3467. https://doi.org/10.1109/TAC.2016.2610945 doi: 10.1109/TAC.2016.2610945
    [17] M. Hong, M. Razaviyayn, J. Lee, Gradient primal-dual algorithm converges to second-order stationary solution for nonconvex distributed optimization over networks, in Proceedings of the 35th International Conference on Machine Learning, (2018), 2009–2018.
    [18] D. Jakovetić, N. Krejić, N. K. Jerinkić, A hessian inversion-free exact second order method for distributed consensus optimization, IEEE Trans. Signal Inf. Process., 8 (2022), 755–770. https://doi.org/10.1109/TSIPN.2022.3203860 doi: 10.1109/TSIPN.2022.3203860
    [19] Q. Liu, J. Wang, A second-order multi-agent network for bound-constrained distributed optimization, IEEE Trans. Autom. Control, 60 (2015), 3310–3315. https://doi.org/10.1109/TAC.2015.2416927 doi: 10.1109/TAC.2015.2416927
    [20] P. Lin, W. Ren, J. A. Farrell, Distributed continuous-time optimization: nonuniform gradient gains, finite-time convergence, and convex constraint set, IEEE Trans. Autom. Control, 62 (2016), 2239–2253. https://doi.org/10.1109/TAC.2016.2604324 doi: 10.1109/TAC.2016.2604324
    [21] J. Lu, C. Y. Tang, Zero-gradient-sum algorithms for distributed convex optimization: The continuous-time case, IEEE Trans. Autom. Control, 57 (2012), 2348–2354. https://doi.org/10.1109/TAC.2012.2184199 doi: 10.1109/TAC.2012.2184199
    [22] S. Rahili, W. Ren, Distributed continuous-time convex optimization with time-varying cost functions, IEEE Trans. Autom. Control, 62 (2016), 1590–1605. https://doi.org/10.1109/TAC.2016.2593899 doi: 10.1109/TAC.2016.2593899
    [23] Z. Feng, G. Hu, C. G. Cassandras, Finite-time distributed convex optimization for continuous-time multiagent systems with disturbance rejection, IEEE Trans. Control. Netw. Syst., 7 (2019), 686–698. https://doi.org/10.1109/TCNS.2019.2939642 doi: 10.1109/TCNS.2019.2939642
    [24] Y. Tang, K. Zhu, Optimal consensus for uncertain high-order multi-agent systems by output feedback, Int. J. Robust Nonlin. Control, 32 (2022), 2084–2099. https://doi.org/10.1002/rnc.5928 doi: 10.1002/rnc.5928
    [25] X. Wang, G. Wang, S. Li, Distributed finite-time optimization for integrator chain multiagent systems with disturbances, IEEE Trans. Autom. Control, 65 (2020), 5296–5311. https://doi.org/10.1109/TAC.2020.2979274 doi: 10.1109/TAC.2020.2979274
    [26] G. Li, X. Wang, S. Li, Finite-time distributed approximate optimization algorithms of higher order multiagent systems via penalty-function-based method, IEEE Trans. Syst. Man Cybern., 52 (2022), 6174–6182. https://doi.org/10.1109/TSMC.2021.3138109 doi: 10.1109/TSMC.2021.3138109
    [27] L. Wang, J. Dong, C. Xi, Event-triggered adaptive consensus for fuzzy output-constrained multi-agent systems with observers, J. Franklin Inst., 357 (2020), 82–105. https://doi.org/10.1016/j.jfranklin.2019.09.033 doi: 10.1016/j.jfranklin.2019.09.033
    [28] M. Shahvali, M. B. Naghibi-Sistani, J. Askari, Distributed adaptive dynamic event-based consensus control for nonlinear uncertain multi-agent systems, Proc. Inst. Mech. Eng. Part I: J. Syst. Control Eng., 236 (2022), 1630–1648. https://doi.org/10.1177/09596518221105669 doi: 10.1177/09596518221105669
    [29] Y. Wu, T. Xu, H. Fang, Command filtered adaptive neural tracking control of uncertain nonlinear time-delay systems with asymmetric time-varying full state constraints and actuator saturation, Proc. Inst. Mech. Eng. Part I: J. Syst. Control Eng., 235 (2021), 1139–1153. https://doi.org/10.1177/0959651820975265 doi: 10.1177/0959651820975265
    [30] B. Beigzadehnoe, Z. Rahmani, A. Khosravi, B. Rezaie, Control of interconnected systems with sensor delay based on decentralized adaptive neural dynamic surface method, Proc. Inst. Mech. Eng. Part I: J. Syst. Control Eng., 235 (2021), 751–768. https://doi.org/10.1177/0959651820966529 doi: 10.1177/0959651820966529
    [31] N. Zhang, J. Xia, T. Liu, C. Yan, X. Wang, Dynamic event-triggered adaptive finite-time consensus control for multi-agent systems with time-varying actuator faults, Math. Biosci. Eng., 20 (2023), 7761–7783. https://doi.org/10.3934/mbe.2023335 doi: 10.3934/mbe.2023335
    [32] K. P. Tee, S. S. Ge, E. H. Tay, Barrier lyapunov functions for the control of output-constrained nonlinear systems, Automatica, 45 (2009), 918–927. https://doi.org/10.1016/j.automatica.2008.11.017 doi: 10.1016/j.automatica.2008.11.017
    [33] K. Zhao, Y. Song, T. Ma, L. He, Prescribed performance control of uncertain euler–lagrange systems subject to full-state constraints, IEEE Trans. Neural Netw. Learn. Syst., 29 (2017), 3478–3489. https://doi.org/10.1109/TNNLS.2017.2727223 doi: 10.1109/TNNLS.2017.2727223
    [34] L. Chen, Asymmetric prescribed performance-barrier lyapunov function for the adaptive dynamic surface control of unknown pure-feedback nonlinear switched systems with output constraints, Int. J. Adapt. Control Signal Process., 32 (2018), 1417–1439. https://doi.org/10.1002/acs.2921 doi: 10.1002/acs.2921
    [35] J. Ni, P. Shi, Adaptive neural network fixed-time leader–follower consensus for multiagent systems with constraints and disturbances, IEEE Trans. Cybern., 51 (2020), 1835–1848. https://doi.org/10.1109/TCYB.2020.2967995 doi: 10.1109/TCYB.2020.2967995
    [36] M. Zamanian, F. Abdollahi, S. K. Yadavar Nikravesh, Finite-time consensus of heterogeneous unknown nonlinear multi-agent systems with external disturbances via event-triggered control, J. Vib. Control, 27 (2021), 1806–1823. https://doi.org/10.1177/1077546320948347 doi: 10.1177/1077546320948347
    [37] J. Yuan, T. Chen, Observer-based adaptive neural network dynamic surface bipartite containment control for switched fractional order multi-agent systems, Int. J. Adapt. Control Signal Process., 36 (2022), 1619–1646. https://doi.org/10.1002/acs.3413 doi: 10.1002/acs.3413
    [38] T. Han, W. X. Zheng, Bipartite output consensus for heterogeneous multi-agent systems via output regulation approach, IEEE Trans. Circuits-II, 68 (2020), 281–285. https://doi.org/10.1109/TCSII.2020.2993057 doi: 10.1109/TCSII.2020.2993057
    [39] T. Han, Z. H. Guan, B. Xiao, H. Yan, Bipartite average tracking for multi-agent systems with disturbances: Finite-time and fixed-time convergence, IEEE Trans. Circuits Syst. I: Regular Papers, 68 (2021), 4393–4402. https://doi.org/10.1109/TCSI.2021.3104933 doi: 10.1109/TCSI.2021.3104933
    [40] Q. Ma, Q. Meng, S. Xu, Distributed optimization for uncertain high-order nonlinear multiagent systems via dynamic gain approach, IEEE Trans. Syst. Man Cybern., 53 (2023), 4351–4357. https://doi.org/10.1109/TSMC.2023.3247456 doi: 10.1109/TSMC.2023.3247456
    [41] X. He, T. Huang, J. Yu, C. Li, Y. Zhang, A continuous-time algorithm for distributed optimization based on multiagent networks, IEEE Trans. Syst. Man Cybern., 49 (2019), 2700–2709. https://doi.org/10.1109/TSMC.2017.2780194 doi: 10.1109/TSMC.2017.2780194
    [42] F. Shojaei, M. M. Arefi, A. Khayatian, H. R. Karimi, Observer-based fuzzy adaptive dynamic surface control of uncertain nonstrict feedback systems with unknown control direction and unknown dead-zone, IEEE Trans. Syst. Man Cybern., 49 (2018), 2340–2351. https://doi.org/10.1109/TSMC.2018.2852725 doi: 10.1109/TSMC.2018.2852725
    [43] X. M. Sun, W. Wang, Integral input-to-state stability for hybrid delayed systems with unstable continuous dynamics, Automatica, 48 (2012), 2359–2364. https://doi.org/10.1016/j.automatica.2012.06.056 doi: 10.1016/j.automatica.2012.06.056
    [44] B. Gharesifard, J. Cortés, Distributed continuous-time convex optimization on weight-balanced digraphs, IEEE Trans. Autom. Control, 59 (2013), 781–786. https://doi.org/10.1109/TAC.2013.2278132 doi: 10.1109/TAC.2013.2278132
    [45] Y. Liu, Q. Zhu, N. Zhao, L. Wang, Adaptive fuzzy backstepping control for nonstrict feedback nonlinear systems with time-varying state constraints and backlash-like hysteresis, Inf. Sci., 574 (2021), 606–624. https://doi.org/10.1016/j.ins.2021.07.068 doi: 10.1016/j.ins.2021.07.068
    [46] Z. Li, Z. Duan, Cooperative Control of Multi-Agent Systems: A Consensus Region Approach, CRC press, Florida, 2017. https://doi.org/10.1201/b17571
    [47] B. Chen, X. Liu, K. Liu, C. Lin, Direct adaptive fuzzy control of nonlinear strict-feedback systems, Automatica, 45 (2009), 1530–1535. https://doi.org/10.1016/j.automatica.2009.02.025 doi: 10.1016/j.automatica.2009.02.025
    [48] K. Li, Y. Li, Adaptive neural network finite-time dynamic surface control for nonlinear systems, IEEE Trans. Neural Netw. Learn. Syst., 32 (2020), 5688–5697. https://doi.org/10.1109/TNNLS.2020.3027335 doi: 10.1109/TNNLS.2020.3027335
    [49] D. Wang, J. Huang, Neural network-based adaptive dynamic surface control for a class of uncertain nonlinear systems in strict-feedback form, IEEE Trans. Neural Netw., 16 (2005), 195–202. https://doi.org/10.1109/TNN.2004.839354 doi: 10.1109/TNN.2004.839354
    [50] J. Yu, P. Shi, W. Dong, B. Chen, C. Lin, Neural network-based adaptive dynamic surface control for permanent magnet synchronous motors, IEEE Trans. Neural Netw. Learn. Syst., 26 (2014), 640–645. https://doi.org/10.1109/TNNLS.2014.2316289 doi: 10.1109/TNNLS.2014.2316289
    [51] X. Zhao, X. Wang, S. Zhang, G. Zong, Adaptive neural backstepping control design for a class of nonsmooth nonlinear systems, IEEE Trans. Syst. Man Cybern., 49 (2018), 1820–1831. https://doi.org/10.1109/TSMC.2018.2875947 doi: 10.1109/TSMC.2018.2875947
    [52] Y. F. Gao, X. M. Sun, C. Wen, W. Wang, Adaptive tracking control for a class of stochastic uncertain nonlinear systems with input saturation, IEEE Trans. Autom. Control, 62 (2016), 2498–2504. https://doi.org/10.1109/TAC.2016.2600340 doi: 10.1109/TAC.2016.2600340
  • This article has been cited by:

    1. Hui Yang, Qing Sun, Jiaxin Yuan, Distributed Adaptive Optimization Algorithm for High-Order Nonlinear Multi-Agent Stochastic Systems with Lévy Noise, 2024, 26, 1099-4300, 834, 10.3390/e26100834
    2. Yang Zhang, Liang Zhao, Jyotindra Narayan, Adaptive control and state error prediction of flexible manipulators using radial basis function neural network and dynamic surface control method, 2025, 20, 1932-6203, e0318601, 10.1371/journal.pone.0318601
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1448) PDF downloads(175) Cited by(2)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog