Loading [MathJax]/jax/element/mml/optable/BasicLatin.js
Research article Special Issues

A novel complex network based dynamic rule selection approach for open shop scheduling problem with release dates

  • In the open shop scheduling problem, resources and tasks are required to be allocated in an optimized manner, but when the arrival of tasks is dynamic, the problem becomes much more difficult. To solve large scale open shop scheduling problem with release dates, heuristic algorithms are more promising compared with metaheuristic algorithms. In this paper, a framework of general scheduling object is developed, under which open shop scheduling problem is described. Then, a complex scheduling network model for open shop scheduling problem is established, and the problem is transformed into reasonably arranging the node traversal order with the goal of traversing all nodes in the network as quickly as possible, on condition that each node has a traversal time and only disconnected nodes can be traversed simultaneously. Considering that network structural features and local time attributes of nodes can provide heuristic information, six single heuristic rules are raised and a novel complex network based dynamic rule selection approach is proposed to solve dynamic open shop problem by switching dynamically the scheduling rules based on real-time production status. Finally, two experiments are carried out and the experimental results show that the proposed heuristic rules have acceptable performance, and the proposed complex network based dynamic rule selection approach is feasible.

    Citation: Zilong Zhuang, Zhiyao Lu, Zizhao Huang, Chengliang Liu, Wei Qin. A novel complex network based dynamic rule selection approach for open shop scheduling problem with release dates[J]. Mathematical Biosciences and Engineering, 2019, 16(5): 4491-4505. doi: 10.3934/mbe.2019224

    Related Papers:

    [1] Na Zhang, Jianwei Xia, Tianjiao Liu, Chengyuan Yan, Xiao Wang . Dynamic event-triggered adaptive finite-time consensus control for multi-agent systems with time-varying actuator faults. Mathematical Biosciences and Engineering, 2023, 20(5): 7761-7783. doi: 10.3934/mbe.2023335
    [2] Xiangfei Meng, Guichen Zhang, Qiang Zhang . Robust adaptive neural network integrated fault-tolerant control for underactuated surface vessels with finite-time convergence and event-triggered inputs. Mathematical Biosciences and Engineering, 2023, 20(2): 2131-2156. doi: 10.3934/mbe.2023099
    [3] Xiaohan Yang, Yinghao Cui, Zhanhang Yuan, Jie Hang . RISE-based adaptive control of electro-hydraulic servo system with uncertain compensation. Mathematical Biosciences and Engineering, 2023, 20(5): 9288-9304. doi: 10.3934/mbe.2023407
    [4] Siyu Li, Shu Li, Lei Liu . Fuzzy adaptive event-triggered distributed control for a class of nonlinear multi-agent systems. Mathematical Biosciences and Engineering, 2024, 21(1): 474-493. doi: 10.3934/mbe.2024021
    [5] Qiushi Wang, Hongwei Ren, Zhiping Peng, Junlin Huang . Dynamic event-triggered consensus control for nonlinear multi-agent systems under DoS attacks. Mathematical Biosciences and Engineering, 2024, 21(2): 3304-3318. doi: 10.3934/mbe.2024146
    [6] Mingxia Gu, Zhiyong Yu, Haijun Jiang, Da Huang . Distributed consensus of discrete time-varying linear multi-agent systems with event-triggered intermittent control. Mathematical Biosciences and Engineering, 2024, 21(1): 415-443. doi: 10.3934/mbe.2024019
    [7] Chaoyue Wang, Zhiyao Ma, Shaocheng Tong . Adaptive fuzzy output-feedback event-triggered control for fractional-order nonlinear system. Mathematical Biosciences and Engineering, 2022, 19(12): 12334-12352. doi: 10.3934/mbe.2022575
    [8] Songyu Hu, Ruifeng Hu, Liping Tang, Weiwei Jiang, Banglin Deng . Quantitative generation of microfluidic flow by using optically driven microspheres. Mathematical Biosciences and Engineering, 2019, 16(6): 6696-6707. doi: 10.3934/mbe.2019334
    [9] Duoduo Zhao, Fang Gao, Jinde Cao, Xiaoxin Li, Xiaoqin Ma . Mean-square consensus of a semi-Markov jump multi-agent system based on event-triggered stochastic sampling. Mathematical Biosciences and Engineering, 2023, 20(8): 14241-14259. doi: 10.3934/mbe.2023637
    [10] Tongyu Wang, Yadong Chen . Event-triggered control of flexible manipulator constraint system modeled by PDE. Mathematical Biosciences and Engineering, 2023, 20(6): 10043-10062. doi: 10.3934/mbe.2023441
  • In the open shop scheduling problem, resources and tasks are required to be allocated in an optimized manner, but when the arrival of tasks is dynamic, the problem becomes much more difficult. To solve large scale open shop scheduling problem with release dates, heuristic algorithms are more promising compared with metaheuristic algorithms. In this paper, a framework of general scheduling object is developed, under which open shop scheduling problem is described. Then, a complex scheduling network model for open shop scheduling problem is established, and the problem is transformed into reasonably arranging the node traversal order with the goal of traversing all nodes in the network as quickly as possible, on condition that each node has a traversal time and only disconnected nodes can be traversed simultaneously. Considering that network structural features and local time attributes of nodes can provide heuristic information, six single heuristic rules are raised and a novel complex network based dynamic rule selection approach is proposed to solve dynamic open shop problem by switching dynamically the scheduling rules based on real-time production status. Finally, two experiments are carried out and the experimental results show that the proposed heuristic rules have acceptable performance, and the proposed complex network based dynamic rule selection approach is feasible.


    Important properties of the HSA, such as fast and accurate responses, a high force/mass ratio and relatively good stiffness, have attracted great interest in the HSA and its applications. In the last two decades, high-performance controller design of the HSA has attracted increasing attention due to the expanded performance requirements of technical systems in the industry [1,2,3,4].

    A large number of machines driven by HSAs often work with high payloads in harsh and mostly external environments. As a result of variables of their environment, such as temperature, dust, humidity, wear, variable loads and disturbances, the HSA is usually subject to large uncertainties during operation. Hence, high-precision control of the HSA has always challenged researchers due to its unmodeled dynamics, large nonlinearities, parametric uncertainties, unmeasurable states in practice, etc. It is well known that it is impossible to determine most of the physical parameters of HSA components. While some HSA parameters are available only with certain accuracy, the other parameters are completely unknown. Dominant nonlinear sources existing in HSAs are impossibility of accurate determining parameters, which are very difficult to handle with high accuracy. These unknown parameters are caused by protection of proprietary data of individual manufacturers or indirect measuring and calculating, pressure losses, transient and turbulent flow conditions, friction, leakage characteristics, and generation discontinuous control signals to HSAs due to effects of saturation and changing the direction of servo valve. Furthermore, variable working conditions during operation, such as oil temperature, the bulk modulus, fluctuating supply pressure and pipe volume will lead to parameter changes, which worsen the existing control performances. These facts make it difficult to realize high-quality control of the HSA, which cannot be achieved without knowing the accurate HSA model [5,6,7,8].

    Further, direct measurement of the whole HSA state vector is not feasible for practical implementation and in addition would require very expensive measurement equipment. It is more convenient to use control algorithms which apply methods based on state reconstruction rather than to perform direct measurements of the states [9].

    In modern control theory, optimal control of the HSA plays a vital role in the controller design. Namely, the main challenge is to design optimal control algorithms that will affect the minimum energy consumption [10,11]. The optimal control design is an offline control technique that usually depends on perfect knowledge of the HSA model, which is not possible to obtain in most practical situations. Even if an approximated model of the HSA can be developed, the dynamic uncertainty, produced by the mismatch between the approximated model and the true HSA model, will degrade the control performance of the traditionally designed optimal controller [9,12]. Therefore, further research on the design of optimal controllers of HSAs is very important and our primary aim for this study.

    The practical applicability of control algorithms is enhanced by the fact that nonlinear systems can be very precisely represented by linear models with online estimated dynamics [13,14]. Many modern engineering applications such as intelligent vehicles [15,16], modernized microgrids [17], microphone sensing [18], strain prediction for fatigue [19], maintaining the security of cyber–physical systems [20], robotic manipulation tasks [21], 2-degree-of-freedom helicopter [22] and requests for online controller design which rely on linear systems.

    Adaptive dynamic programming (ADP) ensures an effective way to achieve high performance of the optimal controller which relies on adaptive control, optimal control and reinforcement learning [9,23,24,25,26]. ADP represents a kind of data-based control technique which can guarantee the stability of the feedback control system [9]. Recently, the field of ADP application has also been expanded to various research areas, including robotic systems, aerospace systems, guided missiles, spacecraft, etc. [27,28,29]. In circumstances of unknown system dynamics and unmeasurable states, of great interest is to use ADP techniques based on measured input/output data from linear systems, which are commonly called output feedback. A main benefit of the output feedback techniques is that knowledge of the HSA dynamics is not needed for their application. For an unknown HSA model, this indirect technique generates a sequence of suboptimal controllers which converge to the optimal control policy with an increasing number of iterations.

    The implementation of ADP algorithms is usually based on periodic selection [30]. In order to save limited communication and computational resources, event-triggered strategies have recently started to be applied in control algorithms based on ADP [31,32,33,34]. Moreover, the number of updates of the control inputs in this way will be smaller compared to the periodic update of the controller, because it is updated only when necessary (e.g., when the performance of the system deteriorates). The implementation of event-triggered algorithms is based on aperiodic sampling. Several event-based controllers have been proposed in the literature, most of which are state-feedback controllers [35,36,37,38,39]. In contrast, this paper will consider the event-triggered ADP-based control problem of HSAs in the case where only output feedback is available.

    This paper considers an online learning technique, where during operation, from measured input/output data, the controller learns to compensate unknown HSA dynamics, various disturbances and modeling errors, ensuring desired performances of the control system. The optimal control law is accomplished iteratively based on output feedback, state reconstruction and ADP. The unknown HSA model is first identified after which the algebraic Riccati equation (ARE) is iteratively solved. To ensure consistency of approximations and obtain unique solutions in each iteration step, some exploration noise must be added to control input to meet the requirements of the persistent excitation condition [40,41,42]. For exploration noise, some persistent excitation is usually applied such as white noise or pseudo random binary signals (PRBS). The selection of exploration noise is a non-trivial task for the most learning problems, as it can affect the accuracy of solutions, especially for large systems [43]. By applying the theory of experimental design, we will use the sum of sinusoidal signals as an exploration noise that will enable the output of the system to carry maximum information about the system, which will shorten the learning time, i.e., speed up the controller design process. Thus, the obtained input and output signals are used to reconstruct the state vector of the model, which is of great practical importance in relation to control techniques with direct state measurement that rely on a large number of sensors.

    Due to implementation of the ADP-based control techniques, it is easier to realize data acquisition for the discrete-time HSA model in relation to its continuous-time model. ADP-based control methodology for discrete-time systems is proposed in [44].

    We chose to use the measured input and output data to reconstruct the state vector of the discretized HSA model, after which ADP-based control can be implemented. The control law is learned iteratively and very efficiently provides solutions for optimal control of HSAs based only on measurements in real time. The main advantage of the proposed control methodology is avoiding the knowledge of system dynamics, which is very important under practical conditions.

    By applying an event-based control strategy, the number of control input updates will be reduced relative to periodic update of the controller, because the controller is only updated when certain conditions are met. In this way, energy, computing and communication resources will be significantly preserved.

    The rest of the paper is organized as follows. The problem of modelling an HSA with unknown dynamics is presented in Section 2. Event-triggered control based on ADP is shown in Section 3. Simulation results show the validity and effectiveness of the event-triggered ADP-based controller for HSAs in the presence of completely model uncertainty in Section 4. Finally, Section 5 gives the concluding remarks.

    The HSA under study is shown in Figure 1, and it consists of a servo valve and a hydraulic cylinder. The analysis of the properties of the HSA comes out from the dynamics of its components, which involves the piston motion dynamics, pressure dynamics at the cylinder and servo valve dynamics. Hence, the model of the HSA is derived from complex nonlinear equations that depend on many parameters which cannot be accurately obtained [7,8].

    Figure 1.  The HSA configuration.

    See Table 1 for the description of the HSA parameters. Using the notation in Figure 1, and defining the area ratio of the piston α=Ab/Aa as well as Va=Va0+yAa, Vb=Vb0+(Ly)αAa and qLi=cLi(papb), where cLi is the internal leakage flow coefficient; cvi>0 denote discharge coefficients, the sign function sg(x)={xx00x<0 and assuming an external leakage negligible, the considered model can be described by the following equations:

    mt¨y=AapaAbpbFf(˙y)KeyFext, (2.1)
    ˙pa=βeVa(y)(qaAa˙yqLiqLea), (2.2)
    ˙pb=βeVb(y)(qb+αAa˙y+qLiqLeb), (2.3)
    qa=qsv1qsv2=cv1sg(xv)sign(pspa)|pspa|cv2sg(xv)sign(pap0)|pap0|, (2.4)
    qb=qsv3qsv4=cv3sg(xv)sign(pspb)|pspb|cv4sg(xv)sign(pbp0)|pbp0|. (2.5)
    Table 1.  Parameters of the HSA.
    Notations Descriptions
    xv The spool valve displacement
    pa, pb Forward and return pressure
    qa, qb Forward and return flows
    y Piston displacement
    L Piston stroke
    Ke Load spring gradient
    pS, p0 Supply and tank pressure
    mt, mp, m total mass, piston mass, payload mass
    Ff Friction force
    Fext Disturbance force
    Aa, Ab Effective areas of the head and rod piston side
    Va, Vb, Va0, Vb0 Fluid volumes of the head and rod piston side and corresponding initial volumes
    qLi, qLe Internal and external leakage flow
    βe Bulk modulus of the fluid

     | Show Table
    DownLoad: CSV

    According to Eqs (2.1)–(2.5), and by defining the state and input variables as

    x(t)=[x1(t)x2(t)x3(t)x4(t)]T[y(t)˙y(t)pa(t)pb(t)]T, (2.6)
    u(t)=xv(t), (2.7)

    the governing nonlinear continuous-time dynamics of the HSA can be expressed in a state-space form as follows:

    ˙x(t)=f(x(t))+g(x(t),u(t))+h(t),y(t)=η(x(t)), (2.8)

    where f(x(t)) and g(x(t),u(t)) are the state dynamics and the input function, respectively:

    f(x(t))=[x21mt(Aax3αAax4Ff(x2)Kex1)βeAax1+Va0(Aax2+cLi(x3x4))βeαAa(Lx1)+Vb0(αAax2+cLi(x3x4))],
    g(x(t),u(t))=[00βeAax1+Va0(cv1sg(u)sign(psx3)|psx3|cv4sg(u)sign(x3p0)|x3p0|)βeαAa(Lx1)+Vb0(cv3sg(u)sign(psx4)|psx4|cv2sg(u)sign(x4p0)|x4p0|)],

    where output function η(x(t))=x1(t) and disturbance function h(t)=[h1(t)Fext/mt+h2(t)h3(t)h4(t)] include loads, unmodelled dynamics and parameter uncertainties.

    One of the main nonlinearities of the cylinder model is the nonlinear friction force Ff, which consists of static friction, Coulomb friction and Stribeck effect of velocity. An extensive study related to acting friction forces upon the HSA can be found in [7]. Further, we consider the linearized model of the HSA, whose parameters are experimentally identified for different working points of the HSA (i.e., different positions and external load conditions) [8]. Now, the model equations are expressed in a more suitable way in terms of the load pressure:

    pL=paαpb, (2.9)

    which leads to simplified dynamic equations. At last, using the new state vector [x1(t)x2(t)x3(t)]T[y(t)˙y(t)pL(t)]T allows us to express the HSA in a more compact form. Taking an operating point x0[y0˙y0pL0]T, and assuming dominance of the first order term from the Taylor series expansion, the linearized continuous-time description of the reduced order is stated as follows

    ˙x(t)=Ax(t)+B(t)u(t), (2.10)
    y(t)=Cx(t), (2.11)

    where A=[0100BCmtAamt0KdKp], B=[00Kx]T and C=[100]. The sensibility constants can be found as follows:

    Kd=Aa(βeVA+α2βeVB)1,Kp=βe(KpACLi(1+α2))VA(1+α2)αβe(KpBα2+CLi(1+α2))VB(1+α2),Kx=βeVAKxAαβeVBKxB,

    where the flow sensibility constants regarding the pressure at the cylinder chambers are stated as:

    KpA={cvxv0pspA0forxv>0cvxv0pA0p0forxv<0, (2.12)
    KpB={cvxv0pB0p0forxv>0cvxv0pspB0forxv<0, (2.13)

    and the flow sensibility constants regarding the spool position are stated as

    KxA={cvpspA0forxv>0cvpA0p0forxv<0, (2.14)
    KxB={cvpB0p0forxv>0cvpspB0forxv<0. (2.15)

    The previously mentioned valve sensibility constants are very significant in defining system stability and other dynamic characteristics [8]. Namely, the flow gain Kx has a direct impact on the stability of the HSA, because it directly affects the gain constant in the open loop of the HSA. Further, direct impact on the damping ratio of the HSA has the flow-pressure constant Kp. Hence, the pressure sensibility Kpx=Kx/Kp is quite high, which explains the ability of the HSA to transfer large friction loads with a small error.

    Let us consider a linear continuous-time model of the HSA with unknown dynamics, as follows:

    ˙x(t)=Ax(t)+Bu(t), (3.1)
    y(t)=Cx(t), (3.2)

    where x(t)Rn, u(t)Rm and y(t)Rr are the system state vector, the control input vector, and the output vector, respectively. ARn×n, BRn×m and CRr×n are unknown system matrices, assuming that (A,B) is controllable and (A,C) is observable.

    For the HSA described by (3.1) and (3.2), the performance index is stated as

    J(x0)=0[yT(τ)Qy(τ)+uT(τ)Ru(τ)]dτ, (3.3)

    where x0Rn is an initial state, Q=QT0 and R=RT>0, with (A,Q1/2C) being observable.

    A control law is also called a policy. The design objective is to find a linear optimal control policy in the form of

    u=Kx, (3.4)

    which minimizes the performance given by index (3.3). The optimal feedback gain matrix Kcan be determined as

    K=R1BTP, (3.5)

    where P=(P)T>0 is a unique symmetric positive definite solution of the well-known ARE

    ATP+PA+CTQCPBR1BTP=0, (3.6)

    under conditions that the system matrices are accurately known, as well as conditions that the pair (A,B) is controllable and the pair (A,Q1/2C) is observable [12]. It should be noted that this optimal control design is mainly applicable to low order simple linear systems. In fact, for high-order large scale systems, it is usually difficult to directly solve P from (3.6), which is nonlinear in P.

    Also, for practical implementation of the control system, it is easier to realize the data acquisition for discrete-time systems than for continuous-time systems. Consequently, we transform the continuous-time HSA into the following discrete-time HSA:

    xk+1=Adxk+Bduk, (3.7)
    yk=Cxk, (3.8)

    where Ad=eAh and Bd=h0(eAτdτ)B, where h>0 is a specific sampling period, assuming ωh=2π/h is the nonpathological sampling frequency whose existence is well known [45]. In other words, the controllability and observability of the original continuous-time HSA system is kept after discretization. Namely, if the state, input and output vectors at the sampled instant kh are xk, uk and yk, respectively, then (Ad,C) and (Ad,Q1/2C) are observable while (Ad,Bd) is controllable.

    As depicted in Figure 2, the ADP-based controller for the discretized HSA system consists of three parts: the state reconstruction, critic, and actor. The state reconstruction provides the relationship between the input/output data and the HSA states, which allows one to solve the optimal control problem of an HSA with unknown dynamics. Based on the input/output data, the critic part of the controller is designed to evaluate the performance of the control policy. The controller learns online in order to maximize its performance. Finally, the actor applies the improved control policy. The updates of the control actions are governed by an event-triggering mechanism to reduce the amount of data transmission from the controller to the HSA system.

    Figure 2.  Event-triggered ADP-based control algorithm for the discretized HSA system.

    The event-triggered design is based on a periodic sampling with a nonpathological h>0. We use ˆuk to represent the sampled value of uk, that is

    ˆuk=ukj,k[kj,kj+1), (3.9)

    where {kj}0 is a monotonically increasing sequence of the sampling time instants, and the control input is only updated at the discrete-time instants: k0,k1,k2, For the convenience of discussions, define the sampling error of the input data as

    Δk=ˆukuk. (3.10)

    Hence, the discrete-time system described by (3.7) and (3.8) can be rewritten as

    xk+1=Adxk+(Bduk+Δk), (3.11)
    yk=Cxk. (3.12)

    Further, the performance index for the discretized system described by (3.7) and (3.8) is

    Jd(x0)=j=0yTjQdyj+uTjRduj, (3.13)

    where Qd=Qh and Rd=Rh. The optimal control low minimizing (3.13) is

    uk=Kdxk, (3.14)

    where the discrete optimal feedback gain matrix is Kd=(R+BTdPdBd)1BTdPdAd, where Pd is the unique symmetric positive definite solution to

    ATdPdAdPd+CTQCATdPdBdKd=0. (3.15)

    Since (3.15) is nonlinear in Pd, it is difficult to directly solve Pd for high-order large-scale systems. Nevertheless, many efficient algorithms have been developed to numerically approximate the solution of (3.15). One of such algorithms was developed by Hewer [46], and it is introduced in the form of Lemma 3.1.

    Lemma 3.1. Let K0Rm×n be any stability feedback gain matrix and Pj be the symmetric positive definite solution of the Lyapunov equation

    (AdBdKj)TPj(AdBdKj)+CTQdC+KTjRdKj=0, (3.16)

    where Kj, j=1,2, can be updated as follows:

    Kj=(R+BTdPj1Bd)1BTdPj1Ad. (3.17)

    Then, it holds that

    1) AdBdKj is a stability matrix

    2) PdPj+1Pj

    3) lim, \mathop {\lim }\limits_{j \to \infty } {P_j} = P_d^*.

    By iteratively solving the Lyapunov equations given by (3.16), which is linear in P_j , and recursively updating the control policy K_j by (3.17), the solution to the nonlinear equation given by (3.15) is numerically approximated [46]. It has been concluded that the sequences \left\{ P_j \right\}_{j = 0}^\infty and \left\{ K_j \right\}_{j = 0}^\infty , computed from this algorithm, converge to P_d^* and K_d^* , respectively. Moreover, for j = 0, 1, \ldots , A_d - B_d K_j is a Schur matrix. It should be noted that the method by Hewers involves a model-based policy iteration (PI) algorithm, which cannot be directly applied to the problem studied in this paper since it is an offline algorithm which depends on the system parameters. To apply this algorithm online for the discretized HSA described by (3.7) and (3.8), we will develop the control algorithm based on ADP via output feedback, which does not depend on the knowledge of HSA matrices.

    Motivated by [44,47], the discrete-time HSA described by (3.7) and (3.8) can be extended by using input/output sequences on the time horizon [k-N, k-1] as follows:

    \begin{equation} \begin{gathered} x_k = A_d^N x_{k-N} + V(N) \bar{u}_{k-1, k-N}, \\ \bar{y}_{k-1, k-N} = U(N) x_{k-N} + T(N)\bar{u}_{k-1, k-N}, \end{gathered} \end{equation} (3.18)

    where

    \begin{equation*} \begin{gathered} \bar{\Delta}_k = \begin{bmatrix} \Delta_{k-1}^T & \Delta_{k-2}^T & \ldots & \Delta_{k-N}^T \end{bmatrix}^T, \\ \bar{u}_{k-1, k-N} = \begin{bmatrix} \hat{u}_{k-1}^T & \hat{u}_{k-2}^T & \ldots & \hat{u}_{k-N}^T \end{bmatrix}^T, \\ \bar{y}_{k-1, k-N} = \begin{bmatrix} y_{k-1}^T & y_{k-2}^T & \ldots & y_{k-N}^T \end{bmatrix}^T, \\ V(N) = \begin{bmatrix} B_d & A_d B_d & \ldots & A_d^{N-1} B_d \end{bmatrix}], \\ U(N) = \begin{bmatrix} (C A_d^{N-1})^T & (C A_d)^T & \ldots & C^T \end{bmatrix}^T, \\ T(N) = \begin{bmatrix} 0 & C B_d & C A_d B_d & \cdots & C A_d^{N-2} B_d \\ 0 & 0 & C B_d & \cdots & C A_d^{N-3} B_d \\ \vdots & \vdots & \ddots & \ddots & \vdots \\ 0 & 0 & \cdots & 0 & C B_d \\ 0 & 0 & \cdots & 0 & 0 \\ \end{bmatrix}, \end{gathered} \end{equation*}

    and the observability index is N = \max(\rho_u, \rho_v) where \rho_u is the minimum integer which ensures that U(\rho_u) has full column rank and \rho_v is the minimum integer which ensures that V(\rho_v) has full row rank [44]. Therefore, there exists a left inverse of U(N) , stated as U^+(N) = \left[U^T(N) U(N) \right]^{-1} U^T(N) . With the state reconstruction in (3.18), the idea of an ADP-based controller with output feedback can be applied to solve the optimal control problem of HSAs with unknown dynamics. A uniqueness of state reconstruction is stated in the form of Lemma 3.2 as follows [48].

    Lemma 3.2. If the conditions of observability and controllability of the system described by (3.7) and (3.8) are fulfilled, then the states of the HSA are uniquely received in terms of measured inputs and outputs signals as follows:

    \begin{equation} x_k = \Theta z_k, \end{equation} (3.19)

    where \Theta = \begin{bmatrix} M_u & M_y \end{bmatrix} has full row rank, M_u = V(N) - M_y T(N) , M_y = A_d^N U^+(N) and z_k = \begin{bmatrix} \bar{u}_{k-1, k-N}^T & \bar{y}_{k-1, k-N}^T \end{bmatrix}^T \in \mathbb{R}^q , where q = N[\dim(u)+\dim(y)] .

    Now, based on (3.16) and (3.17), an online learning strategy using output feedback can be introduced in the form of u^*_k = - \overline{K}_d z_k , providing suboptimal property of the closed-loop system. The discrete-time model (3.11) can be stated as follows:

    \begin{equation} x_{k+1} = A_j x_k + B_d \left( K_j x_k + \hat{u}_k \right), \end{equation} (3.20)

    where A_j = A_d - B_d K_j . Setting \bar{K}_j = K_j \Theta and \bar{P}_j = \Theta^T P_j \Theta , from (3.16) and (3.20), it can be obtained

    \begin{equation} \begin{split} & z_{k+1}^T \bar{P}_j z_{k+1} - z_k^T \bar{P}_j z_k = \\ & \left( \bar{K}_j z_k + \hat{u}_k \right)^T \begin{bmatrix} B_d^T \bar{P}_j B_d & B_d^T \bar{P}_j A_d \end{bmatrix} \begin{bmatrix} -\bar{K}_j z_k + \hat{u}_k \\ 2 z_k \end{bmatrix} - \left( y_k^T Q y_k + z_k^T \bar{K}_j^T R \bar{K}_j z_k \right) = \\ & \left[ \hat{u}_k^T \otimes \hat{u}_k^T - (z_k^T \otimes z_k^T)(\bar{K}_j^T \otimes \bar{K}_j^T) \right]\text{vec}(\bar{H}_j^1) + \\ & 2 \left[ (z_k^T \otimes z_k^T)(I_q \otimes \bar{K}_j^T)+(z_k^T \otimes u_k^T) \right] \text{vec}(\bar{H}_j^2) - \left( y_k^T Q y_k + z_k^T \bar{K}_j^T R \bar{K}_j z_k \right)\overset{\wedge}{ = } \\ & \phi^1 \text{vec}(\bar{H}_j^1)+\phi^2 \text{vec}(\bar{H}_j^2)-\left( y_k^T Q y_k + z_k^T \bar{K}_j^T R \bar{K}_j z_k \right), \end{split} \end{equation} (3.21)

    where \bar{H}_j^1 = B_d^T \bar{P}_j B_d , \bar{H}_j^2 = B_d^T \bar{P}_j A_d \Theta , \phi ^1 = \hat{u}_k^T \otimes \hat{u}_k^T -(z_k^T \otimes z_k^T)(\bar{K}_j^T \otimes \bar{K}_j^T) and \phi ^2 = 2 \left[(z_k^T \otimes z_k^T)(I_q \otimes \bar{K}_j^T) + (z_k^T \otimes \hat{u}_k^T) \right] .

    The symbol \otimes is used to denote a Kronecker product operator. The vector function \text{vec}(V) = \begin{bmatrix} v_1^T & v_2^T & \ldots & v_m^T \end{bmatrix}^T is stated as an mn -vector formed by stacking the columns of V \in \mathbb{R}^{n\times m} on top of one another, where v_i \in \mathbb{R}^n denotes the columns of matrix V . For an arbitrary symmetric matrix M \in \mathbb{R}^{n\times n} , \text{vecs}(M) = \left[m_{11}, 2m_{12}, \ldots, 2m_{1n}, m_{22}, 2m_{23}, \ldots, 2m_{n-1, n}, m_{nn} \right]^T \in \mathbb{R}^{n(n+1)/2} and for an arbitrary column vector v \in \mathbb{R}^n , \tilde{v} = \left[v_1^2, v_1v_2, \ldots, v_1v_n, v_2^2, v_2v_3, \ldots, v_{n-1}v_n, v_n^2 \right]^T \in \mathbb{R}^{n(n+1)/2} .

    The convergence of the online learning control using output feedback is guaranteed under the rank condition stated in the form of Lemma 3.3 [47]. Lemma 3.3 is about the condition of persistent excitation in adaptive control theory [49,50].

    Lemma 3.3. Let us suppose that for a sufficiently large s \in \mathbb{Z}_+ , it holds that

    \begin{equation} \mathit{\text{rank}}(\Gamma) = \left( \dim(u)+\dim(z) \right) \left(\dim(u) + \dim(z) + 1\right)/2, \end{equation} (3.22)

    where

    \begin{equation} \Gamma = \left[ \eta_{k0} \otimes \eta_{k0}, \eta_{k1} \otimes \eta_{k1}, \cdots, \eta_{ks} \otimes \eta_{ks} \right], \; \mathit{\text{where}} \; k_0 < k_1 < \cdots < k_s \in \mathbb{Z}_+ \; \mathit{\text{and}} \; \eta_{kj} = [\hat{u}_{kj}^T, z_{kj}^T]^T, j = \overline{0, s}; \end{equation} (3.23)

    then \left(\bar{P}_j, \bar{H}_j^1, \bar{H}_j^2 \right) can be uniquely solved based on \bar{K}_j and measurable online data during the period k \in [k_0, k_s] . Further, \bar{K}_{j+1} is obtained as follows:

    \begin{equation} \bar{K}_{j+1} = \left( R + \bar{H}_j^1 \right)^{-1} \bar{H}_j^2. \end{equation} (3.24)

    Some exploration noise e_k , which satisfies the persistent excitation condition, must be added into the input signal during the online learning phase due to the satisfaction of the rank condition given by (3.22), without affecting the convergence of the learning phase [43,51,52]. Note that (3.21) is called the policy evaluation, which is used to uniquely solve \bar{P}_j , and (3.24) is the policy improvement, which is used to update the control gain \bar{K}_{j+1} . Finally, we present the ADP-based online learning control algorithm in Figure 3.

    Figure 3.  Flowchart of ADP-based controller design.

    It should be noted that solving (3.21) instead of (3.16), completely eliminates the original request on the accurate knowledge of the HSA dynamics. Now, we only need to measure u_k and y_k . Namely, having in mind the expression for z_k , we can see that the control policy \hat{u}_k = -\overline{K}_k^* z_k + \Delta_k contains only the previously measured input-output data. With the event-triggered control law \hat{u}_k , the system given by (3.20) is globally asymptotically stable (GAS) at the origin if

    \begin{equation} \left\| \Delta_k \right\|^2 \le \frac{\alpha \gamma \left\| y_k \right\|^2 + \lambda_{\min} (R_d) \left\| \hat{u}_k \right\|^2}{\eta}, \end{equation} (3.25)

    where \alpha \in (0, 1) and \eta is a positive constant satisfying \eta \ge \lambda_{\max} \left(R_d + B_d^T \bar{P}_d B_d \right) .

    The convergence of the ADP-based control algorithm is presented in the form of Theorem 3.4. For Hurwitz feedback matrix A - B K , K \in R^{m\times n} is called stabilizing feedback gain matrix for a linear system \dot{x} = A x + B u .

    Theorem 3.4. If the condition of Lemma 3.3 is fulfilled, with some initial stabilizing feedback gain matrix \overline{K}_0 , then the sequences \left\{ \overline{P}_j \right\}_{j = 0}^{\infty } and \left\{ \overline{K}_j \right\}_{j = 0}^{\infty} received from this algorithm, converge to their optimal values \overline{P}^* and \overline{K}^* , respectively [46,47].

    Proof. If P_{j}^{{}} = P_{j}^{T} represents the solution of (3.16), under the stability feedback gain matrix \overline{K}_j , then K_{j+1} is uniquely obtained from (3.17). It can be easily shown that \overline{P}_j and \overline{K}_{j+1} fulfill (3.21) and (3.24). Now, setting \overline{P} and \overline{K} as solutions of (3.21) and (3.24), Lemma 3.3 provides that \overline{P}_j = \overline{P} and \overline{K}_{j+1} = \overline{K} are uniquely stated. Furthermore, from Lemma 3.1, we have that \underset{j \to \infty}{\mathop{\lim }} \overline{K}_j = \overline{K}_d^* and \underset{j \to \infty}{\mathop{\lim}} \overline{P}_j = \overline{P}_d^* . The proof of convergence is proved.

    The hybrid nature of the controller is shown in Figure 4. It is shown there that the feedback gain or policy is updated at discrete times by using (3.24) after the solution to (3.21) has been determined. On the other hand, the control input is a discrete time signal depending on the state z(k) at each time k . From Figure 4, it can be seen that the control gains are updated at discrete times, but the control signal is piecewise continuous.

    Figure 4.  Hybrid nature of control signal.

    In this section, we apply the proposed event-triggered ADP-based control design to the HSA. In the case of unknown dynamics and unmeasurable states of the HSA, it is meaningful to use the ADP-based method. Consequently, we conduct simulations on the HSA given by the linearized continuous-time description of (2.10) and (2.11) to show the effectiveness of the ADP-based control algorithm. A basic condition for energy savings in many hydraulically driven industrial systems is a high-quality design of event-triggered ADP control for the HSA.

    For this purpose, the HSA is discretized by applying the periodic sampling period h = {0.1}\ {\rm{s}} and the zero-order holder. The approximated optimal feedback gain and performance index for the discretized model of the HSA are iteratively obtained.

    The effectiveness of the ADP-based control algorithm will be considered for the HSA model described by (2.10) and (2.11) with the following parameters: the viscous friction B_C = {200} \ {\rm{N}}\ {\rm{s}}\ {\rm{m}}^{-1} , the supply pressure p_S = {45} \ {\rm{bar}} , the tank pressure p_0 = {1.6} \ {\rm{bar}} , the bulk modulus of the fluid \beta_e = 2 \times 10^8 \mathrm{~Pa} , the total mass m = {25}\ {\rm{kg}} , the initial chamber volumes V_{a0} = V_{b0} = 8.2 \times 10^{-6} \mathrm{~m}^3 , the load spring gradient K_e = 10^{-1} , the effective area of the head side of the piston A_a=4.91 \times 10^{-4} \mathrm{~m}^2, the effective area of the rod side of the piston A_b=2.43 \times 10^{-4} \mathrm{~m}^2 , the internal leakage coefficient c_{Li} = 5 \times 10^{-14} , the piston stroke L = {1} \ {\rm{m}} and discharge coefficients of valve orifices c_{vi} = 1.15 , i = \overline{1, 4} .

    For the purpose of demonstrating the event-triggered ADP method with the HSA, the weight matrices, Q and R , are chosen to be identity matrices, the observability index is N = 3 , initial state vector is x_0 = \begin{bmatrix} 5 & -5 & -10 \end{bmatrix} and the convergence threshold \varepsilon is selected as {{10}^{-1}} .

    It should be noted that our event-driven ADP control design does not require exact knowledge of the HSA matrices. But, only for numerical verification via simulation, it is assumed that the system matrices in (2.10) and (2.11) are known.

    To verify the benefits of the ADP based online learning controller, Figure 5 depicts the errors between \bar{P}_j and \bar{P}_d^* and \bar{K}_j and \bar{K}_d^* , which indicate the convergence of \bar{P}_j and \bar{K}_j .

    Figure 5.  Convergence of \bar{P}_j and \bar{K}_j to their respective optimal values \bar{P}^* and \bar{K}^* during the learning process.

    The evolution of the maximum cost for HSA is shown in Figure 6(a), where V_1 is the maximum cost by using the initial control policy, and V_7 is the maximum cost by using the control policy after seven iterations. It can be seen that the approximated cost function V_7 has been remarkably reduced relative to the initial cost V_1 . Figure 6(b) shows the 3D plot of the approximation error of the cost function. This error is close to zero which confirms that good approximation of the optimal cost function is achieved during the learning process.

    Figure 6.  (a) Comparison of the cost functions during learning; (b) error between the optimal and approximated cost function signal.

    The improved control policy and the initial control policy are compared in Figure 7(a). Further, Figure 7(b) shows the 3D plot of the difference between the approximated control obtained by using the online ADP-based control algorithm and the optimal control. This error is close to zero, which confirms that good approximation of the optimal input is also achieved during the learning process.

    Figure 7.  (a) Comparison of the control policies during the learning process; (b) error between the optimal and approximated input signal.

    Figure 8 shows the control input and the states of the HSA system desribed by (2.10) and (2.11) by using the ADP-based controller with periodic sampling.

    Figure 8.  Control input and states of the HSA model by using the ADP-based controller.

    To illustrate the benefits of the event-triggered ADP method, the control input and the states of the original HSA system described by (2.10) and (2.11), as obtained by using the event-triggered ADP-based controller is shown in Figure 9.

    Figure 9.  Control input and states of the HSA model by using the event-triggered ADP-based controller.

    The comparison of sampling numbers by using the event-triggered ADP controller versus the ADP controller with periodic sampling is shown in Figure 10.

    Figure 10.  Comparison of the total sampling numbers.

    It can be observed that similar control effects have been achieved by the two methods, however, for the event-triggered ADP method, the control input is updated only when the squared norm of the triggering error reaches the threshold, and it is kept constant otherwise. It is also shown that about 54\% communication between the controller and the HSA is reduced by using the event-triggered ADP method instead of the ADP method. The sequence of steps of event-triggered sampling is depicted in Figure 11.

    Figure 11.  Sequence of steps of event-triggered sampling.

    This paper has considered the event-triggered data-driven optimal controller of the HSA with completely unknown dynamics as based on an ADP framework. A basic advantage of the presented control methodology is its ability to avoid the knowledge of entire system dynamics, which is very important in real conditions. By using the output feedback and the state reconstruction method an applied ADP-based control technique has been shown to be a useful tool for digital implementation in a real HSA. For that purpose, a discrete-time control policy was iteratively learned based on the discretized HSA model. The learned control policy very efficiently ensures online solutions to data-driven optimal control problems for the HSA. The presented online control policy only uses measured input/output data to learn the optimal control gain. Then, to reduce the communication between the controller and the HSA, an output feedback event-triggered ADP controller has been designed. The simulation results have shown the validity and effectiveness of the applied control approach for the HSA.

    This research was supported in part by the Serbian Ministry of Education, Science and Technological Development under grant 451-03-47/2023-01/200108, the National Natural Science Foundation of China under grants 61773181, 61976081, 62073001, 62103293 and 62203153,111 Project under grant B23008, the Fundamental Research Funds for the Central Universities under grant JUSRP51733B, the Anhui Provincial Key Research and Development Project under grant 2022i01020013 and the Natural Science Fund for Excellent Young Scholars of Henan Province under grant 202300410127.

    The authors declare that there is no conflict of interest.



    [1] Y. T. Leung, Handbook of Scheduling: Algorithms, Models, and Performance Analysis. CRC Press, Inc. 2004.
    [2] M. Pinedo, Scheduling: Theory, Algorithms, and Systems, 4th ed., Boston, MA: Springer, New York, 2012.
    [3] D. Mourtzis, Internet based collaboration in the manufacturing supply chain, Cirp J. Manufact. Sci. Technol., 4 (2011), 296–304.
    [4] T. F. Gonzalez and S. Sahni, Open shop scheduling to minimize finish time, J. ACM, 23 (1976), 665–679.
    [5] D. Bai, Z. Zhang, Q. Zhang, et al., Open shop scheduling problem to minimize total weighted completion time, Eng. Optimiz., 49 (2017), 98–112.
    [6] J. Blazewicz, W. Domschke and E. Pesch, The job shop scheduling problem: Conventional and new solution techniques, Eur. J. Oper. Res., 93 (1996), 1–33.
    [7] R. L. Graham, E. L. Lawler, J. K. Lenstra, et al., Optimization and approximation in deterministic sequencing and scheduling: a survey, Annals discrete mathematics (1979), 287–326.
    [8] M. Pinedo, Scheduling: theory, algorithms, and systems. Prentice Hall, USA, 2002.
    [9] D. Bai and L. Tang, Open shop scheduling problem to minimize makespan with release dates, Appl. Math. Model., 37 (2013), 2008–2015.
    [10] E. L. Lawler, J. K. Lenstra, A. H. Kan, et al., Minimizing maximum lateness in a two-machine open shop, Math. Oper. Res., 6 (1981), 153–158.
    [11] R. Chen, W. Huang and G. Tang, Dense open-shop schedules with release times, Theor. Comput. Sci., 407 (2008), 389–399.
    [12] P. Brucker, J. L. Hurink, B. Jurisch, et al. A branch & bound algorithm for the open-shop problem, Discret Appl. Math., (1997), 43–59.
    [13] U. Dorndorf, P. Erwin and T. Phanhuy, Solving the open shop scheduling problem, J. Sched., 4 (2001), 157–174.
    [14] C. Liaw, A hybrid genetic algorithm for the open shop scheduling problem, Eur. J. Oper. Res., 124 (2000), 28–42.
    [15] F. Ahmadizar and M. H. Farahani, A novel hybrid genetic algorithm for the open shop scheduling problem, Int. J. Adv. Manuf. Technol., 62 (2012), 775–787.
    [16] X. Li, L. Gao, Q. Pan, et al., An Effective Hybrid Genetic Algorithm and Variable Neighborhood Search for Integrated Process Planning and Scheduling in a Packaging Machine Workshop, IEEE Trans. Syst. Man Cybern., (2018), 1–13.
    [17] X. Li, L. Gao, W. Wang, et al., Particle swarm optimization hybridized with genetic algorithm for uncertain integrated process planning and scheduling with interval processing time, Comput. Ind. Eng., (2019).
    [18] X. Li, S. Xiao, C. Wang, et al., Mathematical modeling and a discrete artificial bee colony algorithm for the welding shop scheduling problem, Memet. Comput., (2019), 1–19.
    [19] X. Li, C. Lu, L. Gao, et al., An Effective Multiobjective Algorithm for Energy-Efficient Scheduling in a Real-Life Welding Shop, IEEE Trans. Ind. Inform., 14 (2018), 5400–5409.
    [20] A. Barabasi and R. Albert, Emergence of Scaling in Random Networks, Science, 286 (1999), 509–512.
    [21] D. J. Watts and S. H. Strogatz, Collective dynamics of 'small-world' networks, Nature, 393 (1998), 440–442.
    [22] Q. Xuan and T. J. Wu, Network model and heuristic scheduling rule designing method for complex open shop problems, J. Zhejiang University Eng. Sci. Edition, 45 (2011), 961–968.
    [23] Q. Xuan and T. J. Wu, Open shop complex scheduling network model and characteristic analysis, J. Zhejiang University. Eng. Sci., 45 (2011), 589–595.
    [24] X. Li, Y. Yuan, W. Sun, et al. Bottleneck identification in job-shop based on network structure characteristic, Comput. Integr. Manuf., 4 (2016), 23.
    [25] D. H. Wolpert and W. G. Macready, No free lunch theorems for optimization, IEEE T. Evol. Comput., 1 (1997), 67–82.
    [26] A. S. Manne, On the job-shop scheduling problem, Oper. Res., 8 (1960), 219–223.
    [27] L. Lu, D. Chen, X. Ren, et al. Vital nodes identification in complex networks, Phys. Rep. Rev. Sec. Phys. Lett., 650 (2016), 1–63.
    [28] T. Bian and Y. Deng, Identifying influential nodes in complex networks: A node information dimension approach, Chaos, 28 (2018).
    [29] E. D. Taillard, Benchmarks for basic scheduling problems, Eur. J. Oper. Res., 64 (1993), 278–285.
    [30] G. C. Ciro, F. Dugardin, F. Yalaoui, et al., Open shop scheduling problem with a multi-skills resource constraint: a genetic algorithm and an ant colony optimisation approach, Int. J. Prod. Res., 54 (2016), 4854–4881.
    [31] A. Clauset, C. R. Shalizi and M. E. J. Newman, Power-law distributions in empirical data, SIAM Rev., 51 (2009), 661–703.
    [32] Y. S. Virkar, Power-Law Distributions and Binned Empirical data, Ann. Appl. Stat., 8 (2014), 89–119.
  • This article has been cited by:

    1. Ibrahim Yousif, Liam Burns, Fadi El Kalach, Ramy Harik, Leveraging computer vision towards high-efficiency autonomous industrial facilities, 2024, 0956-5515, 10.1007/s10845-024-02396-1
    2. Xin Cai, Bingpeng Gao, Xinyuan Nan, A collective neurodynamic approach to distributed resource allocation with event-triggered communication, 2024, 10, 2199-4536, 5071, 10.1007/s40747-024-01436-w
    3. Lu Chen, Fei Hao, Optimal tracking control for unknown nonlinear systems with uncertain input saturation: A dynamic event-triggered ADP algorithm, 2024, 564, 09252312, 126964, 10.1016/j.neucom.2023.126964
    4. Farhad Pourkamali-Anaraki, Jamal F. Husseini, Evan J. Pineda, Brett A. Bednarcyk, Scott E. Stapleton, Two-stage surrogate modeling for data-driven design optimization with application to composite microstructure generation, 2024, 138, 09521976, 109436, 10.1016/j.engappai.2024.109436
    5. Jinan Yang, Jialei Deng, Jiahou Zhao, Sujuan Jiao, Xinhua Long, A novel parallel multi-harmonic global multi-channel control algorithm for helicopter active vibration control, 2024, 142, 09670661, 105772, 10.1016/j.conengprac.2023.105772
    6. Yazhou Wang, Gang Wang, Huike Xu, Jianhui Liu, Zhen Wang, A prediction model of gear radial composite deviation based on digital twin mesh, 2025, 240, 02632241, 115619, 10.1016/j.measurement.2024.115619
    7. Qingyu Shi, Xia Huang, Bo Meng, Zhen Wang, Neural network-based iterative learning control for trajectory tracking of unknown SISO nonlinear systems, 2023, 232, 09574174, 120863, 10.1016/j.eswa.2023.120863
    8. Xiaogang Deng, Jiayan Li, Deep one-class classification model assisted by radius constraint for anomaly detection of industrial control systems, 2024, 138, 09521976, 109357, 10.1016/j.engappai.2024.109357
    9. Mustafa Demetgul, Qi Zheng, Ibrahim Nur Tansel, Jürgen Fleischer, Monitoring the misalignment of machine tools with autoencoders after they are trained with transfer learning data, 2023, 128, 0268-3768, 3357, 10.1007/s00170-023-12060-2
    10. Heng Zhao, Huanqing Wang, Xiaoheng Chang, Adil M. Ahmad, Xudong Zhao, Neural network-based adaptive critic control for saturated nonlinear systems with full state constraints via a novel event-triggered mechanism, 2024, 675, 00200255, 120756, 10.1016/j.ins.2024.120756
    11. Xinyi He, Chang Liu, Xiaodi Li, A practical leader–follower hybrid control scheme for wheeled mobile robots, 2024, 184, 09600779, 114954, 10.1016/j.chaos.2024.114954
    12. Shunyi Zhao, Zheng Zhou, Chengxi Zhang, Jin Wu, Fei Liu, Guangyi Shi, Localization of underground pipe jacking machinery: A reliable, real-time and robust INS/OD solution, 2023, 141, 09670661, 105711, 10.1016/j.conengprac.2023.105711
    13. Zishuo Dong, Xu Li, Feng Luan, Lingming Meng, Jingguo Ding, Dianhua Zhang, Fusion of theory and data-driven model in hot plate rolling: A case study of rolling force prediction, 2024, 245, 09574174, 123047, 10.1016/j.eswa.2023.123047
    14. Weizhen Wang, Xin Chen, Jiangbo Jia, Kaili Wu, Mingyang Xie, Optimal formation tracking control based on reinforcement learning for multi-UAV systems, 2023, 141, 09670661, 105735, 10.1016/j.conengprac.2023.105735
    15. Maria Letizia Corradini, Resilience enhancement to loss of actuator effectiveness in a Model-Free Adaptive framework, 2024, 361, 00160032, 106957, 10.1016/j.jfranklin.2024.106957
    16. Xinggui Zhao, Bo Meng, Zhen Wang, Event-triggered integral sliding mode control for uncertain networked linear control systems with quantization, 2023, 20, 1551-0018, 16705, 10.3934/mbe.2023744
    17. Yu Wan, Xuehui Gao, Extended-state-observer-based output feedback control for hydraulic systems with performance constraint, 2024, 112, 0924-090X, 18333, 10.1007/s11071-024-09957-2
    18. Yongwei Zhang, Shunchao Zhang, Integral sliding mode-based event-triggered optimal fault tolerant tracking control of continuous-time nonlinear systems, 2024, 79, 09473580, 101021, 10.1016/j.ejcon.2024.101021
    19. Yawu Wang, Zhichao Xu, Jundong Wu, Yue Zhang, Chun-Yi Su, Modelling and model-based tracking control of soft twisted and coiled actuators, 2023, 141, 09670661, 105722, 10.1016/j.conengprac.2023.105722
    20. Hao Shen, Ziwei Li, Jing Wang, Jinde Cao, Nonzero-sum games using actor-critic neural networks: A dynamic event-triggered adaptive dynamic programming, 2024, 662, 00200255, 120236, 10.1016/j.ins.2024.120236
    21. Youness Boutyour, Abdellah Idrissi, Dynamic confidence-based constraint adjustment in distributional constrained policy optimization: enhancing supply chain management through adaptive reinforcement learning, 2024, 0956-5515, 10.1007/s10845-024-02492-2
    22. Bo Huang, Sirui Zheng, Hamido Fujita, Jin Liu, A multi-task learning model for recommendation based on fusion of dynamic and static neighbors, 2024, 133, 09521976, 108190, 10.1016/j.engappai.2024.108190
    23. Haoming Zou, Guoshan Zhang, Zhiguo Yan, Wanquan Liu, Dynamic event-triggered finite-horizon robust suboptimal control of multi-player systems with input disturbances, 2025, 611, 09252312, 128665, 10.1016/j.neucom.2024.128665
    24. Héctor Escobar-Cuevas, Erik Cuevas, Jorge Gálvez, Karla Avila, A novel hybrid search strategy for evolutionary fuzzy optimization approach, 2024, 36, 0941-0643, 2633, 10.1007/s00521-023-09161-0
    25. Zongsheng Huang, Xiaoyang Gao, Tieshan Li, Yue Long, Hanqing Yang, Prescribed performance event-triggered fuzzy optimal tracking control for strict-feedback nonlinear systems, 2024, 658, 00200255, 120014, 10.1016/j.ins.2023.120014
    26. Huanqing Wang, Muxuan Li, Haikuo Shen, Dynamic event‐triggered adaptive neural nonsingular fixed‐time attitude control for multi‐UAVs systems, 2024, 38, 0890-6327, 3102, 10.1002/acs.3863
    27. Haoran Zhang, Chunhui Zhao, Jinliang Ding, Constrained Reinforcement Learning-Based Closed-Loop Reference Model for Optimal Tracking Control of Unknown Continuous-Time Systems, 2024, 21, 1545-5955, 7312, 10.1109/TASE.2023.3340726
    28. Naseem Ahmad, Shoulin Hao, Tao Liu, Yihui Gong, Qing-Guo Wang, Data-driven set-point learning control with ESO and RBFNN for nonlinear batch processes subject to nonrepetitive uncertainties, 2024, 146, 00190578, 308, 10.1016/j.isatra.2023.12.044
    29. Wei Xie, Gan Yu, David Cabecinhas, Carlos Silvestre, Weidong Zhang, Wei He, Robust collision-free formation control of quadrotor fleets: Trajectory generation and tracking with experimental validation, 2024, 145, 09670661, 105842, 10.1016/j.conengprac.2024.105842
    30. Paolo Righettini, Roberto Strada, Monica Tiboni, Filippo Cortinovis, Jasmine Santinelli, A systematic management and control methodology for high energy saving in applications equipped with hydraulic servo-axes, 2024, 145, 09670661, 105847, 10.1016/j.conengprac.2024.105847
    31. Ting Shi, Peng Shi, Jonathon Chambers, Dynamic Event-Triggered Model Predictive Control Under Channel Fading and Denial-of-Service Attacks, 2024, 21, 1545-5955, 6448, 10.1109/TASE.2023.3325534
    32. Chunping Xiong, Qian Ma, Guopeng Zhou, ADP‐based robust consensus for multi‐agent systems with unknown dynamics and random uncertain channels, 2024, 34, 1049-8923, 4051, 10.1002/rnc.7177
    33. Yanzhe Wang, Qian Yang, Weiwei Qu, A collision-free transition path planning method for placement robots in complex environments, 2024, 10, 2199-4536, 8481, 10.1007/s40747-024-01585-y
    34. Jinggao Sun, Huiyong Lu, Huaicheng Yan, Zhicheng Kou, Dynamic event‐triggered model‐free adaptive control for networked control systems with random packet loss, 2024, 1049-8923, 10.1002/rnc.7647
    35. Ziwen Gu, Yatao Shen, Zijian Wang, Jiayi Qiu, Wenmei Li, Chun Huang, Yaqun Jiang, Peng Li, Load forecasting model considering dynamic coupling relationships using structured dynamic-inner latent variables and broad learning system, 2024, 133, 09521976, 108180, 10.1016/j.engappai.2024.108180
    36. Jue Wang, Huihui Pan, Weichao Sun, Event-Triggered Adaptive Output Constraint Tracking Control of Uncertain MIMO Nonlinear Systems With Sensor and Actuator Faults, 2024, 21, 1545-5955, 6774, 10.1109/TASE.2023.3330966
    37. Lokesh Soni, Neeta Pandey, A Reliable and high performance Radiation Hardened Schmitt Trigger 12T SRAM cell for space applications, 2024, 176, 14348411, 155161, 10.1016/j.aeue.2024.155161
    38. Yuzhu Huang, Zhaoyan Zhang, Xiong Yang, Backstepping based neural H∞ optimal tracking control for nonlinear state constrained systems with input delay and disturbances, 2024, 595, 09252312, 127869, 10.1016/j.neucom.2024.127869
    39. Seyed Adel Alizadeh Kolagar, Alireza Taheri, Ali F. Meghdari, NAO Robot Learns to Interact with Humans through Imitation Learning from Video Observation, 2023, 109, 0921-0296, 10.1007/s10846-023-01938-8
    40. Deshuai Zheng, Jin Yan, Tao Xue, Yong Liu, A knowledge-based task planning approach for robot multi-task manipulation, 2023, 2199-4536, 10.1007/s40747-023-01155-8
    41. Nan Jiang, Qingping Xiang, Hongzhi Wang, Bo Zheng, Time series compression based on reinforcement learning, 2023, 648, 00200255, 119490, 10.1016/j.ins.2023.119490
    42. Chenye Hu, Jingyao Wu, Chuang Sun, Xuefeng Chen, Ruqiang Yan, Intelligent temporal detection network for boundary-sensitive flight regime recognition, 2023, 126, 09521976, 106949, 10.1016/j.engappai.2023.106949
    43. Ding Wang, Hongyu Ma, Jin Ren, Ning Gao, Junfei Qiao, Adaptive critic design with weight allocation for intelligent learning control of wastewater treatment plants, 2024, 133, 09521976, 108284, 10.1016/j.engappai.2024.108284
    44. Haixiu Xie, Yuanwei Jing, Yantong Liu, Jiqing Chen, Event-based adaptive fuzzy tracking control for nonlinear systems with guaranteed performance, 2023, 643, 00200255, 119267, 10.1016/j.ins.2023.119267
    45. Yuxue Li, Xiaoyuan Zhu, Guodong Yin, Robust actuator fault detection for quadrotor UAV with guaranteed sensitivity, 2023, 138, 09670661, 105588, 10.1016/j.conengprac.2023.105588
    46. Shuofeng Weng, Chaochun Yuan, Youguo He, Jie Shen, Long Chen, Lizhang Xu, Zhihao Zhu, Qiuye Yu, Zeyu Sun, Neural network energy management strategy for plug-in hybrid electric combine harvesters based on quasi-periodic samples, 2024, 136, 09521976, 109051, 10.1016/j.engappai.2024.109051
    47. Chunyang Sheng, Qinghui Wang, Tao Su, Haixia Wang, Induction motor torque closed-loop vector control system based on flux observation and harmonic current suppression, 2024, 142, 09670661, 105755, 10.1016/j.conengprac.2023.105755
    48. Zitao Chen, Kairui Chen, Ruizhi Tang, Optimal synchronization with L2-gain performance: An adaptive dynamic programming approach, 2024, 179, 08936080, 106566, 10.1016/j.neunet.2024.106566
    49. Haipeng Wang, Application of new features based on artificial intelligent robot technology in medium-scale urban design pedigree and intelligent management and control, 2024, 22, 26673053, 200379, 10.1016/j.iswa.2024.200379
    50. Qingyong Yang, Shu-Chuan Chu, Jeng-Shyang Pan, Jyh-Horng Chou, Junzo Watada, Dynamic multi-strategy integrated differential evolution algorithm based on reinforcement learning for optimization problems, 2024, 10, 2199-4536, 1845, 10.1007/s40747-023-01243-9
    51. Julio Yuzo Yassuda, Cristiano Marcos Agulhari, Emerson Ravazzi Pires da Silva, Sampled-data robust control of a 2-DoF helicopter modeled using a quasi-LPV framework, 2024, 145, 09670661, 105870, 10.1016/j.conengprac.2024.105870
    52. Zhenguo Zhang, Tianhao Ma, Yadan Zhao, Shuai Yu, Fan Zhou, Adaptive dynamic programming-based multi-fault tolerant control of reconfigurable manipulator with input constraint, 2024, 10, 2199-4536, 8341, 10.1007/s40747-024-01550-9
    53. Qing-xin Meng, Jian-wei Liu, Nonstationary online convex optimization with multiple predictions, 2024, 654, 00200255, 119862, 10.1016/j.ins.2023.119862
    54. Adam Zielonka, Andrzej Sikora, Marcin Woźniak, Fuzzy rules intelligent car real-time diagnostic system, 2024, 135, 09521976, 108648, 10.1016/j.engappai.2024.108648
    55. HaiTao Wang, XiangShuai Zhai, Tao Wen, ZiDu Yin, Yang Yang, Data-driven hierarchical learning approach for multi-point servo control of Pan–Tilt–Zoom cameras, 2024, 136, 09521976, 108987, 10.1016/j.engappai.2024.108987
    56. Antai Li, Datong Qin, Zheng Guo, Yu Xia, Chang Lv, Wet clutch pressure hysteresis compensation control under variable oil temperatures for electro-hydraulic actuators, 2023, 141, 09670661, 105723, 10.1016/j.conengprac.2023.105723
    57. Yuan Wang, Zhenbin Du, Yanming Wu, Pseudo-partial-derivative information-driven adaptive fault-tolerant tracking control for discrete-time systems, 2024, 10, 2199-4536, 2531, 10.1007/s40747-023-01280-4
    58. M. Tanhaeean, S.F. Ghaderi, M. Sheikhalishahi, A decision-making framework for optimal maintenance management: An integrated simulation-mathematical programming-expert system approach, 2023, 185, 03608352, 109671, 10.1016/j.cie.2023.109671
    59. Yue Wang, Yonghui Yang, Libing Wu, Adaptive fault-tolerant consensus control of multi-agent systems with event-triggered inputs, 2023, 650, 00200255, 119594, 10.1016/j.ins.2023.119594
    60. Guozeng Cui, Hui Xu, Jinpeng Yu, Qian Ma, Muwei Jian, Event-Triggered Fixed-Time Adaptive Fuzzy Control for Nontriangular Nonlinear Systems With Unknown Control Directions, 2024, 5, 2691-4581, 2397, 10.1109/TAI.2023.3318895
    61. Xiaolei Ji, Fei Hao, Distributed asynchronous event-triggered cooperative control for virtually coupled train set subject to gradient terrain and input saturation, 2023, 360, 00160032, 11809, 10.1016/j.jfranklin.2023.09.029
    62. Ding Wang, Hongyu Ma, Junfei Qiao, Multilayer adaptive critic design with digital twin for data-driven optimal tracking control and industrial applications, 2024, 133, 09521976, 108228, 10.1016/j.engappai.2024.108228
    63. Yuejie Yao, Yiping Luo, Jinde Cao, Finite-time guarantee-cost H∞ consensus control of second-order multi-agent systems based on sampled-data event-triggered mechanisms, 2024, 174, 08936080, 106261, 10.1016/j.neunet.2024.106261
    64. Yu Wan, Wenlong Yue, Xuehui Gao, Qiang Chen, Ruiyin Xu, Adaptive finite-time prescribed performance tracking control for hydraulic servo systems with friction compensation, 2024, 564, 09252312, 126967, 10.1016/j.neucom.2023.126967
    65. Bo Dong, Zhendong Ding, Tianjiao An, Yiming Cui, Xinye Zhu, Integral reinforcement learning-based event-triggered optimal tracking control for modular robot manipulators via non-zero-sum game, 2024, 35, 0957-0233, 096205, 10.1088/1361-6501/ad50f8
    66. Wenyan Ye, Ping Zhang, Haohsuan Chang, A data-driven optimal time-delayed control approach and its application to aerial manipulators, 2024, 142, 09670661, 105754, 10.1016/j.conengprac.2023.105754
    67. Onuchukwu Godwin Chike, Norhayati Ahmad, Wan Fahmin Faiz Wan Ali, Neural network prediction of thermal field spatiotemporal evolution during additive manufacturing: an overview, 2024, 134, 0268-3768, 2107, 10.1007/s00170-024-14256-6
    68. Quangui He, Wei Liu, Formation control for linear multi-agent systems with asynchronously sampled outputs, 2024, 658, 00200255, 119992, 10.1016/j.ins.2023.119992
    69. Amir Veisi, Hadi Delavari, Deep reinforcement learning optimizer based novel Caputo fractional order sliding mode data driven controller, 2025, 140, 09521976, 109725, 10.1016/j.engappai.2024.109725
    70. Shanke Li, Kun Peng, Fei Hui, Ziqi Li, Cheng Wei, Wenbo Wang, A Decision-Making Approach for Complex Unsignalized Intersection by Deep Reinforcement Learning, 2024, 73, 0018-9545, 16134, 10.1109/TVT.2024.3408917
    71. Chenhang Yan, Liping Yan, Yuezu Lv, Yuanqing Xia, Dynamic Event-Triggered Byzantine-Resilient Output Regulation in Continuous-Time High-Order Multiagent Systems With Static/Dynamic Leader, 2024, 71, 1549-8328, 5532, 10.1109/TCSI.2024.3381191
    72. Jianying Li, Hailong Yang, Hui Ji, Characterization of Two-Cylinder Parallel Electro-hydraulic Force/Position Synchronization Based on RBF Fuzzy Neural Network Control, 2024, 1562-2479, 10.1007/s40815-024-01846-5
    73. Xiong Yang, Qinglai Wei, Adaptive Dynamic Programming for Robust Event-Driven Tracking Control of Nonlinear Systems With Asymmetric Input Constraints, 2024, 54, 2168-2267, 6333, 10.1109/TCYB.2024.3418904
    74. Chengli Fan, Chunyi Xiao, Tao Xu, Dengxiu Yu, C.L. Philip Chen, Decentralized event-trigger-based predefined-time adaptive control of large-scale nonlinear systems, 2024, 00160032, 107446, 10.1016/j.jfranklin.2024.107446
    75. Weiqi Liu, Shuai Sui, C.L. Philip Chen, Event-triggered predefined-time output feedback fuzzy adaptive control of permanent magnet synchronous motor systems, 2025, 142, 09521976, 109882, 10.1016/j.engappai.2024.109882
    76. A. Aziz Khater, Mohamed Fekry, Mohammad El-Bardini, Ahmad M. El-Nagar, Deep reinforcement learning-based adaptive fuzzy control for electro-hydraulic servo system, 2025, 0941-0643, 10.1007/s00521-024-10741-x
    77. Wenhui Dou, Shihong Ding, Ju H. Park, Keqi Mei, An Adaptive Generalized Super-Twisting Algorithm via Event-Triggered Control, 2025, 22, 1545-5955, 393, 10.1109/TASE.2024.3351122
    78. Bo Dong, Yuhang Gao, Tianjiao An, Hucheng Jiang, Bing Ma, Nonzero-sum game-based decentralized approximate optimal control of modular robot manipulators with coordinate operation tasks using value iteration, 2025, 36, 0957-0233, 026209, 10.1088/1361-6501/ad880d
    79. Feng Qin, Azlan Mohd Zain, Kai-Qing Zhou, Norfadzlan Bin Yusup, Didik Dwi Prasetya, Rozita Abdul Jalil, Zaheera Zainal Abidin, Mahadi Bahari, Yusri Kamin, Mazlina Abdul Majid, Hybrid Harmony Search Algorithm Integrating Differential Evolution and Lévy Flight for Engineering Optimization, 2025, 13, 2169-3536, 13534, 10.1109/ACCESS.2025.3529714
    80. Suhuan Zhang, Fanglai Zhu, Xufeng Ling, Event-triggered UIO-based security control for discrete-time systems under deception attacks, 2025, 00200255, 121902, 10.1016/j.ins.2025.121902
    81. Yuebing Wen, Shuhua Teng, Qiang Li, Jianping Tan, Yuwei Song, Shiyuan Sun, Investigating the Symmetric Control of a Hydraulic System Based on Status Feedback, 2025, 17, 2073-8994, 246, 10.3390/sym17020246
    82. Guhui Li, Zidong Wang, Xingzhen Bai, Zhongyi Zhao, Hongli Dong, Event-Triggered Set-Membership Filtering for Active Power Distribution Systems Under Fading Channels: A Zonotope-Based Approach, 2025, 22, 1545-5955, 1139, 10.1109/TASE.2024.3360600
    83. Cong Guan, Tao Jiang, Yi-Chen Li, Zongzhang Zhang, Lei Yuan, Yang Yu, Constraining an Unconstrained Multi-agent Policy with offline data, 2025, 08936080, 107253, 10.1016/j.neunet.2025.107253
    84. Zhikai Yao, Xianglong Liang, Shuping Wang, Jianyong Yao, Model-Data Hybrid Driven Control of Hydraulic Euler–Lagrange Systems, 2025, 30, 1083-4435, 131, 10.1109/TMECH.2024.3390129
    85. Giseung Park, Whiyoung Jung, Seungyul Han, Sungho Choi, Youngchul Sung, Adaptive multi-model fusion learning for sparse-reward reinforcement learning, 2025, 09252312, 129748, 10.1016/j.neucom.2025.129748
  • Reader Comments
  • © 2019 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(5048) PDF downloads(498) Cited by(12)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog