
The FeldsteinHorioka1980 study on investment flows through the correlation of domestic saving and investment concluded that liberalization of capital markets does not necessarily lead to a movement of capital looking for a better allocation of resources, as classical theory would suggest. Ever since, literature has been prolific regarding this "puzzle", with arguments for and against this conclusion. This paper aims to analyze the issue from a different perspective. In recent years, the stock markets of Chile, Colombia, Mexico and Peru joined the Latin American Integrated Market through an agreement that allows investors in any of the participating markets to invest in the others as if they were investing locally. Compositional methods are used to assess the hypothesis of a potential flow of capital between markets generated by the creation of the joint market. First, cross-sectional methods for compositional data were used to test the hypothesis. As a result, it was not possible to find a change in the composition of the investment in the four markets produced by the creation of the joint market. Secondly, vector autoregressive models were estimated and tested for structural breaks in the parameters. However, these models were not found to be informative. In conclusion, it was not possible to reject the Felstein-Horioka hypothesis, supporting the idea that liberalization is not enough to generate capital flows between markets.
Citation: Juan David Vega Baquero, Miguel Santolino. Capital flows in integrated capital markets: MILA case[J]. Quantitative Finance and Economics, 2022, 6(4): 622-639. doi: 10.3934/QFE.2022027
[1] | Luca Franzoi, Riccardo Montalto . Time almost-periodic solutions of the incompressible Euler equations. Mathematics in Engineering, 2024, 6(3): 394-406. doi: 10.3934/mine.2024016 |
[2] | Lars Eric Hientzsch . On the low Mach number limit for 2D Navier–Stokes–Korteweg systems. Mathematics in Engineering, 2023, 5(2): 1-26. doi: 10.3934/mine.2023023 |
[3] | Piermarco Cannarsa, Rossana Capuani, Pierre Cardaliaguet . C1;1-smoothness of constrained solutions in the calculus of variations with application to mean field games. Mathematics in Engineering, 2019, 1(1): 174-203. doi: 10.3934/Mine.2018.1.174 |
[4] | Mirco Piccinini . A limiting case in partial regularity for quasiconvex functionals. Mathematics in Engineering, 2024, 6(1): 1-27. doi: 10.3934/mine.2024001 |
[5] | Youchan Kim, Seungjin Ryu, Pilsoo Shin . Approximation of elliptic and parabolic equations with Dirichlet boundary conditions. Mathematics in Engineering, 2023, 5(4): 1-43. doi: 10.3934/mine.2023079 |
[6] | Alessandra De Luca, Veronica Felli . Unique continuation from the edge of a crack. Mathematics in Engineering, 2021, 3(3): 1-40. doi: 10.3934/mine.2021023 |
[7] | Yuzhe Zhu . Propagation of smallness for solutions of elliptic equations in the plane. Mathematics in Engineering, 2025, 7(1): 1-12. doi: 10.3934/mine.2025001 |
[8] | Massimiliano Giona, Luigi Pucci . Hyperbolic heat/mass transport and stochastic modelling - Three simple problems. Mathematics in Engineering, 2019, 1(2): 224-251. doi: 10.3934/mine.2019.2.224 |
[9] | Sudarshan Tiwari, Axel Klar, Giovanni Russo . Interaction of rigid body motion and rarefied gas dynamics based on the BGK model. Mathematics in Engineering, 2020, 2(2): 203-229. doi: 10.3934/mine.2020010 |
[10] | Edgard A. Pimentel, Miguel Walker . Potential estimates for fully nonlinear elliptic equations with bounded ingredients. Mathematics in Engineering, 2023, 5(3): 1-16. doi: 10.3934/mine.2023063 |
The FeldsteinHorioka1980 study on investment flows through the correlation of domestic saving and investment concluded that liberalization of capital markets does not necessarily lead to a movement of capital looking for a better allocation of resources, as classical theory would suggest. Ever since, literature has been prolific regarding this "puzzle", with arguments for and against this conclusion. This paper aims to analyze the issue from a different perspective. In recent years, the stock markets of Chile, Colombia, Mexico and Peru joined the Latin American Integrated Market through an agreement that allows investors in any of the participating markets to invest in the others as if they were investing locally. Compositional methods are used to assess the hypothesis of a potential flow of capital between markets generated by the creation of the joint market. First, cross-sectional methods for compositional data were used to test the hypothesis. As a result, it was not possible to find a change in the composition of the investment in the four markets produced by the creation of the joint market. Secondly, vector autoregressive models were estimated and tested for structural breaks in the parameters. However, these models were not found to be informative. In conclusion, it was not possible to reject the Felstein-Horioka hypothesis, supporting the idea that liberalization is not enough to generate capital flows between markets.
Within numerous control systems, various measures are implemented to mitigate costs or curtail energy consumption. For some special cases, specific control values are deliberately maintained at zero for designated durations using handoff control strategies, thereby effectively minimizing the associated control costs. In [1], a novel control methodology was introduced to solve the maximum hands-off control problem characterized by L0 norm, which is to find the sparsest control among all permissible controls. This control methodology holds significant utility in minimizing electricity or fuel consumption, showcasing its practical relevance and applicability. In [1], within the car's start-stop system, the engine is automatically deactivated when the vehicle comes to a halt or operates at speeds below a predetermined threshold. This deliberate shutdown serves to diminish CO2 emissions and decrease fuel consumption as part of the system's efficiency measures. Handoff control demonstrates effectiveness within networked control systems as well [2].
Sparse optimization has emerged as a prominent subject of interest among scholars in recent years. The technology of sparse optimization has garnered extensive attention and exploration across various domains, including compressed sensing [3], image and signal processing [4,5], machine learning [6,7], Alzheimer's disease biomarkers [8], and other fields. Maximum hands-off control shares a relationship with sparsity, a characteristic pertinent to numerous optimization problems. In the realm of sparse optimization, sparsity refers to a condition where a substantial portion of elements within a matrix or vector assumes zero values. Leveraging sparsity offers the potential to economize on both time and compression expenses. Moreover, it serves as a mechanism to sift through extensive datasets, extracting pertinent information and thereby streamlining the complexity of the problem [9]. Within various optimization problems, the L0 norm of a vector commonly serves as a means to quantify and define sparsity [10,11]. Due to its non-convex and non-continuous nature, the L0 norm presents inherent complexity in analysis. Consequently, numerous scholars resort to exploring the variational property of the L0 norm and employing subdifferentiation techniques to address its intricacies. In [12], a technique known as convex relaxation was introduced as a substitute for the L0 norm, employing the more tractable L1 norm. This transformation enables the formulation of the problem in a linear programming paradigm, amenable to resolution using methods such as the interior point method or the simplex method. In [13,14], the approach revolves around the non-convex relaxation method, leveraging the Lp relaxation technique, where 0≤p≤1, to yield solutions with increased sparsity. In [15,16], a focus is directed toward the Difference of Convex functions (DC) algorithm. This method entails the representation of the L0 norm as a difference between two convex functions. Subsequently, leveraging the DC algorithm facilitates the resolution of the resultant relaxation model. In [17], by addressing the non-convex nature of the logarithmic function, a proposal was made to substitute the logarithm function for the L0 norm. In [18], the resolution involves addressing a sparse optimization problem that encapsulates two competing objectives: measurement error minimization and sparsity maximization. This problem is approached and solved through the utilization of a multi-objective evolutionary algorithm. In [19], an effective fault diagnosis framework was constructed by integrating L0 norm sparse constraint optimization with principal component analysis, aimed at mitigating the extent of sparsity. However, in [12,13,14,15,16,17,18,19], the sparse optimization problems do not take into account the constraints of the dynamical system. The focus of the study presented in [20] revolves around the L1 objective optimal control problem specifically tailored for linear discrete systems. This formulation leads to diverse sparse solutions based on the selection of distinct problem parameters. The study detailed in [21] introduces a methodical approach for synthesizing sparse optimal control structures governed by linear discrete systems. This method guarantees sparsity within solutions, enabling the specification of a predetermined count of zero elements within the control structure. In [22], the utilization of the Smith iteration and Altering-Direction-Implicit (ADI) iteration methods was explored for obtaining numerical solutions in a large sparse optimal control problem governed by linear discrete-time Riccati equations, leveraging Newton's method. In [23], provably optimal sparse solutions to overdetermined linear systems with non-negativity constraints were researched in a least-squares sense by implicit enumeration. Nevertheless, in [20,21,22,23], the sparse optimal control problem governed by the linear dynamical system failed to incorporate the constraints associated with the nonlinear dynamical system. The work presented in [24] delved into the examination of the Newton-like method of undetermined equations. This exploration led to the discovery of sparse solutions pertaining to sparse optimal control problems governed by an affine nonlinear system. The analysis of optimal control using the value function was conducted employing the dynamic programming method. In [25], a dynamic programming approach was introduced to effectively approximate the sparse optimal control problem governed by an affine nonlinear system in a numerical context. The maximum hands off control problem governed by the class of affine nonlinear systems was studied in [26]. Nonetheless, in [24,25,26], the nonlinear dynamical systems involved in the sparse optimal control problem are affine nonlinear systems, where the state variables and control inputs are separable. This paper delves into the investigation of the sparse optimal control problem governed by a general nonlinear dynamical system where the state variables and control inputs are inseparable.
The conventional approach to optimal control involves determining the most effective control strategy within specified constraints [27]. This strategy aims to either maximize or minimize the performance index. Typically, deriving an analytical solution for the optimal control problem governed by a nonlinear dynamic system proves challenging. Therefore, the resolution of such problems often necessitates numerical methods to obtain an effective solution [28]. The numerical solutions to optimal control problems can be divided into two solution strategies: first optimize then discretize (indirect methods), and first discretize then optimize (direct methods) [29]. For problems that do not contain inequality constraints, indirect methods derive the first-order necessary conditions for the optimal control problem, namely, the Euler-Lagrange equations that include the initial and boundary value conditions. The solution to this initial-boundary value problem mainly involves two types of methods: the control vector iteration method and methods such as multiple shooting and collocation [30]. The direct method encompasses techniques such as the direct collocation method and control parameterization [31,32]. Control parameterization involves discretizing solely the control function, approximating it using fixed basis functions within specific subintervals [33]. The coefficients within this linear combination of basis functions serve as the optimal decision variables. Typically, predetermined switching times govern the transitions between values for each control component, often evenly divided within specified intervals [34]. To enhance solution accuracy, the time range of the control function is frequently subdivided more finely, leading to a greater number of decision variables. However, this denser division amplifies computational costs. Simultaneously, to mitigate the need for extensive time range subdivisions, one must consider incorporating the switching times as additional decision variables [35]. The traditional time-scaling transformation operates by mapping variable switching times with fixed points within a redefined time horizon. This process yields an optimization problem wherein the revised switching times remain constant. The application of the time-scaling transformation finds extensive utility across diverse domains, including mixed integer programming [36] and singular optimal control [37]. Yet, in practical scenarios, simultaneous switching of time poses notable challenges [38]. The sequential adaptive switching time optimization technique (SASTOT) presented by [39], suggests that the interval for control switching times can vary, affording the flexibility to freely select the number of segments for each approximate control component. The method initiates by applying control parameterization and the time-scaling transformation to a single control component initially, leaving the remaining components untouched. Subsequently, within the newly introduced time range, the time-scaling transformation introduces a subsequent new time range. This iterative process continues until the final control component undergoes processing in a similar manner. This modification introduces a significant level of flexibility into the control strategy. Compared with the traditional time-scale transformation techniques, the SASTOT can accurately identify sparse positions for each control.
In this study, our focus lies on the maximum hands-off control problem governed by a nonlinear dynamical system (MHCPNDS) with a maximum hands-off control constraint characterized by the L0 norm. The optimization variables encompass the count of segments for each control component and the respective switching times for each of these components. The examination of the MHCPNDS poses challenges in obtaining an analytical solution. Hence, the adoption of a numerical solution becomes imperative. In practical scenarios, simultaneous switching of all control components is not optimal. Addressing this, this paper introduces the SASTOT, which allows for a flexible selection of the number of segments and switching times for each individual control element. The integration of control parameterization with the SASTOT serves as a transformative approach to addressing the MHCPNDS. The non-smooth term that involves the maximum operator is approximated through a smoothing function, and we further illustrate the convergence of this approximation technique. The attainment of a sparse solution for the MHCPNDS involves the utilization of a gradient-based algorithm. Empirical assessments through numerical experiments substantiate the efficacy of the algorithm put forth in the study.
The contributions of the paper are three-fold:
1) Distinguished from the above-mentioned sparse optimization problem, which is amenable to analytical solutions, we consider the sparse optimal control problem within a nonlinear dynamic system. Given the intrinsic complexity of the nonlinear dynamic system, pinpointing an analytical solution poses a significant challenge. Taking these into consideration, we propose a numerical solution method, based on the gradient formulae of the cost function (Theorems 5–8) and the gradient-based algorithm, without requiring the linearization of the nonlinear dynamical systems.
2) A smoothing function has been introduced to mitigate the roughness of the non-smooth term involving the maximum operator. Several significant theorems (Theorems 1–4) have been established, illustrating that the smoothing function effectively addresses the shortcomings associated with constraint qualification noncompliance.
3) Setting itself apart from the time-scaling transformation technique, this paper introduces the SASTOT method that enables flexible determination of the number of segments and the specific switching times for each control element individually.
The rest of the paper is organized as follows. In Section 2, we present the MHCPNDS. In Section 3, the MHCPNDS is transformed by using control parameterization and the SASTOT. In Section 4, we deal with the nonsmoothness of the objective function by using the smoothing technique. In Section 5, the gradient-based algorithm is used to solve the resulting smooth problem. In Section 6, numerical results are presented. In Section 7, we draw some conclusions and suggest some future research directions.
Let In be the set of {1,2,...,n}. For a continuous time control v:[0,T]→R, the Lp norm is defined by
‖v‖p≜(∫T0|v|pdt)1p,p∈[1,+∞). | (2.1) |
The L0 norm is defined by
‖v‖0≜qM(supp(v)), | (2.2) |
where qM is the Lebesgue measure and the set supp(v) is called a support set of v defined by
supp(v)={t∈[0,T]:v(t)≠0}. | (2.3) |
In some cases, it is possible to significantly reduce the control effort by keeping the control value at exactly zero over a time interval. Maximum hands-off control can be called sparse control, which means that the more zeros the control value is in unit time, the more sparse it is. It makes the time interval of the control completely zero.
In order to obtain the sparse solution of maximum hands-off control, we consider a nonlinear dynamical system defined by
{dx(t)dt=f(t,x(t),u(t)),t∈[0,T],x(0)=ξ,x(T)=0, | (2.4) |
where x(t):=(x1(t),...,xn(t))T∈Rn denotes the state vector at time t, u(t):=(u1(t),...,um(t))T∈Rm denotes control input vector at time t, ξ denotes the initial state value, T≥0 denotes a given terminal time, and f(t,x(t),u(t)) denotes a nonlinear function vector defined on Rn.
Let x(⋅|u) denote the solution of system (2.4) satisfying
x(t|u)≥0,t∈[0,T]. | (2.5) |
For all t on [0,T], there is also the following constraint on each component of the control input vector u(t)
maxi∈Im|ui(t)|≤1. | (2.6) |
A control input vector u(t)∈Rm that satisfies constraint (2.6) is called a candidate control input vector. Let L[T,ξ] be the set consisting of all candidate control input vectors.
Definition 1. The maximum hands-off control constraint characterized by the L0 norm is defined by
A0(u)≜1Tm∑i=1λi‖ui‖0≤μ, | (2.7) |
where λi≥0,i∈Im, are given weights, u∈L[T,ξ] denotes the admissible control, and μ is a small positive number.
Definition 1 is used to characterize the sparsity of maximum hands-off control. The cost functional is defined by
J(u)=π0(x(T|u))+∫T0ϑ(x(t|u))dt, | (2.8) |
where π0:Rm→R and ϑ:Rm→R are continuously differentiable functions.
Then, our maximum hands-off control problem can be formulated as follows.
Problem A: Given system (2.4), choose an admissible control u∈L[T,ξ] to minimize the cost functional defined in (2.8) subject to boundary condition (2.6) and maximum hands-off control constraint (2.7).
The L0 norm, owing to its non-convex and discontinuous characteristics, inherently introduces complexity into analytical procedures. The L0 norm can be well approximated by the L1 norm [20]. The maximum hands-off control constraint characterized by the L0 norm is approximated by Definition 2.
Definition 2. The maximum hands-off control constraint characterized by the L1 norm is defined by
A1(u)≜1Tm∑i=1λi‖ui‖1=1Tm∑i=1λi∫T0|ui|dt≤μ, | (3.1) |
where λi≥0,i∈Im, are given weights, u∈L[T,ξ] denotes the admissible control, and μ is a small positive number.
Definition 2 serves to delineate the computationally efficient sparsity of maximum hands-off control. By Definition 2, Problem A can be well approximated by the following Problem B.
Problem B: Given system (2.4), choose the control u∈L[T,ξ] to minimize the objective functional defined in (2.8) subject to boundary condition (2.6) and maximum hands-off control constraint (3.1).
Control parameterization stands as an effective methodology for resolving the optimal control problem, commonly approached through approximation via piecewise constant functions [29,40,41,42]. Increasing the level of detail in the time level partition leads to greater accuracy as it allows for a more intricate and nuanced representation of the time-dependent processes or phenomena under consideration [43]. Employing control parameterization enables the derivation of a finite-dimensional approximation for Problem B. Ultimately, the gradient-based algorithm is employed to resolve the resultant approximation problem.
The conventional time-scaling transformation necessitates simultaneous switching of all control components, a condition challenging to attain in practical applications [44]. Hence, the introduction of SASTOT emerges as a solution, combining both time-scaling transformation and control parameterization methodologies [39]. Within this methodology, each control component possesses the capability to adaptively adjust and independently select its switching time, allowing for diverse switching points across components. Empirical validation showcases a notable reduction in computational complexity alongside an improvement in the accuracy of calculations. For a clearer exposition of this approach, this article examines two distinct control inputs. Let ˜u(t)=[˜u1(t),˜u2(t)]T. Then, the dynamical system is
˜f(t,˜x(t),˜u1(t),˜u2(t))=˜f(t,˜x(t),˜u(t)). | (3.2) |
The interval [0,T] is divided into q1 subintervals [δl1−11,δl11], l1∈Iq1, where σ1:=[δ11,δ21,…,δq11]T are variable switching time vector, and
0=δ11≤δ21≤… ≤δq11=T. | (3.3) |
The control component ˜u1(t) can be approximated by
˜u1(t)≈˜uq11(t|δ1,σ1)=q1∑l1=1ηl11χ[δl1−11,δl11)(t), t∈[0,T],l1∈Iq1, | (3.4) |
where ηl11 is the piecewise constant function value on the l1th subinterval of the control component u1(t) satisfying
a1≤ηl11≤b1,l1∈Iq1, | (3.5) |
and χ[δl1−11,δl11)(t) is an indicator function defined by
χ[δl1−11,δl11)(t)={1,ift∈[δl1−11,δl11)),0,ift∉[δl1−11,δl11)). | (3.6) |
Then, we obtain the following system:
{d˜x(t)dt=q1∑l1=1˜f(t,˜x(t),ηl11,˜u2(t)), t∈[0,T],˜x(0)=ξ,˜x(T)=0. | (3.7) |
Let ˜x(⋅|η1,u2) denote the solution of system (3.7).
After applying the time-scaling transformation [45] to ˜u1(t) and mapping the variable switching times {δ01,δ11,…,δq11} to a fixed number of switching times {0,1,…,q1} in the new time horizon, we define the vector ϕ1:=[ϕ11,…,ϕq11]T∈Rq1, where ϕl11=δl11−δl1−11≥ρ, ρ is an extremely small positive number, l1∈Iq1, and ϕ11+ϕ21+⋯+ϕq11=T.
Then, we introduce a new time variable p and define a time scaling function ν1(p,ϕ1)
t(p)≜ν1(p,ϕ1)=⌊p⌋∑l1=1ϕl11+ϕ(⌊p⌋+1)1(p−⌊p⌋), p∈[0,q1], | (3.8) |
where θ(p)=˜u2(ν1(p,ϕ1))=˜u2(t) and ⌊p⌋ is a floor function of the time variable p. With the new time variable p, the dynamical system is redefined on the subinterval [l1−1,l1), l1∈Iq1,
{dˆx(p)dp=ϕl11ˆf(p,ˆx(p),ηl11,θ(p)),ˆx(0)=ξ,ˆx(T)=0. | (3.9) |
Let ˆx(⋅|η1,θ) denote the solution of system (3.9).
The interval [0,q1] is divided into q2 subintervals [pl2−12,pl22], l1∈Iq2, where σ2=[p02,p12,…,pq22]T is the variable switching time vector, and
0=p02≤p12≤… ≤pq22=T. | (3.10) |
The control component ˜u2(t) can be approximated by
˜u2(t)≈θ(p):=q2∑l2=1ηl22χ[pl2−12,pl22)(p),p∈[0,q1], | (3.11) |
where ηl22 is the piecewise constant function value on the l2th subinterval of the control component θ(p) satisfying
a2≤ηl22≤b2,l2∈Iq2. | (3.12) |
With the new time variable p, system (3.9) is redefined on the subinterval [pl2−12,pl22), l2∈Iq2,
{dˇx(p)dp=ϕ⌊p⌋1ˇf(p,ˇx(p),η⌊p⌋1,ηl22),ˇx(0)=ξ,ˇx(0)=0. | (3.13) |
After applying a time-scaling transformation to ˇu2(t) and mapping the variable switching times {δ02,δ12,…,δq22} to a fixed number of switching times {0,1,…,q2} in the new time horizon, we define the vector ϕ2:=[ϕ12,…,ϕq22]T∈Rq2, where ϕl22=δl22−δl2−12≥ρ, ρ is an extremely small positive number, l2∈Iq2, and ϕ12+ϕ22+⋯+ϕq22=T.
Then, we introduce a new time variable w and define a time scaling function ν2(w,ϕ2)
p(w)≜ν2(w,ϕ2)=⌊w⌋∑l2=1ϕl22+ϕ(⌊w⌋+1)2(w−⌊w⌋), w∈[0,q2], | (3.14) |
where ⌊w⌋ is the floor function of the time variable w. System (3.13) is redefined on the subinterval [l2−1,l2),l2∈Iq2,
{dˇx(w)dw=ϕl22ϕ⌊ν2(w)⌋1ˇf(w,ˇx(w),η⌊ν2(w)⌋1,ηl22),ˇx(0)=ξ,ˇx(0)=0. | (3.15) |
Let ˇx(⋅|η1,η2,ϕ1,ϕ2) be the solution of system (3.15). Then the continuous state constraint (2.5) becomes
ˇx(w|η1,η2,ϕ1,ϕ2)≥0,∀w∈[l2−1,l2), l2∈Iq2. | (3.16) |
The cost function (2.8) and the maximum hands-off control constraint (3.1) become
¯J(η1,η2,ϕ1,ϕ2)=π0(˜x(q2|η1,η2,ϕ1,ϕ2))+q2∑l2=1∫l2l2−1ϕl22ϕ⌊ν2(w)⌋1ϑ(˜x(w|η1,η2,ϕ1,ϕ2))dw, | (3.17) |
A2(η1,η2,ϕ1,ϕ2)=1q2q2∑l2=1∫l2l2−1ϕl22ϕ⌊ν2(w)⌋1[λ1|η⌊ν2(w)⌋1|+λ2|ηl22|]dw≤μ. | (3.18) |
With these in mind, Problem B can be approximated by the following Problem C.
Problem C: Given system (3.15), choose the quadruple (η1,η2,ϕ1,ϕ2)∈Rq1×Rq1×Rq2×Rq2 to minimize the objective function defined by (3.17) subject to continuous inequality constraints (3.16), boundary conditions (3.5) and (3.12), and maximum hands-off control constraint (3.18).
Because of |η⌊ν2(w)⌋1|=2max{η⌊ν2(w)⌋1,0}−η⌊ν2(w)⌋1,|ηl22|=2max{ηl22,0}−ηl22, the maximum hands-off control constraint (3.18) can be equivalently transformed as A3(η1,η2,ϕ1,ϕ2) defined in (3.19).
A3(η1,η2,ϕ1,ϕ2)=1q2q2∑l2=1∫l2l2−1ϕl22ϕ⌊ν2(w)⌋1[λ1(2max{η⌊ν2(w)⌋1,0}−η⌊ν2(w)⌋1)+λ2(2max{ηl22,0}−ηl22)]dw≤μ | (3.19) |
Define
J1(η1,η2,ϕ1,ϕ2,ρ)=¯J(η1,η2,ϕ1,ϕ2)+ρH(η1,η2,ϕ1,ϕ2), | (3.20) |
where ρ is the penalty parameter and
H(η1,η2,ϕ1,ϕ2)=q2∑l2=1∫l2l2−1max{−˜x(w|η1,η2,ϕ1,ϕ2),0}dw+q1∑l1=1[max{a1−ηl11,0}+max{ηl11−b1,0}]+q2∑l2=1[max{a2−ηl21,0}+max{ηl22−b2,0}]+A3(η1,η2,ϕ1,ϕ2). |
Thus, Problem C can be equivalently transformed into Problem D.
Problem D: Given system (3.15), choose the quadruple (η1,η2,ϕ1,ϕ2)∈Rq1×Rq1×Rq2×Rq2 to minimize the objective function defined in (3.20).
Remark 1. According to [29], it has been established that when the penalty parameter ρ exceeds its threshold value ρ∗, the solution derived for Problem D represents an exact solution for Problem C.
The fundamental concept underlying the smooth technique involves the approximation of the non-smooth maximum operator through the utilization of a smoothing function [46], which is defined by
P{G,ρ,q,ε}={0,ifG<−ερq,ρq2εG2+G+ε2ρq,if−ερq≤G<0,G+ε2ρq,ifG≥0, | (4.1) |
where ρ is a penalty factor, q is the number of continuous inequality constraints, and ε>0 is the smoothing parameter.
Theorem 1. ([46]) For ε>0, we have
0≤P{G,ρ,q,ε}−max{G,0}≤ε2ρq. | (4.2) |
By the smoothing process, the maximum hands-off control constraint A3(η1,η2,ϕ1,ϕ2) defined in (3.19) can be approximated by
˜A3(η1,η2,ϕ1,ϕ2,ρ,q2,ε)=1q2q2∑l2=1∫l2l2−1ϕl22ϕ⌊ν2(w)⌋1[λ1(2P{η⌊ν2(w)⌋1,ρ,q2,ε}−η⌊ν2(w)⌋1)+λ2(2P{ηl22,ρ,q2,ε}−ηl22)]dw. | (4.3) |
Based on the smoothing function (4.1), the cost function (3.20) can be approximated by
J2(η1,η2,ϕ1,ϕ2,ρ,q1,q2,ε)=¯J(η1,η2,ϕ1,ϕ2)+ρ˜H(η1,η2,ϕ1,ϕ2,ρ,q1,q2,ε), | (4.4) |
where
˜H(η1,η2,ϕ1,ϕ2,ρ,q1,q2,ε)=q2∑l2=1∫l2l2−1P{−˜x(w|η1,η2,ϕ1,ϕ2),ρ,q2,ε}dw+˜A3(η1,η2,ϕ1,ϕ2,ρ,q2,ε)+q2∑l2=1[P{a2−ηl22,ρ,q2,ε}+P{ηl22−b2,ρ,q2,ε}]+q1∑l1=1[P{a1−ηl11,ρ,q1,ε}+P{ηl11−b1,ρ,q1,ε}]. | (4.5) |
Based on the smoothing function (3.20), Problem D can be approximated by Problem E.
Problem E: Given system (3.15), choose the quadruple (η1,η2,ϕ1,ϕ2)∈Rq1×Rq1×Rq2×Rq2 to minimize the cost function defined in (4.4).
The application of smoothing techniques may introduce discrepancies between Problems D and E. This section entails the derivation of errors between Problems D and E concerning the smoothing function (3.20). It has been demonstrated that for a sufficiently small smoothing parameter ε, the solution to Problem D can be acquired by sequentially solving a series of Problem E while incrementing the values of the penalty factor ρ.
Theorem 2. If ρ>0, q>0, and ε>0, then
0≤J2(η1,η2,ϕ1,ϕ2,ρ,q1,q2,ε)−J1(η1,η2,ϕ1,ϕ2,ρ)≤5ε2+(λ1+λ2)T2εq2. | (4.6) |
Proof. Based on Theorem 1, we have for ρ>0, q1, q2>0, and ε>0,
0≤P{G,ρ,q,ε}−max{G,0}≤ε2ρq. |
Then,
0≤J2(η1,η2,ϕ1,ϕ2,ρ,q1,q2,ε)−J1(η1,η2,ϕ1,ϕ2,ρ)=ρ{q2∑l2=1∫l2l2−1(P{˜x(w|η1,η2,ϕ1,ϕ2),ρ,q2,ε}−max{˜x(w|η1,η2,ϕ1,ϕ2),0})dw+q1∑l1=1[(P{ηl11−a1,ρ,q1,ε}−max{ηl11−a1,0})+(P{b1−ηl11,ρ,q1,ε}−max{b1−ηl11,0})]+q2∑l2=1[(P{ηl22−a2,ρ,q2,ε}−max{ηl21−a2,0})+(P{b2−ηl22,ρ,q2,ε}−max{b2−ηl22,0})]+1q2q2∑l2=1∫l2l2−1ϕl22ϕ⌊ν2(w)⌋1[2λ1(P{η⌊ν2(w)⌋1,ρ,q2,ε}−max{η⌊ν2(w)⌋1,0})+2λ2(P{ηl22,ρ,q2,ε}−max{ηl22,0})]dw}≤ρ[5ε2ρ+2(λ1+λ2)T2ε2ρq2]=5ε2+(λ1+λ2)T2εq2, |
which completes the proof.
Theorem 3. Let (η∗1,η∗2,ϕ∗1,ϕ∗2) be the solution of Problem D, and let (η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2) be the solution of Problem E. Then,
0≤J2(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ,q1,q2,ε)−J1(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ)≤5ε2+(λ1+λ2)T2εq2. |
Proof. Based on Theorem 2, we have
0≤J2(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ,q1,q2,ε)−J1(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ)≤5ε2+(λ1+λ2)T2εq2,%(λ1+λ2)T2εc+5ε2cΛ−α0≤J2(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ,q1,q2,ε)−J1(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ)≤5ε2+(λ1+λ2)T2εq2. |
Since (η∗1,η∗2,ϕ∗1,ϕ∗2) is the solution of Problem D, we have
J1(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ)>J1(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ), |
which yields
J2(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ,q1,q2,ε)−J1(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ)≤J2(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ,q1,q2,ε)−J1(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ). | (4.7) |
Since (η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2) is the solution of Problem D, we obtain
J2(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ,q1,q2,ε)<J2(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ,q1,q2,ε), |
which yields
J2(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ,q1,q2,ε)−J1(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ)≤J2(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ,q1,q2,ε)−J1(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ). | (4.8) |
Based on (4.7) and (4.8), we have
0≤J2(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ,q,ε)−J1(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ)≤J2(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ,q,ε)−J1(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ)≤J2(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ,q,ε)−J1(η∗1,η∗2,ϕ∗1,ϕ∗2,ρ)≤5ε2+(λ1+λ2)T2εq2, |
which completes this proof.
Remark 2. Theorem 3 states that when the smoothing parameter ε assumes a sufficiently small value, the solution of Problem E tends to approximate the solution of Problem D.
To obtain an error estimation between the solutions of Problems E and C outlined in Theorem 4, the definition of ε-feasibility to Problem D is given in Definition 3.
Definition 3. A vector (ηε1,ηε2,ϕε1,ϕε2)∈Rq1×Rq1×Rq2×Rq2 is called ε-feasible to Problem D if
P{(ηl22)ε−a2,ρ,q2,ε}≤ε,P{b2−(ηl22)ε,ρ,q2,ε}≤ε,P{(ηl11)ε−a1,ρ,q1,ε}≤ε,P{b1−(ηl11)ε,ρ,q1,ε}≤ε,P{−˜x(w|ηε1,ηε2,ϕε1,ϕε2),ρ,q2,ε}≤ε,P{(η⌊ν2(w)⌋1)ε,ρ,q2,ε}≤ε,P{(ηl22)ε,ρ,q2,ε}≤ε. |
Theorem 4. Let (η∗1,η∗2,ϕ∗1,ϕ∗2) and (η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2) be the solution of Problems D and E, respectively. Furthermore, let (η∗1,η∗2,ϕ∗1,ϕ∗2) be feasible to Problem D and (η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2) be ε-feasible to Problem D. Then
−5ε2−(λ1+λ2)T2εq2≤¯J(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2)−¯J(η∗1,η∗2,ϕ∗1,ϕ∗2)≤5ε2+(λ1+λ2)T2εq2. | (4.9) |
Proof. Since (η∗1,η∗2,ϕ∗1,ϕ∗2) is the solution of Problem D, we have
H(η∗1,η∗2,ϕ∗1,ϕ∗2)=q2∑l2=1∫l2l2−1max{−˜x(w|η∗1,η∗2,ϕ∗1,ϕ∗2,0}dw+q1∑l1=1[max{a1−(ηl11)∗,0}+max{(ηl11)∗−b1,0}]+q2∑l2=1[max{a2−(ηl21)∗,0}+max{(ηl22)∗−b2,0}]+A3(η∗1,η∗2,ϕ∗1,ϕ∗2)=0, |
where
A3(η∗1,η∗2,ϕ∗1,ϕ∗2)=1q2q2∑l2=1∫l2l2−1(ϕl22)∗(ϕ⌊ν2(w)⌋1)∗[λ1(2max{(η⌊ν2(w)⌋1)∗,0}−(η⌊ν2(w)⌋1)∗)+λ2(2max{(ηl22)∗,0}−(ηl22)∗)]dw=0. |
Since (η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2) is ε-feasible to Problem D, we obtain
0≤ρ˜H(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ,q1,q2,ε)≤5ε2+(λ1+λ2)T2εq2. |
Based on Theorem 3, we get
0≤¯J(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2)+ρ˜H(η∗∗1,η∗∗2,ϕ∗∗1,ϕ∗∗2,ρ,q1,q2,ε)−¯J(η∗1,η∗2,ϕ∗1,ϕ∗2)−ρH(η∗1,η∗1,ϕ∗1,ϕ∗2)≤5ε2+(λ1+λ2)T2εq2, |
which completes this proof.
Remark 3. In this scenario, as outlined in Theorem 4, an error estimation between the solutions of Problems E and C is provided, specifically under the condition of ρ>ρ∗. Consequently, the solution for SOP (Sum of Optimization Problems) can be approximately obtained by iteratively solving a sequence of Problem E.
In this section, we will derive the gradient formulae of the cost function J2(η1,η2,ϕ1,ϕ2,ρ,q1,q2,ε) with respect to the quadruple (η1,η2,ϕ1,ϕ2)∈Rq1×Rq1×Rq2×Rq2 based on Theorems 5–8, whose proofs are similar to the proofs of Theorems 1 and 2 in [39].
Theorem 5. For each (η1,η2,ϕ1,ϕ2), we have
∂˜x(w|η1,η2,ϕ1,ϕ2)∂η1=M1(w|η1,η2,ϕ1,ϕ2), w∈[0,q2], | (5.1) |
where M1(⋅|η1,η2,ϕ1,ϕ2) is the solution to the following system on w\in[l_{2}-1, l_{2}], l_{2}\in I_{q_{2}}:
\begin{equation} \begin{array}{cc} \left \{\begin{array}{l} \dot{\boldsymbol{M}_{1}}(w) = \phi_{2}^{l_{2}}\phi_{1}^{\lfloor \nu_{2}(w) \rfloor}\bigg\{{ \frac{\partial\boldsymbol{f}(w, \hat{\boldsymbol{x}}(w), \eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, \eta_{2}^{l_{2}})}{\partial \tilde{\boldsymbol{x}}}}\boldsymbol{M}_{1}(w)+ \frac{\partial\boldsymbol{f}(w, \hat{\boldsymbol{x}}(w), \eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, \eta_{2}^{l_{2}})}{\partial {\boldsymbol{\eta_{1}}}} \bigg\}.\\ \boldsymbol{M}_{1}(0) = 0. \end{array} \right. & \\ \end{array} \end{equation} | (5.2) |
On the basis of Theorem 5, the gradient of J_{2} with respect to \boldsymbol{\eta_{1}} is given as follows:
\begin{eqnarray} \begin{split} \frac{\partial J_{2}}{\partial \boldsymbol{\eta_{1}}} & = \frac{\pi_{0}(\tilde{\boldsymbol{x}}(q_{2}|\boldsymbol {\eta_{1}, \eta_{2}}))}{\partial \tilde {\boldsymbol{x}}}\boldsymbol{M}_{1}(q_{2}) +\sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}}\phi_{2}^{l_{2}}\phi_{1}^{\lfloor \nu_{2}(w) \rfloor} \frac{\vartheta(\tilde{\boldsymbol{x}}(w|\boldsymbol {\eta_{1}, \eta_{2}}))}{\partial\tilde{\boldsymbol{x}}}\boldsymbol{M}_{1}(w) {\rm{d}}w\\ &+\rho\bigg[\sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}} \frac{P\left\{-\tilde{x}(w| \boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}), \rho, q_{2}, \varepsilon\right\}}{\partial \tilde {\boldsymbol{x}}}\boldsymbol{M}_{1}(w) {\rm{d}}w\\ &+\sum\limits_{l_{1} = 1}^{q_{1}} \frac{(P\left\{a_{1}-\eta_{1}^{l_{1}}, \rho, q_{1}, \varepsilon\right\}+P\left\{\eta_{1}^{l_{1}}-b_{1}, \rho, q_{1}, \varepsilon\right\})}{\partial \boldsymbol{\eta_{1}}}\\ &+\sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}} \frac{\phi_{2}^{l_{2}}\phi_{1}^{\lfloor \nu_{2}(w) \rfloor}}{q_{2}}\bigg\{ \frac{\partial[\lambda_{1}(2P\left\{\eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, c, q_{2}, \varepsilon\right\}-\eta_{1}^{\lfloor \nu_{2}(w) \rfloor})]}{\partial \boldsymbol{\eta_{1}}}\bigg\} {\rm{d}}w\bigg]. \end{split} \end{eqnarray} | (5.3) |
Theorem 6. For each (\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}) ,
\begin{eqnarray} {\frac{\partial \tilde{\boldsymbol{x}}(w|\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}})}{\partial \boldsymbol{\eta_{2}}}} = \boldsymbol{M}_{2}(w|\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}), \ w\in[0, q_{2}], \end{eqnarray} |
where \boldsymbol{M}_{2}(\cdot|\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}) is the solution to the following system on w\in[l_{2}-1, l_{2}], \ l_{2}\in I_{q_{2}}:
\begin{equation} \begin{array}{cc} \left \{\begin{array}{l} \dot{\boldsymbol{M}_{2}}(w) = \phi_{2}^{l_{2}}\phi_{1}^{\lfloor \nu_{2}(w) \rfloor}\bigg\{{ \frac{\partial\boldsymbol{f}(w, \hat{\boldsymbol{x}}(w), \eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, \eta_{2}^{l_{2}})}{\partial \tilde{\boldsymbol{x}}}}\boldsymbol{M}_{2}(w)+ \frac{\partial\boldsymbol{f}(w, \hat{\boldsymbol{x}}(w), \eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, \eta_{2}^{l_{2}})}{\partial {\boldsymbol{\eta_{2}}}} \bigg\}.\\ \boldsymbol{M}_{2}(0) = 0. \end{array} \right. & \\ \end{array} \end{equation} | (5.4) |
On the basis of Theorem 6, the gradient of J_{2} with respect to \boldsymbol{\eta_{2}} is given as follows:
\begin{eqnarray} \begin{split} \frac{\partial J_{2}}{\partial \boldsymbol{\eta_{2}}} & = \frac{\pi_{0}(\tilde{\boldsymbol{x}}(q_{2}|\boldsymbol {\eta_{1}, \eta_{2}}))}{\partial \tilde {\boldsymbol{x}}}\boldsymbol{M}_{2}(q_{2}) +\sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}}\phi_{2}^{l_{2}}\phi_{1}^{\lfloor \nu_{2}(w) \rfloor} \frac{\vartheta(\tilde{\boldsymbol{x}}(w|\boldsymbol {\eta_{1}, \eta_{2}}))}{\partial\tilde{\boldsymbol{x}}}\boldsymbol{M}_{2}(w) {\rm{d}}w\\ &+\rho\bigg[\sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}} \frac{P\left\{-\tilde{x}(w| \boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}), \rho, q_{2}, \varepsilon\right\}}{\partial \tilde {\boldsymbol{x}}}\boldsymbol{M}_{2}(w) {\rm{d}}w\\ &+\sum\limits_{l_{1} = 1}^{q_{1}} \frac{(P\left\{a_{2}-\eta_{2}^{l_{2}}, \rho, q_{2}, \varepsilon\right\}+P\left\{\eta_{2}^{l_{2}}-b_{2}, \rho, q_{2}, \varepsilon\right\})}{\partial \boldsymbol{\eta_{2}}}\\ &+\sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}} \frac{\phi_{2}^{l_{2}}\phi_{1}^{\lfloor \nu_{2}(w) \rfloor}}{q_{2}}\bigg\{ \frac{\partial[\lambda_{2}(2P\left\{\eta_{2}^{l_{2}}, \rho, q_{2}, \varepsilon\right\}-\eta_{2}^{l_{2}})]}{\partial \boldsymbol{\eta_{2}}}\bigg\} {\rm{d}}w\bigg]. \end{split} \end{eqnarray} |
Theorem 7. For each (\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}) , we have
\begin{eqnarray} {\frac{\partial \tilde{\boldsymbol{x}}(w|\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}})}{\partial \boldsymbol{\phi_{1}}}} = \boldsymbol{N}_{1}(w|\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}), \ w\in[0, q_{2}]{, } \end{eqnarray} |
where \boldsymbol{N}_{1}(\cdot|\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}) is the solution to the following system on w\in[l_{2}-1, l_{2}], \ l_{2}\in I_{q_{2}}:
\begin{equation} \begin{array}{cc} \left \{\begin{array}{l} \dot{\boldsymbol{N}_{1}}(w) = \phi_{2}^{l_{2}}\bigg\{{ \frac{\phi_{1}^{\lfloor \nu_{2}(w) \rfloor}\partial\boldsymbol{f}(w, \hat{\boldsymbol{x}}(w), \eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, \eta_{2}^{l_{2}})}{\partial \tilde{\boldsymbol{x}}}}\boldsymbol{N}_{1}(w)+ \frac{\phi_{1}^{\lfloor \nu_{2}(w) \rfloor}\partial\boldsymbol{f}(w, \hat{\boldsymbol{x}}(w), \eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, \eta_{2}^{l_{2}})}{\partial {\boldsymbol{\phi_{1}}}} \bigg\}.\\ \boldsymbol{N}_{1}(0) = 0. \end{array} \right. & \\ \end{array} \end{equation} | (5.5) |
Based on Theorem 7, the gradient of J_{2} with respect to \boldsymbol{\phi_{1}} is given as follows:
\begin{eqnarray} \begin{split} \frac{\partial J_{2}}{\partial \boldsymbol{\phi_{1}}} & = \frac{\pi_{0}(\tilde{\boldsymbol{x}}(q_{2}|\boldsymbol {\eta_{1}, \eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}))}{\partial \boldsymbol{\phi_{1}}}\boldsymbol{N}_{1}(q_{2})+ \sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}}\phi_{2}^{l_{2}} \frac{\phi_{1}^{\lfloor \nu_{2}(w) \rfloor}\vartheta(\tilde{\boldsymbol{x}}(w|\boldsymbol {\eta_{1}, \eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}))}{\partial \boldsymbol{\phi_{1}}}\boldsymbol{N}_{1}(w) {\rm{d}}w\\ &+\rho\bigg[\sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}} \frac{P\left\{-\tilde{x}(w| \boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}), \rho, q_{2}, \varepsilon\right\}}{\partial \boldsymbol{\phi_{1}}}\boldsymbol{N}_{1}(w) {\rm{d}}w\\ &+\sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}} \frac{\phi_{2}^{l_{2}}}{q_{2}}\bigg\{ \frac{\phi_{1}^{\lfloor \nu_{2}(w) \rfloor}\partial[\lambda_{1}(2P\left\{\eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, \rho, q_{2}, \varepsilon\right\}-\eta_{1}^{\lfloor \nu_{2}(w) \rfloor})+\lambda_{2}(2P\left\{\eta_{2}^{l_{2}}, \rho, q_{2}, \varepsilon\right\}-\eta_{2}^{l_{2}})}{\partial \boldsymbol{\phi_{1}}}\bigg\} {\rm{d}}w\bigg]. \end{split} \end{eqnarray} |
Remark 4. For this, we introduce the inverse of \nu_{2}(w) . The gradient of J_{2} with respect to \boldsymbol{\phi}_{2} is complicated to derive because the switching time is in the time range of w . \nu_{2}^{-1}(p) is a function of \boldsymbol{\phi}_{2} , and at every connection point, this inverse function is not differentiable. \nu_{2}^{-1}(r), r\in I_{q_{1}} is represented by w_{r} , and the gradient of w_{r} with respect to \phi_{2}^{l_{2}}, \ l_{2}\in I_{q_{2}} at r\in I_{q_{1}-1} is given in Remark 1 as follows.
Remark 5. If point w_{r} coincide with w\in I_{q_{2}-1}, then we have
\begin{eqnarray} \frac{\partial w_{r}^{-}}{\partial \phi_{2}^{l_{2}}}: = \begin{array}{cc} \left \{\begin{array}{l} 0 \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; {\rm{if}}\; l_{2} > e, \\ (-p+\sum\limits_{j = 1}^{e-1}\phi_{2}^{j})/(\phi_{2}^{e})^2, \; \; \; \; \; {\rm{if}}\; l_{2} = e, \\ - \frac{1}{\phi_{2}^{e+1}}, \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; {\rm{if}}\; l_{2} < e, \end{array} \right. & \\ \end{array} \end{eqnarray} | (5.6) |
and
\begin{eqnarray} \frac{\partial w_{r}^{+}}{\partial \phi_{2}^{l_{2}}}: = \begin{array}{cc} \left \{\begin{array}{l} 0 \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; {\rm{if}}\; l_{2} > e+1, \\ (-p+\sum\limits_{j = 1}^{e}\phi_{2}^{j})/(\phi_{2}^{e+1})^2, \; \; \; \; \; {\rm{if}}\; l_{2} = e+1, \\ - \frac{1}{\phi_{2}^{e+1}} \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; \; {\rm{if}}\; l_{2} < e+1. \end{array} \right. & \\\end{array} \end{eqnarray} | (5.7) |
Remark 6. If point w_{h} does not coincide with w\in I_{q_{2}-1} , then we have \frac{\partial w_{r}^{+}}{\partial \phi_{2}} = \frac{\partial w_{r}^{-}}{\partial \phi_{2}} .
With the above discussion, the gradient of states with respect to \boldsymbol{\phi_{2}} is given below.
Theorem 8. For each (\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}) , we have
\begin{eqnarray} {\frac{\partial \tilde{\boldsymbol{x}}(w|\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}})}{\partial \boldsymbol{\phi_{2}}}} = \boldsymbol{N}_{2}(w|\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}), \ w\in[0, q_{2}], \end{eqnarray} |
where \boldsymbol{N}_{2}(\cdot|\boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}) is the solution to the following system on w\in[l_{2}-1, l_{2}], \ l_{2}\in I_{q_{2}}:
\begin{equation} \begin{array}{cc} \left \{\begin{array}{l} \dot{\boldsymbol{N}_{2}}(w) = \phi_{1}^{\lfloor \nu_{2}(w) \rfloor}\bigg\{{ \frac{\phi_{2}^{l_{2}}\partial\boldsymbol{f}(w, \hat{\boldsymbol{x}}(w), \eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, \eta_{2}^{l_{2}})}{\partial \tilde{\boldsymbol{x}}}}\boldsymbol{N}_{2}(w)+ \frac{\phi_{2}^{l_{2}}\partial\boldsymbol{f}(w, \hat{\boldsymbol{x}}(w), \eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, \eta_{2}^{l_{2}})}{\partial {\boldsymbol{\phi_{2}}}} \bigg\}, \\ \boldsymbol{N}_{2}(w_{r}^{+}) = \boldsymbol{N}_{2}(w_{r}^{-})-H_{r, \boldsymbol{f}, \boldsymbol{h}}^{+}+H_{r, \boldsymbol{f}, \boldsymbol{h}}^{-}, \\ \boldsymbol{N}_{2}(0) = 0, \\ \end{array} \right. & \\ \end{array} \end{equation} | (5.8) |
with
H_{r, \boldsymbol{f}, \boldsymbol{h}}^{-} = \phi_{1}^{\lfloor \nu_{2}(w_{h}^{-}) \rfloor}\phi_{2}^{\lfloor w_{h}^{-} \rfloor} \Big[\boldsymbol{f}(\tilde{\boldsymbol{x}}(w_{h}^{-}))+h_{1}(\tilde{\boldsymbol{x}}(w_{h}^{-}))\eta_{1}^{\lfloor \nu_{2}(w_{h}^{-}) \rfloor}+ h_{2}(\tilde{\boldsymbol{x}}(w_{h}^{-}))\eta_{2}^{\lfloor w_{h}^{-} \rfloor}\Big] \frac{\partial w_{r}^{-}}{\partial \phi_{2}} and
H_{r, \boldsymbol{f}, \boldsymbol{h}}^{+} = \phi_{1}^{\lfloor \nu_{2}(w_{h}^{+}) \rfloor}\phi_{2}^{\lfloor w_{h}^{+} \rfloor} \Big[\boldsymbol{f}(\tilde{\boldsymbol{x}}(w_{h}^{+}))+h_{1}(\tilde{\boldsymbol{x}}(w_{h}^{+}))\eta_{1}^{\lfloor \nu_{2}(w_{h}^{+}) \rfloor}+ h_{2}(\tilde{\boldsymbol{x}}(w_{h}^{+}))\eta_{2}^{\lfloor w_{h}^{+} \rfloor}\Big] \frac{\partial w_{r}^{+}}{\partial \phi_{2}} .
Based on Theorem 8, the gradient of J_{2} with respect to \boldsymbol{\phi_{2}} is given as follows:
\begin{eqnarray} \begin{split} \frac{\partial J_{2}}{\partial \boldsymbol{\phi_{2}}} & = \frac{\pi_{0}(\tilde{\boldsymbol{x}}(q_{2}|\boldsymbol {\eta_{1}, \eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}))}{\partial \boldsymbol{\phi_{2}}}\boldsymbol{N}_{2}(q_{2})+ \sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}}\phi_{2}^{l_{2}} \frac{\phi_{1}^{\lfloor \nu_{2}(w) \rfloor}\vartheta(\tilde{\boldsymbol{x}}(w|\boldsymbol {\eta_{1}, \eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}))}{\partial \boldsymbol{\phi_{2}}}\boldsymbol{N}_{2}(w) {\rm{d}}w\\ &+\rho\bigg[\sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}} \frac{P\left\{-\tilde{x}(w| \boldsymbol{\eta_{1}}, \boldsymbol{\eta_{2}}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}), \rho, q_{2}, \varepsilon\right\}}{\partial \boldsymbol{\phi_{2}}}\boldsymbol{N}_{2}(w) {\rm{d}}w\\ &+\sum\limits_{l_{2} = 1}^{q_{2}} \int_{l_{2}-1}^{l_{2}} \frac{\phi_{1}^{\lfloor \nu_{2}(w) \rfloor}}{q_{2}}\bigg\{ \frac{\phi_{2}^{l_{2}}\partial[\lambda_{1}(2P\left\{\eta_{1}^{\lfloor \nu_{2}(w) \rfloor}, \rho, q_{2}, \varepsilon\right\}-\eta_{1}^{\lfloor \nu_{2}(w) \rfloor})+\lambda_{2}(2P\left\{\eta_{2}^{l_{2}}, \rho, q_{2}, \varepsilon\right\}-\eta_{2}^{l_{2}})}{\partial \boldsymbol{\phi_{2}}}\bigg\} {\rm{d}}w\\ &-H_{r, \boldsymbol{f}, \boldsymbol{h}}^{+}+H_{r, \boldsymbol{f}, \boldsymbol{h}}^{-}\bigg]. \end{split} \end{eqnarray} |
where H_{r, \lambda_{i}}^{+} and H_{r, \lambda_{i}}^{-} are defined as H_{r, \boldsymbol{f}, \boldsymbol{h}}^{+} and H_{r, \boldsymbol{f}, \boldsymbol{h}}^{-} , respectively.
Remark 7. The gradient formulae of the cost function J_{2}(\boldsymbol {\eta_{1}, \eta_{2}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}}}, \rho, q_1, q_2, \varepsilon) with respect to the quadruple (\boldsymbol {\eta_{1}, \eta_{2}, \boldsymbol{\phi_{1}}, \boldsymbol{\phi_{2}} })\in \mathbb{R}^{q_{1}}\times \mathbb{R}^{q_{1}}\times \mathbb{R}^{q_{2}}\times \mathbb{R}^{q_{2}} is calculated by the variational method (Theorems 5–8). Therefore, the gradient-based algorithm easily obtains the optimal solution for Problem E.
We consider a nonlinear maximum hands-off control problem with the cost functional defined by
\begin{eqnarray} \underset{i}{\min}\, h_{0} = \int_{0}^{1}\left\{-6x_{1}^2(t)-12x_{2}(t)+3u_{1}(t)+u_{2}(t)\right\} {\rm{d}}t, \end{eqnarray} | (6.1) |
governed by the nonlinear dynamical system
\begin{eqnarray} \begin{array}{cc} \left \{\begin{array}{l} \dot{x}_{1}(t) = u_{2}(t), \\ \dot{x}_{2}(t) = -{x}_{1}^2(t)+{u}_{1}(t), t\in [0, 1], \\ \boldsymbol{x} = (1, 0)^{\mathrm{T}}. \end{array} \right. & \\ \end{array} \end{eqnarray} | (6.2) |
with the maximum hands-off control constraint
\begin{equation} A_1(u_1, u_2) = \sum\limits_{i = 1}^{2}\lambda_{i} \int_{0}^{1}|u_{i}(t)| {\rm{d}}t \leq \mu = 10, \end{equation} | (6.3) |
and the bound constraints of the control inputs
\begin{equation} \label{exampleboundconstraint} -1\leq {u}_{j}(t)\leq 1, t\in [0, 1], j\in I_2. \nonumber \end{equation} |
Case 1: The optimal control problem governed by nonlinear dynamical system (6.2) without the maximum hands-off control constraint (6.3).
Case 2: The optimal control problem governed by nonlinear dynamical system (6.2) with the maximum hands-off control constraint (6.3).
For Cases 1 and 2, the number of segments for the two controls are, respectively, 10 and 20. For Cases 1 and 2, by using the proposed method we respectively obtain the optimal control strategies {\mathbf u}^{1*} and {\mathbf u}^{2*} plotted in Figures 1 and 2. Their corresponding optimal cost and sparsity levels are given in Table 1. From Figure 2, it is observed that for the first control, sparse control happened in the first and from the fourth to the tenth segments. For the second control, sparse control occurred in the second, the sixteenth, from the ninth to the thirteenth, and from the eighteenth to the twentieth segments.
Type | Control strategies | Optimal cost | A_1(u_1, u_2) |
Case 1 | {\mathbf u}^{1*} | h_0({\mathbf u}^{1*})=0.3392 | A_1({\mathbf u}^{1*}) = 21.2052 |
Case 2 | {\mathbf u}^{2*} | h_0({\mathbf u}^{2*})=0.3737 | A_1({\mathbf u}^{2*}) = 7.303 |
The graphical representation in Figure 1 illustrates that the optimal control strategies demonstrate a considerable density. In contrast, Figure 2 portrays these strategies with a noticeable level of sparsity. Nevertheless, it is essential to highlight that the inclusion of the maximum hands-off control constraint (6.3) results in a marginal increase in the cost function value compared to the scenario without this constraint (6.3). Specifically, the cost function slightly exceeds its counterpart when the maximum hands-off control constraint is imposed.
The observed trend in Figure 2 reveals a noteworthy pattern in the sparsity of the sparse optimal control strategies. It becomes evident that there is a rapid initial increase in sparsity, accompanied by a relatively minor escalation in cost. Therefore, based on the findings presented in this paper, it can be inferred that the proposed method exhibits the capability to generate solutions of superior quality.
Furthermore, from Table 1, it is pertinent to note that the value of the cost function, specifically h_0({\mathbf u}^{2*}) , slightly exceeds that of h_0({\mathbf u}^{1*}). The smaller the value of A_1({\mathbf u}^{*}) , the sparser {\mathbf u}^{*} becomes. From Table 1, it follows that {\mathbf u}^{2*} is sparser than {\mathbf u}^{1*} . So, we can conclude that this proposed method achieves a balance between system performance and sparsity, thereby suggesting its efficacy in providing solutions that optimize both aspects effectively.
Solving the sparse optimal control problem within the framework of linear dynamical systems often allows for an analytical solution. However, when addressing the maximum hands-off control problem within the domain of nonlinear dynamical systems, obtaining an analytical solution proves to be considerably challenging. This paper necessitates the numerical solution of the sparse optimal control problem which is governed by nonlinear dynamical systems. We employ control parameterization in conjunction with the sequential adaptive switching time optimization technique to approximate the maximum hands-off control problem by a sequence of finite-dimensional optimization problem. This approach allows for varying control switching times without imposing uniformity and offers flexibility in selecting control components. The resolution of the sparse optimal control problem relies on a gradient-based algorithm. To showcase the efficacy of the proposed methodology, an illustrative example is provided.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
This work was supported in part by the National Key Research and Development Program of China under Grant 2022YFB3304600; in part by the National Natural Science Foundation of China under Grants 11901075, 12271307 and 12161076; in part by the China Postdoctoral Science Foundation under Grant 2019M661073; in part by the Fundamental Research Funds for the Central Universities under Grants 3132022201, 3132023206, 3132023535 and DUT22LAB305; in part by the Guandong Province Natural Science Foundation of China under Grant 2022A1515011761; in part by Chongqing Natural Science Foundation Innovation and Development Joint Fund (CSTB2022NSCQ-LZX0040); in part by Ministry of Higher Education (MoHE) Malaysia through the Foundation Research Grant Scheme (FRGS/1/2021/STG06/SYUC/03/1); in part by the Shandong Province Natural Science Foundation of China under Grant ZR2023MA054; in part by the Fundamental Scientific Research Projects of Higher Education Institutions of Liaoning Provincial Department of Education under JYTMS20230165 (General Project); and in part by the Xinghai Project of Dalian Maritime University.
The authors declare there is no conflict of interest.
[1] | Aitchison J (1986) The statistical analysis of compositional data. Chapman & Hall, London. |
[2] |
Belles-Sampera J, Guillen M, Santolino M (2016) Compositional methods applied to capital allocation problems. J Risk 19: 1–15. https://doi.org/10.21314/JOR.2016.345 doi: 10.21314/JOR.2016.345
![]() |
[3] | Bellod-Redondo JF (1996) Ahorro e inversión en el largo plazo: El caso de la América Latina. El Trimest Econ 43: 1113–1137. |
[4] |
Boonen TJ, Guillen M, Santolino M (2019) Forecasting compositional risk allocations. Insur Math Econ 84: 79–86. https://doi.org/10.1016/j.insmatheco.2018.10.002 doi: 10.1016/j.insmatheco.2018.10.002
![]() |
[5] |
Chen S, Dong H (2020) Dynamic network connectedness of Bitcoin markets: Evidence from realized volatility. Front Phys 8: 582817. https://doi.org/10.3389/fphy.2020.582817 doi: 10.3389/fphy.2020.582817
![]() |
[6] |
Coakley J, Kulasi F, Smith S (1996) Current account solvency and the Feldstein-Horioka puzzle. Econ J 106: 620–627. https://doi.org/10.2307/2235567 doi: 10.2307/2235567
![]() |
[7] | Comas-Cufí M, Martín-Fernández JA, Mateu-Figueras (2016) Log-ratio methods in mixture models for compositional data sets. SORT-STAT OPER RES T 1: 349–374. |
[8] |
Dong H, Chen L, Zhang X, et al. (2020) The asymmetric effect of volatility spillover in global virtual financial asset markets: The case of Bitcoin. Emerg Mark Finance Trade 56: 1293–1311. https://doi.org/10.1080/1540496X.2019.1671819 doi: 10.1080/1540496X.2019.1671819
![]() |
[9] |
Drakos AA, Kouretas GP, Stavroyiannis S, et al. (2017) Is the Feldstein-Horioka puzzle still with us? national saving-investment dynamics and international capital mobility: A panel data analysis across EU member countries. J INT FINANC MARK I 47: 76–88. https://doi.org/10.1016/j.intfin.2016.11.006 doi: 10.1016/j.intfin.2016.11.006
![]() |
[10] |
Egozcue J, Pawlowsky-Glahn V, Mateu-Figueras G, et al. (2003) Isometric logratio transformations for compositional data analysis. Math Geol 35: 279–300. https://doi.org/10.1023/A:1023818214614 doi: 10.1023/A:1023818214614
![]() |
[11] |
Feldstein M, Horioka C (1980) Domestic saving and international capital flows. Econ J 90: 314–329. https://doi.org/10.2307/2231790 doi: 10.2307/2231790
![]() |
[12] |
Ford N, Horioka C (2017) The 'real' explanation of the Feldstein-Horioka puzzle. Appl Econ Lett 24: 95–97. https://doi.org/10.1080/13504851.2016.1164814 doi: 10.1080/13504851.2016.1164814
![]() |
[13] |
Fouquau J, Hurlin C, Rabaud I (2008) The Feldstein-Horioka puzzle: A panel smooth transition regression approach. Econ Model 25: 284–299. https://doi.org/10.1016/j.econmod.2007.06.008 doi: 10.1016/j.econmod.2007.06.008
![]() |
[14] | Fuller WA (1996) Introduction to Statistical Time Series. John Wiley and Sons, New York. |
[15] |
Granger CWJ (1969) Investigating causal relations by econometric models and cross-spectral methods. Econometrica 37: 424–438. https://doi.org/10.2307/1912791 doi: 10.2307/1912791
![]() |
[16] |
Ibarra-Yunez A (2008) Capital mobility and mexico's challenge toward financial integration. Lat Am Econ Rev 9: 256–27. https://doi.org/10.1080/10978520902873250 doi: 10.1080/10978520902873250
![]() |
[17] |
Jia S, Dong H, Yang H (2021) Asymmetric risk spillover of the international crude oil market in the perspective of crude oil dual attributes. Front Environ Sci 9: 720278. https://doi.org/10.3389/fenvs.2021.720278 doi: 10.3389/fenvs.2021.720278
![]() |
[18] | Martín-Fernádez JA, Bren M, Barcelo-Vidal C, et al. (1999) A measure of difference for compositional data based on measures of divergence. |
[19] | McLachlan G, Peel D (2004) Finite Mixture Models. Wiley Series in Probability and Statistics, John Wiley and Sons, New York. |
[20] |
Narayan P (2005) The saving and investment nexus for China: Evidence from cointegration tests. Appl Econ 37: 1979–1990. https://doi.org/10.1080/00036840500278103 doi: 10.1080/00036840500278103
![]() |
[21] | Pawlowsky-Glahn V, Egozcue JJ, Tolosana-Delgado R (2007) Lecture notes on compositional data analysis. |
[22] |
Said SE, Dickey DA (1984) Testing for unit roots in autoregressive-moving average models of unknown order. Biometrika 71: 599–607. https://doi.org/10.1093/biomet/71.3.599 doi: 10.1093/biomet/71.3.599
![]() |
[23] |
Sinha D, Sinha T (1998) An exploration of the long-run relationship between saving and investment in the developing economies: A tale of latin american countries. J Post Keynesian Econ 20: 435–443. https://doi.org/10.1080/01603477.1998.11490162 doi: 10.1080/01603477.1998.11490162
![]() |
[24] | Thomas P, Lovell D (2014) Compositional data analysis (CoDA) approaches to distance in information retrieval. In: Geva, S., et al., The 37th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR '14, Gold Coast, QLD, Australia, 991–994. https://doi.org/10.1145/2600428.2609492 |
Type | Control strategies | Optimal cost | A_1(u_1, u_2) |
Case 1 | {\mathbf u}^{1*} | h_0({\mathbf u}^{1*})=0.3392 | A_1({\mathbf u}^{1*}) = 21.2052 |
Case 2 | {\mathbf u}^{2*} | h_0({\mathbf u}^{2*})=0.3737 | A_1({\mathbf u}^{2*}) = 7.303 |