Research article

A novel nonmonotone trust region method based on the Metropolis criterion for solving unconstrained optimization

  • Received: 13 September 2024 Revised: 17 October 2024 Accepted: 31 October 2024 Published: 08 November 2024
  • MSC : 49M37, 65K05, 90C30

  • In this paper, we propose a novel nonmonotone trust region method that incorporates the Metropolis criterion to construct a new function sequence. This sequence is used to update both the trust region ratio and the iteration criterion, increasing the likelihood of accepting the current trial step and introducing randomness into the iteration process. When the current trial step is not accepted, we introduce an improved nonmonotone line search technique to continue the iteration. This approach significantly reduces the number of subproblems that need to be solved, thereby saving computational resources. The stochastic nonmonotone technique helps the algorithm avoid being trapped in the local optima, and a global convergence is guaranteed under certain conditions. Numerical experiments demonstrate that the algorithm can be more effectively applied to a broader range of problems.

    Citation: Yiting Zhang, Chongyang He, Wanting Yuan, Mingyuan Cao. A novel nonmonotone trust region method based on the Metropolis criterion for solving unconstrained optimization[J]. AIMS Mathematics, 2024, 9(11): 31790-31805. doi: 10.3934/math.20241528

    Related Papers:

    [1] Yueting Yang, Hongbo Wang, Huijuan Wei, Ziwen Gao, Mingyuan Cao . An adaptive simple model trust region algorithm based on new weak secant equations. AIMS Mathematics, 2024, 9(4): 8497-8515. doi: 10.3934/math.2024413
    [2] Ke Su, Wei Lu, Shaohua Liu . An area-type nonmonotone filter method for nonlinear constrained optimization. AIMS Mathematics, 2022, 7(12): 20441-20460. doi: 10.3934/math.20221120
    [3] Bothina El-Sobky, Yousria Abo-Elnaga, Gehan Ashry . A nonmonotone trust region technique with active-set and interior-point methods to solve nonlinearly constrained optimization problems. AIMS Mathematics, 2025, 10(2): 2509-2540. doi: 10.3934/math.2025117
    [4] Auwal Bala Abubakar, Poom Kumam, Maulana Malik, Parin Chaipunya, Abdulkarim Hassan Ibrahim . A hybrid FR-DY conjugate gradient algorithm for unconstrained optimization with application in portfolio selection. AIMS Mathematics, 2021, 6(6): 6506-6527. doi: 10.3934/math.2021383
    [5] Xiaoping Xu, Jinxuan Liu, Wenbo Li, Yuhan Xu, Fuxiao Li . Modified nonmonotonic projection Barzilai-Borwein gradient method for nonnegative matrix factorization. AIMS Mathematics, 2024, 9(8): 22067-22090. doi: 10.3934/math.20241073
    [6] Mouhamad Al Sayed Ali, Miloud Sadkane . Acceleration of implicit schemes for large systems of nonlinear differential-algebraic equations. AIMS Mathematics, 2020, 5(1): 603-618. doi: 10.3934/math.2020040
    [7] Dingyu Zhu, Yueting Yang, Mingyuan Cao . An accelerated adaptive two-step Levenberg–Marquardt method with the modified Metropolis criterion. AIMS Mathematics, 2024, 9(9): 24610-24635. doi: 10.3934/math.20241199
    [8] Xuejie Ma, Songhua Wang . A hybrid approach to conjugate gradient algorithms for nonlinear systems of equations with applications in signal restoration. AIMS Mathematics, 2024, 9(12): 36167-36190. doi: 10.3934/math.20241717
    [9] B. El-Sobky, G. Ashry, Y. Abo-Elnaga . An active-set with barrier method and trust-region mechanism to solve a nonlinear Bilevel programming problem. AIMS Mathematics, 2022, 7(9): 16112-16146. doi: 10.3934/math.2022882
    [10] Xiao Guo, Chuanpei Xu, Zhibin Zhu, Benxin Zhang . Nonmonotone variable metric Barzilai-Borwein method for composite minimization problem. AIMS Mathematics, 2024, 9(6): 16335-16353. doi: 10.3934/math.2024791
  • In this paper, we propose a novel nonmonotone trust region method that incorporates the Metropolis criterion to construct a new function sequence. This sequence is used to update both the trust region ratio and the iteration criterion, increasing the likelihood of accepting the current trial step and introducing randomness into the iteration process. When the current trial step is not accepted, we introduce an improved nonmonotone line search technique to continue the iteration. This approach significantly reduces the number of subproblems that need to be solved, thereby saving computational resources. The stochastic nonmonotone technique helps the algorithm avoid being trapped in the local optima, and a global convergence is guaranteed under certain conditions. Numerical experiments demonstrate that the algorithm can be more effectively applied to a broader range of problems.



    Unconstrained optimization is one of the fundamental problems in the optimization theory [4,26]. It has broad applications across various fields such as engineering design, financial investments, signal processing, and so on [6,18,21]. In engineering design, unconstrained optimization is used to determine the most efficient design parameters for systems or products, which may either include optimizing the dimensions of mechanical components or minimizing material usage while maintaining the structural integrity. In financial investments, unconstrained optimization techniques enable investors to identify the optimal portfolio allocation, thereby balancing the risk and the return without being confined by predefined constraints. This helps investors maximize the expected returns while minimizing the potential losses. In signal processing, unconstrained optimization is employed in tasks such as signal denoising, image quality enhancement, and performance optimization of communication systems. By either minimizing error functions or maximizing the signal-to-noise ratio, unconstrained optimization algorithms can extract meaningful information from noisy data. Therefore, the discussion of methods for solving unconstrained optimization problems becomes very important.

    The trust region method is one of the most commonly used iterative algorithms to solve unconstrained optimization problems [14,27]. The fundamental concept behind the trust region method lies in constructing a surrogate model that approximates the objective function within the neighborhood of the current iteration point. This surrogate model captures the local behavior of the original objective function, thus allowing the algorithm to compute the optimal step size or the trial step within the predefined trust region.

    In most practical implementations, a quadratic model is typically chosen as the approximation function due to its mathematical convenience and the availability of efficient solvers. However, the effectiveness of the quadratic model heavily depends on the proper tuning of its parameters. Researchers have proposed various techniques to solve trust region subproblems, including exact solutions, dog-leg methods, and conjugate gradient methods [1,5,33].

    However, trust region subproblems are often difficult to solve, particularly for complex objective functions and high-dimensional problems, thus leading to increased computational costs. The challenges typically arise from the non-convexity and multimodal nature of the surrogate function within the trust region, requiring frequent subproblem solutions during the iterative process. To address this, researchers have explored methods to improve the efficiency of solving these subproblems by developing more efficient solvers, employing advanced optimization techniques, and exploring alternative approximation functions [8,22,23].

    Overall, trust region methods remain powerful tools for unconstrained optimization, and further advancements in subproblem solvers could enhance their performance and applicability to a wider range of problems. To handle situations where the trial step is not accepted, trust region methods combined with line search techniques have been proposed [11,25]. When the trial step is rejected, these algorithms use the trial step as the search direction and apply a line search technique to determine the corresponding step size. However, these methods require the objective function to monotonically decrease at each iteration, which has some impact on the convergence rate of the algorithm. With the introduction of nonmonotone techniques [12], a new class of nonmonotone trust region algorithms was developed [7], which allowed the sequence of the objective function values to be nonmonotone. Numerical results show that these methods outperform traditional trust region methods. Further research has been conducted in this area [10,28,29].

    However, in some cases, the class of aforementioned nonmonotone techniques that used a maximum of recent function values may ignore some better function values. To address this, scholars have proposed improved nonmonotone strategies. Zhang and Hager [34] introduced a new nonmonotone line search algorithm that replaced the maximum function value with the average of recent function values. Gu and Mo [13] developed a simplified nonmonotone strategy that was applied within the trust region framework, which avoided the complex parameterization process of Zhang and Hager [34]. Zhou et al. [35] proposed a nonmonotone trust region method based on a simple model, which performed well on large-scale problems. Improved nonmonotone techniques have also been applied to solve other problems, such as the nonmonotone BFGS (Broyden, Fletcher, Goldfarb and Shanno) algorithm proposed by Wan et al. [31] for smooth nonlinear equations, where the nonmonotone parameters were updated using known information of the nonlinear equation. Some scholars have developed new nonmonotone line search rules that were applied to unconstrained and box-constrained optimization problems [16,19]. In contrast to previous work, we study a stochastic nonmonotone technique that allows the algorithm to explore a larger solution space during the search process.

    To reduce the number of subproblem computations and improve the overall efficiency of the algorithm, we propose a modified trust region method that incorporates recent advances in this field, with the following key innovations and advantages.

    Construction of a New Function Sequence: To prevent convergence to the local optima and to improve algorithmic performance, we introduce the idea of simulated annealing to construct a new function sequence. Based on this sequence, a new nonmonotone trust region ratio is defined, which leads to corresponding improvements in the iteration criterion of the trust region framework, thus providing a greater flexibility during the iterative process.

    Improved Nonmonotone Strategy: When the current trial step is not accepted, we utilize an improved nonmonotone strategy that leverages the new function sequence to determine the step size. This approach effectively reduces the required number of subproblem solutions, thus enhancing the algorithm's overall efficiency.

    Global Convergence and Efficiency: Our algorithm guarantees a global convergence under certain assumptions. Furthermore, numerical experiments demonstrate that the algorithm significantly reduces the number of iterations, improves the efficiency, and delivers a strong performance.

    The structure of this paper is as follows: in Section 2, we describe the new nonmonotone trust region algorithm; in Section 3, the global convergence of our algorithm is proved under certain assumptions; in Section 4, the related numerical experimental results are given; and finally, the conclusion of this paper is given in Section 5.

    Consider the following unconstrained optimization problem:

    minxRnf(x), (2.1)

    where f: RnR is continuously differentiable. The difficulty in solving the trust region problem often stems from the need to balance multiple objectives: minimizing the approximate objective function, satisfying the trust region constraint, and potentially incorporating additional requirements such as sparsity or regularization. The trust region method computes the trial step dk by solving the following subproblem:

    mindRnmk(d)=gTkd+12dTBkd,s.t.dΔk, (2.2)

    where gk=f(xk), Bk is the approximation of the Hessian matrix at xk, Δk is the trust region radius, and denotes the Euclidean norm.

    As we all know, in trust region method, the trust region ratio is the key to determining whether the algorithm iterates using the current trial step. The following original trust region ratio is monotone:

    rk=AredPred=fkf(xk+dk)mk(0)mk(dk), (2.3)

    where fk=f(xk), Ared is the actual reduction, and Pred is the predicted reduction. The monotone technique may slow down the rate of convergence, particularly in a narrow curved valley. In order to overcome this disadvantage, the nonmonotone trust region ratio [30] has been proposed:

    rk=flkf(xk+dk)mk(0)mk(dk),

    where flk=max0jq(k)f(xkj), N is a fixed positive constant, and

    q(k)={k,ifkN,min{q(k1)+1,N},ifk>N.

    However, this method may result in missing some well-performing function values, so some scholars have proposed improved nonmonotone algorithms. Gu and Mo [13] constructed the following trust region ratio in 2008:

    rk=Dkf(xk+dk)mk(0)mk(dk),

    where

    Dk={fk,ifk=0,ηDk1+(1η)fk,ifk1,

    and 0<η<1 or a variable η.

    Based on the inspiration from the above ideas, we propose a new nonmonotone trust region algorithm, which allows the algorithm to use the existing favorable information, and also avoids the algorithm from falling into local optimal solutions. To achieve these, we consider introducing the idea of simulated annealing [15,17,32] into the trust region method. A new trust region framework is constructed, which includes the improvement of the trust region iteration criterion and the trust region ratio, etc.

    First, we construct a new trust region ratio:

    rk=Rkf(xk+dk)mk(0)mk(dk), (2.4)

    where

    Rk={fk,ifk=0,(1pk1)Rk1+pk1fk,ifk1, (2.5)

    and

    pk={1,ifrkτ,exp{τrkTk},otherwise. (2.6)

    Here, we introduce the above parameter pk [3,20] into the Rk, where Tk is the temperature during annealing, τ is a given constant, and rk is the trust region ratio. It is crucial to control the rate of the annealing cooling process [2], which is usually updated according to the following:

    Tk+1=βTk,where0<β<1.

    From the definition of pk, we know that it is dynamically updated by the value of rk. From this, it can be observed that our trust region ratio rk can be adaptively adjusted according to the existing information.

    Furthermore, considering the relation between rk and τ in the parameter pk, we use the parameter pk to determine whether the current trial step dk is accepted. For a random number lk<1, if pklk is satisfied, then the current trial step is accepted. When rkτ, pk=1, the current trial step is naturally accepted. Otherwise, the current trial step will be accepted with a certain probability, and the larger pk is more probable when the current trial step is accepted. If pk<lk, then the current trial step is rejected. Using our sequence Rk, we propose a new nonmonotone line search:

    f(xk+λidk)Rk+δλigTkdk,

    where 0<δ<1, 0<λ<1. According to the line search, the corresponding step size is obtained for the iteration. The parameter im is computed by this method, which is iterated through xk+1=xk+λimdk. The nonmonotone line search doesn't require the objective function to be monotonically decreasing, which makes the overall algorithm more flexible and faster.

    Combined with the above improvements, we propose the nonmonotone trust region algorithm based on the Metropolis criterion (Algorithm 2.1).

    Algorithm 2.1.

    Step 0. Given an initial point x0Rn, R0=f(x0), Δ0>0, T0>0, ε>0, v>1,

    0<λ<1, 0<δ0.5, 0<η1<η2<1, 0<γ1<1<γ2, 0<β<1,

    0<τη1, B0=ξI,ξ>0, let k:=0.

    Step 1. Compute gk. If gkε, stop.

    Step 2. Solve the subproblem (2.2) to obtain dk.

    Step 3. Compute rk and pk by (2.4) and (2.6), respectively. Additionally, compute the following:

    lk=ev+(e1/vev)×rand(1).

    Step 4. If pklk, then xk+1=xk+dk. Otherwise, compute im, which is the minimum nonnegative integer i which satisfies the following:

    f(xk+λidk)Rk+δλigTkdk;

    then, xk+1=xk+λimdk.

    Step 5. Update the trust region radius:

    Δk+1={γ1Δk,ifrkη1,γ2Δk,ifrkη2,Δk,   otherwise.

    Step 6. Compute fk+1 and Rk+1=(1pk)Rk+pkfk+1, Tk+1=βTk, and compute Bk+1 by a quasi-Newton update. Set k:=k+1, then go to Step 1.

    Remark 2.1. In Step 3, if rkτ, then the function approximation of the previous iteration is better; then, pk=1 and Rk+1=fk+1, which is reduced to the case of the traditional trust region ratio. If rk<τ, then 0<pk<1, and Rk+1 is the convex combination of Rk and fk+1.

    Nonmonotone trust region methods have good global convergence as described in [13,24]. In this section, we will analyze the convergence of our algorithm, which also inherits the properties of nonmonotone trust region methods. First, the following assumptions are made.

    Assumption 3.1. Let L={xRn|f(x)f(x0)}Ω be the level set, where ΩRn is a bounded closed set.

    Assumption 3.2. There exists a positive constant m such that dTBkdmd2, dRn.

    Remark 3.1. Combining with Assumptions 3.1 and 3.2, f(x) is a quadratic continuous differentiable function. It follows that there exists a positive constant M>m, such that BkM.

    Lemma 3.1. Let the sequence {xn} be generated by the Algorithm 2.1. There are

    mk(0)mk(dk)12gkmin{Δk,gkBk}, (3.1)
    gTkdk12gkmin{Δk,gkBk}0. (3.2)

    Lemma 3.2. Let the sequence {xn} be generated by the Algorithm 2.1. Then,

    Rk+1fk+1

    holds for all k.

    Proof. (1) From the definition of Rk+1 and pk, if rkτ, then we have the following:

    Rk+1fk+1=(1pk)(Rkfk+1)=0.

    (2) From Tk+1=βTk, if rk<τ and pk=exp{τrkTk}lk, then we can have limkTk=0. It is known that lk[ev,e1/v], and we can obtain vlnlk1v. Therefore, there exists a positive integer N such that 0<Tklnlk<ε1 holds for any k>N, ε1>0. From the definition of pk, we have the following:

    rkτ+Tklnlkτε1.

    From the definition of rk and (3.1), we have the following:

    Rkfk+1(τε1)[mk(0)mk(dk)]12(τε1)gkmin{Δk,gkBk}. (3.3)

    Let ε1=τ2; then

    Rk+1fk+1=(1pk)(Rkfk+1)14(1pk)τgkmin{Δk,gkBk}0.

    (3) If rk<τ and pklk, then f(xk+λimdk)Rk+δλimgTkdk holds. Combining with (3.2), we obtain the following:

    Rkfk+1δλimgTkdk0.

    Therefore, Rk+1fk+1 holds. The proof is completed.

    Lemma 3.3. Let the sequence {xn} be generated by the Algorithm 2.1. The sequence {Rk} is monotonically decreasing.

    Proof. (1) If rkτ, namely,

    Rkfk+1τ[mk(0)mk(dk)]12τgkmin{Δk,gkBk},

    then

    fk+1Rk12τgkmin{Δk,gkBk}.

    According to the definition of Rk+1, we have the following:

    Rk+1Rk12τpkgkmin{Δk,gkBk}<Rk. (3.4)

    (2) By (3.3) and the definition of Rk+1, if rk<τ and pk>lk, then we obtain the following:

    RkRk+1=pk(Rkfk+1)>14τpkgkmin{Δk,gkBk}. (3.5)

    (3) If rk<τ and pklk, then f(xk+λimdk)Rk+δλimgTkdk holds. Since gTkdk0, then fk+1Rk.

    By the definition of Rk+1, we obtain the following:

    Rk+1=(1pk)Rk+pkfk+1(1pk)Rk+pkRk=Rk.

    Therefore, Rk+1Rk holds. The proof is completed.

    Lemma 3.4. Let the sequence {xn} be generated by the Algorithm 2.1. Then, the step length αk satisfies αk>(1δ)λmM.

    Proof. The proof can be proved similarly to the proof of Lemma 3.3 in [13] and is omitted here.

    To facilitate the subsequent discussion, the definition of the sequence Mk is introduced as Mk=1+ max0ik Bi.

    Lemma 3.5. Assume that the sequences {xk} and {Rk} are generated by the Algorithm 2.1. If there exists a positive constant ε, such that gkε holds, then for any k, there is the following:

    Rk+1Rk12εpkψmin{Δk,εMk},

    where ψ=min{12τ,δ(1δ)λmM}.

    Proof. (1) By (3.4) and (3.5), if pk>lk, then we know that

    Rk+1Rk14τpkgkmin{Δk,gkBk}.

    (2) By combining (3.2), Lemma 3.4, and fk+1Rk+δαkgTkdk, if pklk, then the following holds:

    fk+1Rk12(1δ)λmMδgkmin{Δk,gkBk}.

    According to the definition of Rk+1, we obtain the following:

    Rk+1Rk=pk(fk+1Rk)12(1δ)λmMδpkgkmin{Δk,gkBk}.

    Additionally, if ψ=min{12τ,δ(1δ)λmM} and gkε, then

    Rk+1Rk12εpkψmin{Δk,εMk}.

    The proof is completed.

    Lemma 3.6. Assume that Assumption 3.1 holds. If the sequence {xk} generated by the Algorithm 2.1 does not converge (i.e., there exists a positive constant ε, such that gkε), then

    limkmin{Δk,εMk}=0.

    Proof. Algorithm 2.1 shows that R0=f0. From Lemmas 3.2 and 3.3, we know that fk+1Rk+1Rkf0. Assumption 3.1 shows that if {f(xk)} is bounded, then {Rk} is bounded. According to Lemma 3.5, we can know that limkmin{Δk,εMk}=0.

    Lemma 3.7. If the sequence {xk} generated by the Algorithm 2.1 does not converge (i.e., there exists a positive constant ε such that gkε), then there exists a set J={k|pk<lk} such that the following inequality holds for a sufficiently large kJ:

    xk+1xk(1τ)εmin{Δk,εMk}Mτm,αk=1, (3.6)
    xk+1xk>(1δ)ελMmin{1,εΔkMk},αk<1, (3.7)
    Δk>εMk,kJ.

    Proof. Based on α, we consider the following two cases.

    Case 1. If im=0 (i.e., αk=1), then

    xk+1xk=dkΔk,kJ.

    At this point, there is rk<τ. According to fk+1Rk+1 and the definition of rk, we have the following:

    f(xk)f(xk+dk)Rkf(xk+dk)τ(gTkdk12dTkBkdk). (3.8)

    According to the Taylor's expansion and Remark 3.1, we obtain the following:

    f(xk+dk)f(xk)gTkdk+12Mdk2. (3.9)

    Combining (3.8) and (3.9), we have the following:

    gTkdk12Mdk2τ(gTkdk12dTkBkdk). (3.10)

    Combining (3.10) and Assumption 3.2, we obtain the following:

    (1τ)gTkdk12(Mτm)dk2. (3.11)

    Combining (3.2) and (3.11), we have the following:

    (1τ)gkmin{Δk,gkBk}(Mτm)dk2. (3.12)

    From the definition of Mk and gkε, and combined with xk+1xk=dk and (3.12), we obtain the following:

    (1τ)εmin{Δk,εMk}(Mτm)xk+1xk2.

    Therefore, it is proved that xk+1xk(1τ)εmin{Δk,εMk}Mτm,αk=1.

    Assuming ΔkεMk, there is (3.6) to obtain Δk(1τ)εΔkMτm (i.e., Δk(1τ)εMτm). Combined with ΔkεMk, there is εMkΔk(1τ)εMτm. This contradicts Lemma 3.6, so we have Δk>εMk.

    Case 2. If im>0 (i.e., αk<1), then by combining (3.2), Remark 3.1, Lemma 3.2, and the Step 4 of Algortihm 2.1, according to the Taylor's expansion yield,

    0>Rkf(xk+λ1αkdk)+δλ1αkgTkdkfkf(xk+λ1αkdk)+δλ1αkgTkdk(1δ)λ1αkgTkdk12Mλ2αk2dk212λ1αk[(1δ)εmin{Δk,εMk}Mλ1dkxk+1xk]12λ1αk[(1δ)εmin{Δk,εMk}Mλ1Δkxk+1xk]=12λ1αkΔk[(1δ)εmin{1,εΔkMk}Mλ1xk+1xk].

    Therefore, it is proved that xk+1xk>(1δ)ελMmin{1,εΔkMk},αk<1.

    There is

    xk+1xk=αkdkΔk. (3.13)

    Suppose there exists ΔkεMk. Combining (3.7), we have the following:

    xk+1xk>(1δ)ελM. (3.14)

    Combining (3.13), (3.14), and Assumption 3.1, there are εMkΔkxk+1xk>(1δ)ελM. This contradicts limkmin{Δk,εMk}=0. Therefore, Δk>εMk holds.

    In summary, the proof is completed.

    Lemma 3.8. If the sequence {xk} generated by the Algorithm 2.1 does not converge (i.e., there exists a positive constant ε such that gkε), then Δk>εMk holds for all sufficiently large k.

    Proof. (1) If there are only finitely many k such that Δk+1Δk (i.e., J is a finite set), then there exists a positive constant Δ such that Δk>Δ. According to Lemma 3.6, we know that limk εMk=0.

    (2) We assume that J is an infinite set. It follows from Lemma 3.7 that there exists ˉkJ, such that Δk>εMk holds when kJ,kˉk. When kJ, kˉk, noted as ˜k=max{i|iJ,ik}. We have Δ˜k>εM˜k, and ˜k+sJ, s=1,2,,k˜k. We can obtain Δ˜kΔ˜k+1Δ˜k+2Δk, so Δk>εM˜k holds. According to the definition of Mk, {Mk} is a monotonically increasing sequence.

    The proof is completed.

    Theorem 1. If Assumptions 3.1 and 3.2 hold, and {Bk} satisfies k=01Mk=, then the sequence {xk} generated by the Algorithm 2.1 satisfies the following:

    lim infkgk=0.

    Proof. Suppose there exists a positive constant ε such that gkε. Lemma 3.8 holds, that is Δk>εMk holds for sufficiently large k. By Lemma 3.5, we know that RkRk+112εpkψmin{Δk,εMk}. We have k=0(RkRk+1)k=012ε2pkψ1Mk, then k=01Mk< contradicts k=01Mk=. The theorem is proved.

    In the following, we will show the effectiveness of our Algorithm 2.1 through some numerical experimental results. The numerical experiments are all implemented by using MATLAB (R2017a) on a PC with a CPU of 2.30 GHz and 8.00 GB RAM. The relevant parameters in Algorithm 2.1 and other comparison algorithms are selected as T0=200, v=100, ε=105, Δ0=g0, λ=0.5, δ=0.5, η1=0.25, η2=0.75, γ1=0.5, γ2=2, β=0.9, and τ=0.25. The matrix Bk is updated by the BFGS formula, i.e.,

    Bk+1=Bk+ykyTksTkykBksksTkBksTkBksk.

    In order to explore the influence of the simulated annealing idea, nonmonotone technologies, and different trust region ratios on the algorithm, we perform the following three sets of comparison experiments:

    (a) Algorithm 2.1 is compared with the existing trust region algorithm. The NTRM (Nonmonotone Trust Region Method With Line Search) algorithm [13] is a classical nonmonotone trust region algorithm. The SATRBB (Simulated Annealing-based Trust Region Bazilai-Borwein) algorithm [20] is a trust region algorithm combined with simulated annealing.

    (b) Algorithm 2.1 is compared with the algorithms that combine different line search techniques (i.e., Algorithm 2.1 without a line search (SATR) and with a monotonic line search (ATR)).

    (c) Algorithm 2.1 is compared with algorithms that combine different ratios (i.e., Algorithm 2.1 by using the original trust region ratio (SATR-OR), correcting the numerator portion of the trust region ratio to the form in [13] (SATR-U), and correcting the numerator and denominator portions of the trust region ratio according to Ck in [13] (SATR-UD)).

    Remark 4.1. The SATR-UD algorithm is inspired by Remark in [29], and the form of trust region ratio is as follows:

    rk=Ckf(xk+dk)Ckfkmk(dk).

    The test functions used and their associated dimensionality are shown in Table 1. We require the maximum number of iterations of the algorithm to be 5000. It is worth noting that the algorithm which uses the idea of simulated annealing is random in practice; therefore the relevant data are the average of the results of 10 tests.

    Table 1.  List of test functions.
    Problem Dimension Problem Dimension
    Helical valley 3 Chebyquad 7
    Biggs EXP6 6 Freudenstein and Roth 2
    Gaussian 3 Generalized Rosebrock 50/100/500
    Powell badly scaled 2 Boundary value 100/500
    Box 3 Broyden tridiagonal 1000/2000
    Variable dimension 1000/3000/5000 Separable cubic 100/500/1000
    Watson 1000/2000 Nearly separable 1000/2000/3000/5000
    Penalty1 500/1000/1500 Yang tridiagonal 50/100/500
    Brown and Dennis 4 Allgower 100/300/500
    Wood 4 Schittkowski 100/300
    Extended Rosenbrock 500/1000 Beale 2
    Extended Powell singular 100/500/1000

     | Show Table
    DownLoad: CSV

    To further investigate the efficiency of the algorithm, we use the performance profiles [9] to evaluate and compare the performance of the solvers. Assume that k denotes the number of iterations required, and t denotes the time required for the algorithm. Moreover, there exist ns solvers on the test set S and np problems on test set P. If the computation running time is used as the performance metric, then we define tp,s as computing time required to solve problem p by the solver s. We define the performance ratio as follows:

    rp,s0=tp,s0min{tp,s:sS}.

    There exists a rM such that there is rMrs for any p and s. For the overall evaluation of the solver, we define the following:

    ρs(τ)=1npsize{pP:rp,sτ},

    where ρ is the cumulative distribution function of the performance ratio, which is the probability that the performance ratio rp,s of the solver s is within the best possible ratio factor τ.

    The above three sets of experiments were conducted on the efficiency of solving some problems using the number of iterations and the time required for the algorithm as the performance metrics. The results are shown in Figures 13. It is not difficult to see from Figures 1 and 2 that Algorithm 2.1 has good properties compared with existing algorithms, and the nonmonotone line search technology proposed in this paper greatly improves the performance of the algorithm. Figure 3 shows that although the SATR-UD algorithm has a slight advantage for some problems. However, compared with Algorithm 2.1, there are still some problems that the SATR-UD algorithm cannot solve. All in all, the algorithm proposed in this paper is the preferred solver for the above problem with a high probability. Algorithm 2.1 solves the problems with a higher accuracy and requires fewer iterations.

    Figure 1.  Performance profiles of (a) for the number of iterations k and the CPU time t.
    Figure 2.  Performance profiles of (b) for the number of iterations k and the CPU time t.
    Figure 3.  Performance profiles of (c) for the number of iterations k and the CPU time t.

    In this paper, we proposed a novel stochastic nonmonotone trust region algorithm, which incorporated the simulated annealing for an enhanced performance. This integration of ideas aimed to address some of the challenges faced by traditional optimization methods, especially in terms of avoiding the local minima and achieving a global convergence. At the heart of our algorithm lies the construction of a new sequence, which was governed by the Metropolis criterion. The trust region ratio was modified based on this sequence, and the iterative criterion was adjusted according to the improved ratio. When the current trial step was not accepted, the iteration step size was obtained based on the modified nonmonotone line search. By doing so, we not only reduced the number of solutions required for the trust region subproblem, but also enhanced the algorithm's ability to escape the local minimum and converge to a global optimum. A theoretical analysis showed that our proposed algorithm achieved a global convergence under certain conditions, making it avoid local optimal solutions as much as possible. Furthermore, numerical experiments were conducted to demonstrate the effectiveness of our algorithm in practical applications. These experiments revealed that our algorithm outperformed traditional methods.

    Yiting Zhang: Conceptualization, Methodology, Investigation, Software, Validation, Writing-original draft; Chongyang He: Funding acquisition, Investigation, Software; Wanting Yuan: Validation, Writing-review & editing; Mingyuan Cao: Funding acquisition, Methodology, Software, Writing-review & editing. All authors have read and approved the final version of the manuscript for publication.

    The key project of the natural science foundation joint fund of Jilin province (YDZJ202201ZYTS303, YDZJ202101ZYTS167, 20230508184RC, JJKH20200037KJ); The graduate innovation project of Beihua University (2023003); The science and technology development plan project of Jilin province (20200403193SF).

    All authors declare no conflicts of interest in this paper.



    [1] N. M. Alexandrov, J. E. Dennis, R. M. Lewis, V. Torczon, A trust region framework for managing the use of approximation models in optimization, Struct. Optim., 15 (1998), 16–23. https://doi.org/10.1007/BF01197433 doi: 10.1007/BF01197433
    [2] S. Babaie-Kafaki, R. Ghanbari, N. Mahdavi-Amiri, An efficient and practically robust hybrid metaheuristic algorithm for solving fuzzy bus terminal location problems, Asia-Pac. J. Oper. Res., 29 (2012), 1250009. https://doi.org/10.1142/S0217595912500091 doi: 10.1142/S0217595912500091
    [3] S. Babaie-Kafaki, S. Rezaee, A randomized nonmonotone adaptive trust region method based on the simulated annealing strategy for unconstrained optimization, Int. J. Intell. Comput., 12 (2019), 389–399. https://doi.org/10.1108/IJICC-12-2018-0178 doi: 10.1108/IJICC-12-2018-0178
    [4] J. Bai, L. Jia, Z. Peng, A new insight on augmented Lagrangian method with applications in machine learning, J. Sci. Comput., 99 (2024), 53. https://doi.org/10.1007/s10915-024-02518-0 doi: 10.1007/s10915-024-02518-0
    [5] J. Chen, W. Y. Sun, Nonmonotone adaptive trust-region algorithms with indefinite dogleg path for unconstrained minimization, Northeast. Math. J., 24 (2008), 19–30.
    [6] J. Deepho, A. B. Abubakar, M. Malik, I. K. Argyros, Solving unconstrained optimization problems via hybrid CD-DY conjugate gradient methods with applications, J. Comput. Appl. Math., 405 (2022), 113823. https://doi.org/10.1016/j.cam.2021.113823 doi: 10.1016/j.cam.2021.113823
    [7] N. Y. Deng, Y. Xiao, F. J. Zhou, Nonmonotonic trust region algorithm, J. Optim. Theory Appl., 76 (1993), 259–285. https://doi.org/10.1007/BF00939608 doi: 10.1007/BF00939608
    [8] S. Di, W. Sun, A trust region method for conic model to solve unconstraind optimizaions, Optim. Method. Softw., 6 (1996), 237–263. https://doi.org/10.1080/10556789608805637 doi: 10.1080/10556789608805637
    [9] E. D. Dolan, J. J. More, Benchmarking optimization software with performance profiles, Math. Program., 91 (2002), 201–213. https://doi.org/10.1007/s101070100263 doi: 10.1007/s101070100263
    [10] J. Fu, W. Sun, Nonmonotone adaptive trust-region method for unconstrained optimization problems, Appl. Math. Comput., 163 (2005), 489–504. https://doi.org/10.1016/j.amc.2004.02.011 doi: 10.1016/j.amc.2004.02.011
    [11] E. M. Gertz, Combination trust-region line search methods for unconstrained optimization, San Diego: University of California, 1999.
    [12] L. Grippo, F. Lampariello, S. Lucidi, A nonmonotone line search technique for Newton’s method, SIAM J. Numer. Anal., 23 (1986), 707–716. https://doi.org/10.1137/0723046 doi: 10.1137/0723046
    [13] N. Gu, J. Mo, Incorporating nonmonotone strategies into the trust region method for unconstrained optimization, Comput. Math. Appl., 55 (2008), 2158–2172. https://doi.org/10.1016/j.camwa.2007.08.038 doi: 10.1016/j.camwa.2007.08.038
    [14] M. Hatamian, M. Paripour, F. M. Yaghoobi, N. Karamikabir. An adaptive nonmonotone line search technique for solving systems of nonlinear equations, J. Math., 2021 (2021), 7134561. https://doi.org/10.1155/2021/7134561 doi: 10.1155/2021/7134561
    [15] D. Henderson, S. H. Jacobson, A. W. Johnson, The theory and practice of simulated annealing, In: Handbook of Metaheuristics, Boston: Springer, 2003, 287–319. https://doi.org/10.1007/0-306-48056-5_10
    [16] S. Huang, Z. Wan, J. Zhang, An extended nonmonotone line search technique for large-scale unconstrained optimization, J. Comput. Appl. Math., 330 (2018), 586–604. https://doi.org/10.1016/j.cam.2017.09.026 doi: 10.1016/j.cam.2017.09.026
    [17] S. Kirkpatrick, C. D. Gellat, M. P. Vecchi, Optimization by simulated annealing, Science, 220 (1983), 671–680. https://doi.org/10.1126/science.220.4598.671 doi: 10.1126/science.220.4598.671
    [18] J. Lee, K. Jung, J. H. Kwon, The aerodynamic shape optimization of airfoils using unconstrained trust region methods, Eng. Optimiz., 41 (2009), 459–471. https://doi.org/10.1080/03052150802596068 doi: 10.1080/03052150802596068
    [19] T. Li, Z. Wan, J. Guo, A new nonmonotone spectral projected gradient algorithm for box-constrained optimization problems in m×n real matrix space with application in image clustering, J. Comput. Appl. Math., 438 (2024), 115563. https://doi.org/10.1016/j.cam.2023.115563 doi: 10.1016/j.cam.2023.115563
    [20] X. Li, W. Dong, Z. Peng, A new nonmonotone trust region Barzilai-Borwein method for unconstrained optimization problems, Acta Math. Appl. Sin. Engl. Ser., 37 (2021), 166–175. https://doi.org/10.1007/s10255-021-0997-9 doi: 10.1007/s10255-021-0997-9
    [21] Y. Liu, X. Liu, Application and performances of unconstrained optimization methods in seafloor AVO inversion, Arab. J. Geosci., 9 (2016), 652. https://doi.org/10.1007/s12517-016-2692-3 doi: 10.1007/s12517-016-2692-3
    [22] S. Lu, Z. Wei, L. Li, A trust region algorithm with adaptive cubic regularization methods for nonsmooth convex minimization, Comput. Optim. Appl., 51 (2012), 551–573. https://doi.org/10.1007/s10589-010-9363-1 doi: 10.1007/s10589-010-9363-1
    [23] Q. Ni, Optimality conditions for trust-region subproblems involving a conic model, SIAM J. Optim., 15 (2005), 826–837. https://doi.org/10.1137/S1052623402418991 doi: 10.1137/S1052623402418991
    [24] T. D. Niri, M. Heydari, M. M. Hosseini, Two nonmonotone trust region algorithms based on an improved Newton method, J. Appl. Math. Comput., 64 (2020), 179–194. https://doi.org/10.1007/s12190-020-01350-7 doi: 10.1007/s12190-020-01350-7
    [25] J. Nocedal, Y. Yuan, Combining trust region and line search techniques, Advances in nonlinear programming, Boston, MA: Springer, 1998, 153–175. https://doi.org/10.1007/978-1-4613-3335-7_7
    [26] S. Rezaee, S. Babaie-Kafaki, An adaptive nonmonotone trust region algorithm, Optim. method. softw., 34 (2019), 264–277. https://doi.org/10.1080/10556788.2017.1364738 doi: 10.1080/10556788.2017.1364738
    [27] S. Sun, J. Nocedal, A trust region method for noisy unconstrained optimization, Math. Program., 202 (2023), 445–472. https://doi.org/10.1007/s10107-023-01941-9 doi: 10.1007/s10107-023-01941-9
    [28] W. Sun, J. Han, J. Sun, Global convergence of nonmonotone descent methods for unconstrained optimization problems, J. Comput. Appl. Math., 146 (2002), 89–98. https://doi.org/10.1016/S0377-0427(02)00420-X doi: 10.1016/S0377-0427(02)00420-X
    [29] W. Sun, Nonmonotone trust region method for solving optimization problems, Appl. Math. Comput., 156 (2004), 159–174. https://doi.org/10.1016/j.amc.2003.07.008 doi: 10.1016/j.amc.2003.07.008
    [30] P. L. Toint, Non-monotone trust-region algorithms for nonlinear optimization subject to convex constraints, Math. Program., 77 (1997), 69–94. https://doi.org/10.1007/BF02614518 doi: 10.1007/BF02614518
    [31] Z. Wan, Y. Chen, S. Huang, D. Feng, A modified nonmonotone BFGS algorithm for solving smooth nonlinear equations, Optim. Lett., 8 (2014), 1845–1860. https://doi.org/10.1007/s11590-013-0678-6 doi: 10.1007/s11590-013-0678-6
    [32] X. S. Yang, Nature-inspired optimization algorithms, Academic Press, 2020. https://doi.org/10.1016/C2013-0-01368-0
    [33] G. L. Yuan, Z. X. Wei, A trust region algorithm with conjugate gradient technique for optimization problems, Numer. Func. Anal. Opt., 32 (2011), 212–232. https://doi.org/10.1080/01630563.2010.532273 doi: 10.1080/01630563.2010.532273
    [34] H. Zhang, W. W. Hager, A nonmonotone line search technique and its application to unconstrained optimization, SIAM J. Optim., 14 (2004), 1043–1056. https://doi.org/10.1137/S1052623403428208 doi: 10.1137/S1052623403428208
    [35] Q. Y. Zhou, J. Chen, Z. W. Xie, A nonmonotone trust region method based on simple quadratic models, J. Comput. Appl. Math., 272 (2014), 107–115. https://doi.org/10.1016/j.cam.2014.04.026 doi: 10.1016/j.cam.2014.04.026
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(584) PDF downloads(48) Cited by(0)

Figures and Tables

Figures(3)  /  Tables(1)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog