Loading [MathJax]/jax/output/SVG/jax.js
Research article

An accelerated forward-backward algorithm with a new linesearch for convex minimization problems and its applications

  • Received: 21 January 2021 Accepted: 01 April 2021 Published: 07 April 2021
  • MSC : 65K05, 90C25, 90C30

  • We study and investigate a convex minimization problem of the sum of two convex functions in the setting of a Hilbert space. By assuming the Lipschitz continuity of the gradient of the function, many optimization methods have been invented, where the stepsizes of those algorithms depend on the Lipschitz constant. However, finding such a Lipschitz constant is not an easy task in general practice. In this work, by using a new modification of the linesearches of Cruz and Nghia [7] and Kankam et al. [14] and an inertial technique, we introduce an accelerated algorithm without any Lipschitz continuity assumption on the gradient. Subsequently, a weak convergence result of the proposed method is established. As applications, we apply and analyze our method for solving an image restoration problem and a regression problem. Numerical experiments show that our method has a higher efficiency than the well-known methods in the literature.

    Citation: Adisak Hanjing, Pachara Jailoka, Suthep Suantai. An accelerated forward-backward algorithm with a new linesearch for convex minimization problems and its applications[J]. AIMS Mathematics, 2021, 6(6): 6180-6200. doi: 10.3934/math.2021363

    Related Papers:

    [1] Suparat Kesornprom, Papatsara Inkrong, Uamporn Witthayarat, Prasit Cholamjiak . A recent proximal gradient algorithm for convex minimization problem using double inertial extrapolations. AIMS Mathematics, 2024, 9(7): 18841-18859. doi: 10.3934/math.2024917
    [2] Adisak Hanjing, Panadda Thongpaen, Suthep Suantai . A new accelerated algorithm with a linesearch technique for convex bilevel optimization problems with applications. AIMS Mathematics, 2024, 9(8): 22366-22392. doi: 10.3934/math.20241088
    [3] Kasamsuk Ungchittrakool, Natthaphon Artsawang . A generalized viscosity forward-backward splitting scheme with inertial terms for solving monotone inclusion problems and its applications. AIMS Mathematics, 2024, 9(9): 23632-23650. doi: 10.3934/math.20241149
    [4] Weibo Guan, Wen Song . Convergence rates of the modified forward reflected backward splitting algorithm in Banach spaces. AIMS Mathematics, 2023, 8(5): 12195-12216. doi: 10.3934/math.2023615
    [5] Premyuda Dechboon, Abubakar Adamu, Poom Kumam . A generalized Halpern-type forward-backward splitting algorithm for solving variational inclusion problems. AIMS Mathematics, 2023, 8(5): 11037-11056. doi: 10.3934/math.2023559
    [6] Suparat Kesornprom, Prasit Cholamjiak . A modified inertial proximal gradient method for minimization problems and applications. AIMS Mathematics, 2022, 7(5): 8147-8161. doi: 10.3934/math.2022453
    [7] Hasanen A. Hammad, Habib ur Rehman, Manuel De la Sen . Accelerated modified inertial Mann and viscosity algorithms to find a fixed point of $ \alpha - $inverse strongly monotone operators. AIMS Mathematics, 2021, 6(8): 9000-9019. doi: 10.3934/math.2021522
    [8] Damrongsak Yambangwai, Chonjaroen Chairatsiripong, Tanakit Thianwan . Iterative manner involving sunny nonexpansive retractions for nonlinear operators from the perspective of convex programming as applicable to differential problems, image restoration and signal recovery. AIMS Mathematics, 2023, 8(3): 7163-7195. doi: 10.3934/math.2023361
    [9] Puntita Sae-jia, Suthep Suantai . A new two-step inertial algorithm for solving convex bilevel optimization problems with application in data classification problems. AIMS Mathematics, 2024, 9(4): 8476-8496. doi: 10.3934/math.2024412
    [10] Anjali, Seema Mehra, Renu Chugh, Salma Haque, Nabil Mlaiki . Iterative algorithm for solving monotone inclusion and fixed point problem of a finite family of demimetric mappings. AIMS Mathematics, 2023, 8(8): 19334-19352. doi: 10.3934/math.2023986
  • We study and investigate a convex minimization problem of the sum of two convex functions in the setting of a Hilbert space. By assuming the Lipschitz continuity of the gradient of the function, many optimization methods have been invented, where the stepsizes of those algorithms depend on the Lipschitz constant. However, finding such a Lipschitz constant is not an easy task in general practice. In this work, by using a new modification of the linesearches of Cruz and Nghia [7] and Kankam et al. [14] and an inertial technique, we introduce an accelerated algorithm without any Lipschitz continuity assumption on the gradient. Subsequently, a weak convergence result of the proposed method is established. As applications, we apply and analyze our method for solving an image restoration problem and a regression problem. Numerical experiments show that our method has a higher efficiency than the well-known methods in the literature.



    Throughout this article, we suppose that H is a real Hilbert space with an inner product , and the induced norm . We are interested in studying the following unconstrained convex minimization problem:

    minimizef1(x)+f2(x),subject toxH (1.1)

    where f1,f2:HR{} are two proper, lower semi-continuous and convex functions such that f1 is differentiable on an open set containing the domain of f2. Problem (1.1) has been widely studied due to its applications which can be used in various real-world applications such as in signal and image processing, in regression problems, and in classification problems, etc., see [3,5,8,10,11,13] and the references therein. One of the important topics of studying Problem (1.1) is to invent some efficient procedures for approximating minimizers of f1+f2. Various optimization methods were introduced and developed by many researchers, see [3,5,6,7,8,12,14,15,16,25,26,27,32], for instance.

    If a minimizer x of f1+f2 exists, it is known that x is characterized by the fixed point equation of the forward-backward operator

    x=FBα(x):=proxαf2backward step(xαf1(x))forward step, (1.2)

    where α>0, proxf2 is the proximity operator of f2 and f1 stands for the gradient of f1. The above equation leads to the following iterative method:

    Method 1. Let x1domf2. For k1, let

    xk+1=proxαkf2(xkαkf1(xk)),

    where 0<αk<2L and L is a Lipschitz constant of f1.

    This method is well known as the forward-backward splitting algorithm [8,15], which includes the proximal point algorithm [17,24], the gradient method [4,9] and the CQ algorithm [1] as special cases. It is observed from Method 1 that we need to assume the Lipschitz continuity condition on the gradient of f1 and the stepsize αk depends on the Lipschitz constant L, which is not an easy task to find in general practice (see [3,5,8,12] for other relevant methods).

    In the sequel, we set the standing hypotheses on Problem (1.1) as follows:

    (H1) f1,f2:HR{} are two proper, lower semi-continuous and convex functions with domf2domf1, and domf2 is nonempty, closed and convex;

    (H2) f1 is differentiable on an open set containing domf2. The gradient f1 is uniformly continuous on any bounded subset of domf2 and maps any bounded subset of domf2 to a bounded set in H.

    We note that the second part of (H2) is a weaker assumption than the Lipschitz continuity assumption on f1.

    Cruz and Nghia [7] proposed a technique for selecting the stepsize αk which is independent of the Lipschitz constant L by using the following linesearch process.

    Linesearch 1. Given xdomf2, σ>0, θ(0,1) and δ>0.
    Input α=σ.
      While αf1(FBα(x))f1(x)>δFBα(x)x, do
        α=θα.
      End
    Output α.

     | Show Table
    DownLoad: CSV

    Linesearch 1 is a particular case of the linesearch proposed in [29] for inclusion problems and it was shown that this linesearch is well defined, that is, it stops after finitely many steps, see [7, Lemma 3.1] and [29, Theorem 3.4(a)]. Cruz and Nghia [7] employed the forward-backward iteration where the stepsize αk is generated by Linesearch 1.

    Method 2. Let x1domf2, σ>0, θ(0,1) and δ(0,12). For k1, let

    xk+1=proxαkf2(xkαkf1(xk)),

    where αk:= Linesearch 1(xk,σ,θ,δ).

    In optimization theory, to speed up the convergence of iterative methods, many mathematicians often use the inertial-type extrapolation [20,22] by adding the technical term βk(xkxk1). The control parameter βk is called an inertial parameter, which controls the momentum xkxk1. Using Linesearch 1, Cruz and Nghia [7] also introduced an accelerated algorithm with the inertial technical term as follows.

    Method 3. Let x0=x1domf2, α0=σ>0, θ(0,1), δ(0,12), and t1=1. For k1, let

    tk+1=1+1+4t2k2,βk=tk1tk+1, yk=Pdomf2(xk+βk(xkxk1)),xk+1=proxαkf2(ykαkf1(yk)),

    where αk:= Linesearch 1(yk,αk1,θ,δ).

    The technique of choosing βk in Method 3 was first mentioned in the fast iterative shrinkage-thresholding algorithm (FISTA) by Beck and Teboulle [5]. Weak convergence results of Methods 2 and 3 were obtained for solving Problem (1.1) with (H1) and (H2).

    Recently, Kankam et al. [14] proposed a modification of Linesearch 1 as follows.

    Linesearch 2. Given xdomf2, σ>0, θ(0,1) and δ>0.
    Input α=σ.
      While αmax{f1(FB2α(x))f1(FBα(x)),f1(FBα(x))f1(x)}
        >δ(FB2α(x)FBα(x)+FBα(x)x), do
        α=θα.
       End
    Output α,
    where FB2α(x):=FBα(FBα(x)).

     | Show Table
    DownLoad: CSV

    They showed the well-definedness of Linesearch 2 and introduced the following double forward-backward algorithm whose stepsize is generated by Linesearch 2.

    Method 4. Let x1domf2, σ>0, θ(0,1) and δ(0,18). For k1, let

    yk=proxαkf2(xkαkf1(xk)),xk+1=proxαkf2(ykαkf1(yk)),

    where αk:= Linesearch 2(xk,σ,θ,δ).

    A weak convergence theorem of Method 4 was proved and an application in signal recovery was illustrated, see [14].

    In this paper, inspired and motivated by the results of Cruz and Nghia [7] and Kankam et al. [14], and other related researches, we aim to improve Linesearches 1 and 2 and introduce a new accelerated algorithm using our proposed linesearch for the convex minimization problem of the sum of two convex functions. The paper is organized as follows. Basic definitions, notations and some useful tools for proving our convergence results are given in Section 2. Our main result is in Section 3. In this section, we introduce a new modification of Linesearches 1 and 2 and present a double forward-backward algorithm by using an inertial technique for solving Problem (1.1) with Hypotheses (H1) and (H2). After that, a weak convergence theorem of the proposed method is proved. The complexity of our reduced algorithm is also discussed. In Section 4, we apply the convex minimization problem to an image restoration problem and a regression problem. We analyze and illustrate the convergence behavior of our method, and also compare its efficiency with the well-known methods in the literature.

    The mathematical symbols used throughout this paper are as follows. R, R+ and N are the set of real numbers, the set of nonnegative real numbers, and the set of positive integers, respectively. Id stands for the identity operator on H. Denote weak and strong convergence of a sequence {xk}H to xH by xkx and xkx, respectively. The set of all weak-cluster points of {xk} is denoted by ωw(xk). If C is a nonempty closed convex subset of H, then PC stands for the metric projection from H onto C, i.e., for each xH, PCx is the unique element in C such that xPCx=dist(x,C):=infyCxy.

    Let us recall the concept of the proximity operator which extends the notion of the metric projection. Let f:HR{} be a proper, lower semi-continuous and convex function. The proximity (or proximal) operator [2,18] of f, denoted by proxf is defined for each xH, proxfx is the unique solution of the minimization problem:

    minimizeyHf(y)+12xy2. (2.1)

    If f:=iC is an indicator function on C (defined by iC(x)=0 if xC; otherwise iC(x)=), then proxf=PC.

    The proximity operator can be formulated in the equivalent form

    proxf=(Id+f)1:Hdomf, (2.2)

    where f is the subdifferential of f defined by

    f(x):={uH:f(x)+u,yxf(y),yH},xH.

    Moreover, we have the following useful fact:

    xproxαf(x)αf(proxαf(x)),xH,α>0. (2.3)

    The following is a property of the subdifferential operator.

    Lemma 2.1 ([23]). If f:HR{} is a proper, lower semi-continuous and convex function, then the graph of f defined by Gph(f):={(x,y)H×H:yf(x)} is demiclosed, i.e., if the sequence {(xk,yk)}Gph(f) satisfies that xkx and yky, then (x,y)Gph(f).

    We end this section by providing useful tools for proving our main results.

    Fact. Let x,yH. The following inequalities hold on H:

    x±y2=x2±2x,y+y2. (2.4)

    Lemma 2.2 ([12]). Let {ak} and {tk} be two sequences of nonnegative real numbers such that

    ak+1(1+tk)ak+tkak1,kN.

    Then the following holds

    ak+1Kkj=1(1+2tj),whereK=max{a1,a2}.

    Moreover, if k=1tk<, then {ak} is bounded.

    Lemma 2.3 ([31]). Let {ak} and {bk} be two sequences of nonnegative real numbers such that ak+1ak+bk for all kN. If k=1bk<, then limkak exists.

    Lemma 2.4 (Opial [19]). Let {xk} be a sequence in H such that there exists a nonempty set ΩH satisfying:

    (i) For every pΩ, limkxkp exists;

    (ii) ωw(xk)Ω.

    Then, {xk} converges weakly to a point in Ω.

    In this section, using the idea of Linesearches 1 and 2, we introduce a new linesearch and present an inertial double forward-backward algorithm with the proposed linesearch for solving the convex minimization problem of the sum of two convex functions without any Lipschitz continuity assumption on the gradient. A weak convergence result of our proposed method is analyzed and established.

    We now focus on Problem (1.1) and assume that f1 and f2 satisfy Hypotheses (H1) and (H2). The solution set of (1.1) is denoted by Ω. Also, suppose that Ω. For simplicity, we let F:=f1+f2 and denote by FBα the forward-backward operator of f1 and f2 with respect to α, that is, FBα:=proxαf2(Idαf1). Here, our linesearch is designed as follows.

    Linesearch 3. Given xdomf2, σ>0, θ(0,1), μ(0,1) and δ>0.
    Input α=σ.
      While α{(1μ)f1(FB2α(x))f1(FBα(x))+μf1(FBα(x))f1(x)}
        >δ(FB2α(x)FBα(x)+FBα(x)x), do
        α=θα.
      End
    Output α.

     | Show Table
    DownLoad: CSV

    Remark 3.1. The loop termination condition of Linesearch 3 is weaker than that of Linesearch 2. Indeed, since

    α{(1μ)f1(FB2α(x))f1(FBα(x))+μf1(FBα(x))f1(x)}αmax{f1(FB2α(x))f1(FBα(x)),f1(FBα(x))f1(x)},

    it follows from the well-definedness of Linesearch 2 that our linesearch also stops after finitely many steps, see [14, Lemma 3.2] for more details.

    Using Linesearch 3, we propose the following iterative method with the inertial technical term.

    Method 5.
    Initialization: Take x1=y0domf2, σ>0, θ(0,1), μ(0,12] and δ(0,μ4). Let {βk}R+.
    Iterative steps: For k1, calculate xk+1 as follows:
    Step 1. Compute
      zk=FBαk(xk)=proxαkf2(xkαkf1(xk)),(3.1)
      yk=FBαk(zk)=proxαkf2(zkαkf1(zk)),(3.2)
    where αk:= Linesearch 3(xk,σ,θ,μ,δ).
    Step 2. Compute
      xk+1=Pdomf2(yk+βk(ykyk1)).(3.3)
    Set k:=k+1 and return to Step 1.

     | Show Table
    DownLoad: CSV

    To verify the convergence of Method 5, the following result is needed.

    Lemma 3.2. Let {xk} be the sequence generated by Method 5. For each k1 and pdomf2, we have

    xkp2ykp22αk[F(yk)+F(zk)2F(p)]+(14δμ)(xkzk2+zkyk2).

    Proof. Let k1 and pdomf2. By (2.3), (3.1) and (3.2), we get

    xkzkαkf1(xk)f2(zk)andzkykαkf1(zk)f2(yk).

    Then, by the definition of the subdifferential of f2, we have

    f2(p)f2(zk)xkzkαkf1(xk),pzk=1αkxkzk,pzk+f1(xk),zkp (3.4)

    and

    f2(p)f2(yk)zkykαkf1(zk),pyk=1αkzkyk,pyk+f1(zk),ykp. (3.5)

    By the assumptions on f1 and f2, we have the following fact:

    f1(x)f1(y)f1(y),xy,xdomf1,ydomf2. (3.6)

    From (3.6), we get

    f1(p)f1(xk)f1(xk),pxk, (3.7)

    and

    f1(p)f1(zk)f1(zk),pzk. (3.8)

    Summing (3.4), (3.5), (3.7) and (3.8) yields

    2F(p)F(zk)f2(yk)f1(xk)f1(xk),zkp+f1(zk),ykp+f1(xk),pxk+f1(zk),pzk+1αk[xkzk,pzk+zkyk,pyk]=f1(xk),zkxk+f1(zk),ykzk+1αk[xkzk,pzk+zkyk,pyk]=f1(xk)f1(zk),zkxk+f1(zk),zkxk+f1(yk),ykzk+f1(zk)f1(yk),ykzk+1αk[xkzk,pzk+zkyk,pyk]f1(zk),zkxk+f1(yk),ykzkf1(xk)f1(zk)zkxkf1(zk)f1(yk)ykzk+1αk[xkzk,pzk+zkyk,pyk].

    Using (3.6) again, the above inequality becomes

    2F(p)F(zk)f2(yk)f1(xk)f1(yk)f1(xk)f1(xk)f1(zk)zkxkf1(zk)f1(yk)ykzk+1αk[xkzk,pzk+zkyk,pyk]f1(yk)f1(xk)f1(xk)f1(zk)(ykzk+zkxk)f1(zk)f1(yk)(ykzk+zkxk)+1αk[xkzk,pzk+zkyk,pyk]=f1(yk)f1(xk)+1αk[xkzk,pzk+zkyk,pyk](f1(xk)f1(zk)+f1(zk)f1(yk))(ykzk+zkxk). (3.9)

    Since αk:= Linesearch 3(xk,σ,θ,μ,δ), we have

    αk{(1μ)f1(yk)f1(zk)+μf1(zk)f1(xk)}δ(ykzk+zkxk). (3.10)

    It follows from (3.9) and (3.10) that

    1αk[xkzk,zkp+zkyk,ykp]F(yk)+F(zk)2F(p)(f1(xk)f1(zk)+f1(zk)f1(yk))(ykzk+zkxk)F(yk)+F(zk)2F(p)1μ{(1μ)f1(yk)f1(zk)+μf1(zk)f1(xk)}(ykzk+zkxk)F(yk)+F(zk)2F(p)δαkμ(ykzk+zkxk)2F(yk)+F(zk)2F(p)2δαkμ(ykzk2+zkxk2). (3.11)

    From (2.4), we get

    xkzk,zkp=12(xkp2xkzk2zkp2), (3.12)

    and

    zkyk,ykp=12(zkp2zkyk2ykp2). (3.13)

    Therefore, we conclude from (3.11)-(3.13) that

    xkp2ykp22αk[F(yk)+F(zk)2F(p)]+(14δμ)(xkzk2+zkyk2).

    Now we are ready to prove a convergence result of Method 5, which is the main theorem of this paper.

    Theorem 3.3. Let {xk} be the sequence generated by Method 5. If αkα>0, βk0 for all kN and k=1βk<, then {xk} converges weakly to a point in Ω.

    Proof. Let pΩ, αkα>0, βk0 for all kN and k=1βk<. By applying Lemma 3.2, we have

    xkp2ykp22αk[F(yk)F(p)+F(zk)F(p)]+(14δμ)(xkzk2+zkyk2)(14δμ)xkzk20, (3.14)

    which implies

    ykpxkp. (3.15)

    Let x0=x1. By (3.3) and (3.15), we have

    xk+1p=Pdomf2(yk+βk(ykyk1))pyk+βk(ykyk1)pykp+βkykyk1(1+βk)ykp+βkyk1p(1+βk)xkp+βkxk1p,kN. (3.16)

    By Lemma 2.2, we have xk+1pKkj=1(1+2βj), where K:=max{x1p,x2p}, and {xk} is bounded. Thus, k=1βk(xkp+xk1p)<. In view of (3.16), it follows from Lemma 2.3 that limkxkp exists. Also, we have limkxkp=limkykp. Then, by (3.14), we get

    limkxkzk=0. (3.17)

    Next, we show that ωw(xk)Ω. Pick ˉpωw(xk), then there is a subsequence {xki} of {xk} such that xkiˉp. Thus, zkiˉp. By (H2), we have

    limkf1(xk)f1(zk)=0. (3.18)

    From (2.3), we get

    xkizkiαki+f1(zki)f1(xki)f2(zki)+f1(zki)=F(zki). (3.19)

    Here, by zkiˉp and (3.17)-(3.19), it follows from Lemma 2.1 that 0F(ˉp), that is, ˉpΩ. So, ωw(xk)Ω. Using Lemma 2.4, we obtain that xkp for some pΩ. This completes the proof.

    The following method is obtained by reducing the inertial step in Method 5.

    Method 6.
    Initialization: Take x1domf2, σ>0, θ(0,1), μ(0,12] and δ(0,μ4).
    Iterative steps: For k1, compute
       zk=FBαk(xk)=proxαkf2(xkαkf1(xk)),   (3.20)
       xk+1=FBαk(zk)=proxαkf2(zkαkf1(zk)),    (3.21)
    where αk:= Linesearch 3(xk,σ,θ,μ,δ).

     | Show Table
    DownLoad: CSV

    We also analyze the convergence and the complexity of Method 6.

    Theorem 3.4. Let {xk} be the sequence generated by Method 6. If αkα>0 for all kN, then the following statements hold:

    (i) {xk} converges weakly to a point in Ω;

    (ii) If δ(0,μ8), then {F(xk)} is decreasing and

    F(xk)minxHF(x)[dist(x1,Ω)]22αk,kN. (3.22)

    Proof. Setting βk:=0 in Method 5, (i) is obtained directly from Theorem 3.3. To prove (ii), let δ(0,μ8). We first show that {F(xk)} is decreasing. By Lemma 3.2, we get

    xkp2xk+1p22αk[F(xk+1)+F(zk)2F(p)]+(14δμ)(xkzk2+zkxk+12),pdomf2. (3.23)

    Taking p:=zk in (3.23), we have

    xkzk2xk+1zk22αk[F(xk+1)F(zk)]+(14δμ)(xkzk2+zkxk+12). (3.24)

    Also, taking p:=xk in (3.23), we have

    xk+1xk22αk[F(xk+1)+F(zk)2F(xk)]+(14δμ)(xkzk2+zkxk+12). (3.25)

    Combining (3.24) and (3.25), we get

    xkzk2xk+1zk2xk+1xk24αk[F(xk+1)F(xk)]+2(14δμ)(xkzk2+zkxk+12),

    which implies

    F(xk+1)F(xk)14αk{(18δμ)xkzk2+(38δμ)zkxk+12+xk+1xk2}.

    This together with δ(0,μ8) yields F(xk+1)F(xk), that is, {F(xk)} is decreasing. Next, we will show that (3.22) is true. Pick any pΩ. From (3.23), we have

    xkp2xk+1p22αk[F(xk+1)F(p)+F(zk)F(p)]+(14δμ)(xkzk2+zkxk+12)2α[F(xk+1)F(p)].

    Hence

    F(xi+1)F(p)12α(xip2xi+1p2),iN.

    It follows that

    ki=1F(xi+1)kF(p)12αki=1(xip2xi+1p2)=12α(x1p2xk+1p2)x1p22α.

    By the decreasing property of {F(xk)}, we get

    k[F(xk)F(p)]ki=1F(xi+1)kF(p)x1p22α,pΩ. (3.26)

    It follows from (3.26) that

    F(xk)minxHF(x)infpΩx1p22αk=[dist(x1,Ω)]22αk.

    We note that the stepsize condition on {αk} in Theorems 3.3 and 3.4 needs the boundedness from below by a positive real number. Next, we show that this condition can be guaranteed by the Lipschitz continuity assumption on the gradient.

    Let CH be a nonempty closed convex set. Recall that an operator T:CH is said to be Lipschitz continuous if there exists L>0 such that TxTyLxy for all x,yC.

    Proposition 3.5. Let {αk} be the sequence generated by Linesearch 3 of Method 5 (or Method 6). If f1 is Lipschitz continuous on domf2 with a coefficient L>0, then αkmin{σ,δθL} for all kN.

    Proof. Let f1 be L-Lipschitz continuous on domf2. From Linesearch 3, we know that αkσ for all kN. If αk<σ, then αk=σθmk where mk is the smallest positive integer such that

    αk{(1μ)f1(FB2αk(xk))f1(FBαk(xk))+μf1(FBαk(xk))f1(xk)}δ(FB2αk(xk)FBαk(xk)+FBαk(xk)xk).

    Let ¯αk:=αkθ>0. By the Lipschitz continuity of f1 and the above expression, we have

    ¯αkL(FB2¯αk(xk)FB¯αk(xk)+FB¯αk(xk)xk)¯αk(f1(FB2¯αk(xk))f1(FB¯αk(xk))+f1(FB¯αk(xk))f1(xk))¯αk{(1μ)f1(FB2¯αk(xk))f1(FB¯αk(xk))+μf1(FB¯αk(xk))f1(xk)}>δ(FB2¯αk(xk)FB¯αk(xk)+FB¯αk(xk)xk),

    which implies that αk>δθL. Consequently, αkmin{σ,δθL} for all kN.

    Remark 3.6. It is worth mentioning again that the Lipschitz continuity assumption on the gradient of f1 is sufficient for Hypothesis (H2). However, if we assume this assumption further, the calculation of the stepsize αk generated by Linesearch 3 is still independent of the Lipschitz constant.

    In this section, we apply the convex minimization problem, Problem (1.1), to an image restoration problem and a regression problem. To slove these problems, we analyze and illustrate the convergence behavior of our method (Method 5) and compare its efficiency with Methods 1-4 and 6. We also consider the following method with another linesearch strategy [21]:

    Linesearch 4. Given xdomf2, ˉσ>0, and ˉθ(0,1).
    Input α=ˉσ.
       While F(FBα(x))>f1(x)+FBα(x)x,f1(x)+12αFBα(x)x2+f2(FBα(x)), do
         α=ˉθα.
       End
    Output α.

     | Show Table
    DownLoad: CSV

    Method 7. Let x1=y0domf2, ˉσ>0, ˉθ(0,1), ρk>0, and t1=1. For k1, let

    yk=proxαkf2(xkαkf1(xk)),tk+1=1+1+4ρkt2k2,βk=tk1tk+1, xk+1=yk+βk(ykyk1),

    where αk:= Linesearch 4(xk,ˉσ,ˉθ).

    Note that Method 7 is well known as the FISTA with backtracking, see [5,25].

    All experiments and visualizations are done with MATLAB and are performed on a laptop computer (Intel Core-i5/4.00 GB RAM/Windows 8/64-bit).

    Many problems rising in image and signal processing, especially the image restoration problem, can be formulated as the following equation:

    y=Ax+ε, (4.1)

    where xRN is an original image, yRM is the observed image, ε is an additive noise and ARM×N is the blurring operation. To approximate the original image, we need to minimize the value of ε by using the LASSO model [28]:

    minxRN{12yAx22+λx1}, (4.2)

    where λ>0 is a regularization parameter, 1 is the l1-norm, and 2 is the Euclidean norm. In this situation, we apply Problem (1.1) to the LASSO model (4.2) by setting

    f1(x)=12yAx22andf2(x)=λx1,

    where A=RW, R is the matrix representing the blur operator and W is the inverse of a three stage Haar wavelet transform. We use the peak signal-to-noise ratio (PSNR) in decibel (dB) [30] and the structural similarity index metric (SSIM) [33] as the image quality measures, which are formulated as follows:

    PSNR(xk)=10log10(2552MSE(k)),

    where MSE(k)=1Mxkx22, M is the number of image samples, and x is the original image and

    SSIM(xk,x)=(2νxkνx+C1)(2ξxkx+C2)(ν2xk+ν2x+C1)(ξ2xk+ξ2x+C2),

    where {νxk,ξxk} and {νx,ξx} denote the mean intensity and standard deviation set of the deblerring image xk and the original image x, respectively, ξxnx denotes its cross correlation, and C1, C2 are small constants value to avoid instability problem when the denominator is too close to zero.

    Example 4.1. Consider two grayscale images, Cameraman (see Figure 1 (a)) and Boy (see Figure 2 (a)), with size of 256×256, which are contaminated by Gaussian blur of filter size 9×9 with standard deviation ˆσ=4, noise 105 (see Figure 1 (b) and Figure 2 (b)).

    Figure 1.  Restoration for Cameraman image. (c)-(i): Restored images by the deblurring methods.
    Figure 2.  Restoration for Boy image. (c)-(i): Restored images by the deblurring methods.

    Firstly, we test the efficiency of Method 5 for recovering the Caneraman image at the 300th iteration by choosing various linesearch parameters and different starting points as shown in Table 1.

    Table 1.  The test for recovering the Caneraman image of Method 5.
    Initial points x1 Linesearch parameters PSNR SSIM
    σ θ μ δ
    Blurred image 5 0.1 0.1 0.02 31.554 0.6193
    1 0.5 0.1 0.02 31.096 0.6081
    0.1 0.9 0.5 0.12 28.721 0.5717
    10 0.9 0.5 0.12 33.051 0.6708
    Random selection 5 0.1 0.1 0.02 30.554 0.6043
    1 0.5 0.1 0.02 29.660 0.5934
    0.1 0.9 0.5 0.12 28.694 0.5640
    10 0.9 0.5 0.12 32.798 0.6554

     | Show Table
    DownLoad: CSV

    It is observed from Table 1 that the linesearch parameter σ influences the image recovery performance of Method 5, while choosing the starting point x1 has no significant impact on the convergence behavior of our method.

    Next, we test and analyze the image recovery performance of Methods 1-7 for recovering the images (Cameraman and Boy) by setting the parameters as in Table 2 and by choosing the blurred images as the starting points. The maximum iteration number for all methods is fixed at 500. The comparative experiments for recovering the Cameraman and Boy images are as follows. Original images, blurry image contaminated by Gaussian blur and restored images by Methods 1-7 are shown in Figures 1 and 2. The PSNR and SSIM results are shown in Figures 3 and 4. It can be seen that Method 5 gives the higher values of PSNR and SSIM than the others, meanwhile Method 6 gives the better results than the other methods without the inertial step. Therefore, our method has the highest image recovery efficiency comparing with the studied methods.

    Table 2.  The parameters for the deblurring methods.
    Methods
    Parameters 1 2 3 4 5 6 7
    λ=105
    αk=k(k+1)L, L=λmax(AA) - - - - - -
    σ=10, θ=0.9, δ=0.1 - - -
    μ=0.5 - - - - -
    βk={kk+1if1k50012kotherwise - - - - - -
    ˉσ=10, ˉθ=0.9, ρk=1 - - - - - -

     | Show Table
    DownLoad: CSV
    Figure 3.  Plot of PSNR and SSIM results by the deblurring methods.
    Figure 4.  The comparison of PSNR and SSIM values for blurred images and restored images by the deblurring methods.

    We first introduce a brief concept of extreme learning machine (ELM) for regression problems. Let S={(xk,tk):xkRn,tkRm,k=1,2,...,N} be a training set of N distinct samples where xk is an input data and tk is a target. For any single hidden layer of ELM, the output of the i-th hidden node is

    hi(x)=G(ai,bi,x),

    where G is an activation function, ai and bi are parameters of the i-th hidden node. The output function of ELM for a single-hidden layer feedforward neural networks (SLFNs) with M hidden nodes is

    Ok=Mi=1ξihi(xk),

    where ξi is the output weight of the i-th hidden node. The hidden layer output matrix H is defined by

    H=[G(a1,b1,x1)G(aM,bM,x1)G(a1,b1,xN)G(aM,bM,xN)]N×M.

    The main goal of ELM is to find ξ=[ξ1,...,ξM] such that

    Hξ=T, (4.3)

    where T=[t1,...,tN]. From (4.3), we can estimate the weight ξ by ξ=HT where H=(HH)1H is the pseudo-inverse matrix of H.

    Example 4.2. This example aims to predict the sine function by using the ELM method defined as follows:

    Create randomly 10 distinct points y1,y2,...,y10[4,4];

    Create a training matrix R=[y1y2...y10] and a training target matrix S=[sin(y1)sin(y2)...sin(y10)];

    Create a testing matrix V=[43.993.98...4] and a target matrix T=[sin(4)sin(3.99)...sin(y10)];

    Using sigmoid as an activation function generates the hidden layer output matrix H1 of the training matrix R with the hidden node M=100;

    Choose a regularization parameter λ=105. Formulate a convex minimization problem via the LASSO model that

    f1(ξ)=12H1ξS22andf2(ξ)=λξ1,

    and find the optimal weight ξk of this problem by employing Methods 1-5 with certain number of iterations;

    Using sigmoid as an activation function generates the hidden layer output matrix H2 of the testing matrix V with the hidden node M=100;

    Calculate the output Ok=H2ξk and calculate the mean squared error (MSE) at k-th iteration by

    MSE(k)=1NOkT22,

    where N is the number of distinct samples.

    The parameters for each prediction method are set as in Table 3.

    Table 3.  The parameters for the prediction methods.
    Methods
    Parameters 1 2 3 4 5 6 7
    αk=k(k+1)L, L=λmax(HH) - - - - - -
    σ=0.1, θ=0.49, δ=0.1 - -
    μ=0.5 - - - - -
    βk={kk+1if1k1000012kotherwise - - - - - -
    ˉσ=105, ˉθ=0.49, ρk=1 - - - - - -

     | Show Table
    DownLoad: CSV

    Now a graph of predicting the sine function by Methods 1-7 at the 500th iteration is shown in Figure 5. We show a comparative result of iteration numbers, mean squared error values and times in Table 4 when the stopping criterion is set as: MSE103 or at the 10000th iteration. Also, a graph of the mean squared error is shown in Figure 6. It can be observed that Method 5 gives a better performance for predicting the sine function than the other tested methods.

    Figure 5.  The regression for the sine function by the prediction methods at the 500th iteration.
    Table 4.  The comparative result by the prediction methods.
    Methods Iteration No. Time (s) MSE
    1 10000 0.7986 0.0408
    2 10000 12.1314 0.0032
    3 3999 1.3463 9.9995×104
    4 10000 17.7834 0.0014
    5 338 0.6531 9.9239×104
    6 10000 18.3197 0.0014
    7 10000 3.0340 0.0067

     | Show Table
    DownLoad: CSV
    Figure 6.  Plot of the MSE value by the prediction methods.

    In this paper, we discuss the convex minimization problem of the sum of two convex functions in a Hilbert space. The challenge of removing the Lipschitz continuity assumption on the gradient of the function attracts us to study the concept of the linesearch process. We introduce a new linesearch and propose an inertial forward-backward algorithm whose stepsize does not depend on any Lipschitz constant for solving the considered problem without any Lipschitz continuity condition on the gradient. It is shown that the sequence generated by our proposed method converges weakly to a minimizer of the sum of those two convex functions under some mild control conditions. As applications, we employ our method to recover blurry images in an image restoration problem and predict the sine function in a regression problem. The results of the experiments show that our method has a higher efficiency than the well-known methods in [5,7,14,15,25].

    This research was supported by Thailand Science Research and Innovation under the project IRN62W0007 and Chiang Mai University.

    The authors declare no conflict of interest.



    [1] C. Byrne, Iterative oblique projection onto convex sets and the split feasibility problem, Inverse Probl., 18 (2002), 441-453. doi: 10.1088/0266-5611/18/2/310
    [2] H. H. Bauschke, P. L. Combettes, Convex analysis and monotone operator theory in Hilbert spaces, New York: Springer, 2011.
    [3] L. Bussaban, S. Suantai, A. Kaewkhao, A parallel inertial S-iteration forward-backward algorithm for regression and classification problems, Carpathian J. Math., 36 (2020), 35-44. doi: 10.37193/CJM.2020.01.04
    [4] D. P. Bertsekas, J. N. Tsitsiklis, Parallel and distributed computation numerical methods, Belmont: Athena Scientific, 1997.
    [5] A. Beck, M. Teboulle, A fast iterative shrinkage-thresholding algorithm for linear inverse problems, SIAM J. Imaging Sci., 2 (2009), 183-202. doi: 10.1137/080716542
    [6] P. L. Combettes, J. C. Pesquet, A Douglas-Rachford splitting approach to nonsmooth convex variational signal recovery, IEEE J. STSP, 1 (2007), 564-574.
    [7] J. Y. B. Cruz, T. T. A. Nghia, On the convergence of the forward-backward splitting method with linesearchs, Optim. Method. Softw., 31 (2016), 1209-1238. doi: 10.1080/10556788.2016.1214959
    [8] P. L. Combettes, V. R. Wajs, Signal recovery by proximal forward-backward splitting, Multiscale Model. Simul., 4 (2005), 1168-1200. doi: 10.1137/050626090
    [9] J. C. Dunn, Convexity, monotonicity, and gradient processes in Hilbert space, J. Math. Anal. Appl., 53 (1976), 145-158. doi: 10.1016/0022-247X(76)90152-9
    [10] I. Daubechies, M. Defrise, C. D. Mol, An iterative thresholding algorithm for linear inverse problems with a sparsity constraint, Commun. Pure Appl. Math., 57 (2004), 1413-1457. doi: 10.1002/cpa.20042
    [11] M. Figueiredo, R. Nowak, An EM algorithm for wavelet-based image restoration, IEEE T. Image Process, 12 (2003), 906-916. doi: 10.1109/TIP.2003.814255
    [12] A. Hanjing, S. Suantai, A fast image restoration algorithm based on a fixed point and optimization, Mathematics, 8, (2020), 378.
    [13] E. Hale, W. Yin, Y. Zhang, A fixed-point continuation method for l1-regularized minimization with applications to compressed sensing, Rice University: Department of Computational and Applied Mathematics, 2007.
    [14] K. Kankam, N. Pholasa, P. Cholamjiak, On convergence and complexity of the modified forward-backward method involving new linesearches for convex minimization, Math. Meth. Appl. Sci., 42 (2019), 1352-1362. doi: 10.1002/mma.5420
    [15] P. L. Lions, B. Mercier, Splitting algorithms for the sum of two nonlinear operators, SIAM J. Numer. Anal., 16 (1979), 964-979. doi: 10.1137/0716071
    [16] L. J. Lin, W. Takahashi, A general iterative method for hierarchical variational inequality problems in Hilbert spaces and applications, Positivity, 16 (2012), 429-453. doi: 10.1007/s11117-012-0161-0
    [17] B. Martinet, Régularisation d'inéquations variationnelles par approximations successives, Rev. Française Informat. Rech. Opér., 4 (1970), 154-158.
    [18] J. J. Moreau, Fonctions convexes duales et points proximaux dans un espace hilbertien, C. R. Acad. Sci. Paris Sér. A Math., 255 (1962), 2897-2899.
    [19] A. Moudafi, E. Al-Shemas, Simultaneous iterative methods for split equality problem, Trans. Math. Program. Appl., 1 (2013), 1-11.
    [20] Y. Nesterov, A method for solving the convex programming problem with convergence rate O(1/k2), Dokl. Akad. Nauk SSSR., 269 (1983), 543-547.
    [21] Y. Nesterov, Gradient Methods for Minimizing Composite Objective Function, CORE Report, 2007. Available from: http://www.ecore.be/DPs/dp 1191313936.pdf.
    [22] B. T. Polyak, Some methods of speeding up the convergence of iteration methods, USSR Comput. Math. Math. Phys., 4 (1964), 1-17.
    [23] R. T. Rockafellar, On the maximal monotonicity of subdifferential mappings, Pac. J. Math., 33 (1970), 209-216. doi: 10.2140/pjm.1970.33.209
    [24] R. T. Rockafellar, Monotone operators and the proximal point algorithm, SIAM J. Control Optim., 17 (1976), 877-898.
    [25] K. Scheinberg, D. Goldfarb, X. Bai. Fast first order methods for composite convex optimization with backtracking, Found. Comput. Math., 14 (2014), 389-417. doi: 10.1007/s10208-014-9189-9
    [26] S. Suantai, K. Kankam, P. Cholamjiak, A novel forward-backward algorithm for solving convex minimization problem in Hilbert spaces, Mathematics, 8 (2020), 42. doi: 10.3390/math8010042
    [27] S. Suantai, P. Jailoka, A. Hanjing, An accelerated viscosity forward-backward splitting algorithm with the linesearch process for convex minimization problems, J. Inequal. Appl., 2021 (2021), 42. doi: 10.1186/s13660-021-02571-5
    [28] R. Tibshirani, Regression shrinkage and selection via the lasso, J. R. Stat. Soc. B Methodol., 58 (1996), 267-288.
    [29] P. Tseng, A modified forward-backward splitting method for maximal monotone mappings, SIAM J. Control Optim., 38 (2000), 431-446. doi: 10.1137/S0363012998338806
    [30] K. Thung, P. Raveendran, A survey of image quality measures, In: Proceedings of the International Conference for Technical Postgraduates (TECHPOS), Kuala Lumpur, Malaysia, 14-15 December 2009, 1-4.
    [31] K. Tan, H. K. Xu, Approximating fixed points of nonexpansive mappings by the ishikawa iteration process, J. Math. Anal. Appl., 178 (1993), 301-308. doi: 10.1006/jmaa.1993.1309
    [32] K. Toh, S. Yun. An accelerated proximal gradient algorithm for nuclear norm regularized linear least squares problems, Pac. J. Optim., 6 (2010), 615-640.
    [33] Z. Wang, A. C. Bovik, H. R. Sheikh, E. P. Simoncelli, Image quality assessment: from error visibility to structural similarity, IEEE T. Image Process., 13 (2004), 600-612. doi: 10.1109/TIP.2003.819861
  • This article has been cited by:

    1. José A. Vásquez-Coronel, Marco Mora, Karina Vilches, A Review of multilayer extreme learning machine neural networks, 2023, 56, 0269-2821, 13691, 10.1007/s10462-023-10478-4
    2. Adisak Hanjing, Panadda Thongpaen, Suthep Suantai, A new accelerated algorithm with a linesearch technique for convex bilevel optimization problems with applications, 2024, 9, 2473-6988, 22366, 10.3934/math.20241088
  • Reader Comments
  • © 2021 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2495) PDF downloads(189) Cited by(2)

Figures and Tables

Figures(6)  /  Tables(4)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog