1.
Introduction
Convex optimization problem (COP) is considered very important in the current literature as it covers a diverse range of problems with possible applications in signal processing, image processing and machine learning. As a consequence, the tremendous progress in studying the COP has led the emergence of a theory of convex optimization and a useful interface linking various branches of sciences.
Monotone operator theory is a prominent framework for various nonlinear problems and closely related with the theory of convex optimization. One of the fundamental themes in monotone operator theory is to compute zeros of the (maximal-) monotone operators. The importance of this concept stems from the fact that the sub-differential operator associated with a proper, convex and lower semicontinuous (PCLS) function is not only a maximal monotone operator but also solves the convex minimization problem. It is remarked that most of the practical phenomenon can be reformulated as zero point problem. This formalism includes variational inequalities, evolution equations, complementarity problems and inclusions [12].
The class of split feasibility problems (SFP) has an extraordinary utility and broad applicability in medical image reconstruction, signal processing and computerized tomography [15,17,18,21]. Some interesting and crucial results regarding the SFP with areas of feasible applications are established in [16,19,20]. The first prototype strategies for computing the optimal solution of the split common null point problem (SCNPP) can be found in [16]. Since then, different variants of these strategies have been proposed and analyzed for SCNPP and other instances of SFP [19,20,29].
Another useful formalism for modelling various nonlinear phenomenon is the fixed point problem (FPP) of the operator under consideration. Most of the problems in diverse areas such as mathematical economics, variational inequality theory, control theory and game theory can be reformulated in terms of FPP. It is remarked that various nonlinear fixed point operators play equivalent important role in COP. In 2015, Takahashi et al.[31] investigated a unified formalism of null point problem and FPP in Hilbert spaces. Since then, FPP associated with different nonlinear operators are jointly investigated with (split common-) null point problem in this domain. It is therefore natural to investigate FPP associated with an infinite family of operators jointly with SCNPP in Hilbert spaces.
A variety of strategies combining iterative optimization algorithms and fixed point algorithms have been introduced and analyzed to construct an optimal solution of the SCNPP and FPP. Each strategy has certain shortcomings in terms of convergence characteristic and/or rate of convergence. The hybrid shrinking projection algorithm is a well-known strategy for the strong convergence characteristic whereas the inertial extrapolation technique, essentially due to [27] and see also [1,2,3,4,5,6,7,8,9,10,11,32], enhances the rate of convergence of the algorithm under consideration.
Our main contributions in this ongoing fruitful research direction are as follows:
(1) We posit a framework to jointly investigate SCNPP and FPP associated with an infinite family of operators in Hilbert spaces. For the case of an infinite family of fixed point operators, we exploit the construction of an auxiliary operator defined in [28,34];
(2) We employ an algorithmic approach combining the hybrid shrinking projection algorithm with the inertial extrapolation technique to construct the common optimal solution of the problems as mentioned in item (1);
(3) We establish the strong convergence analysis of the proposed algorithm by employing the suitable constraints in accordance with the standard techniques and tools in the current literature;
(4) We posit different frameworks, as an application of the framework mentioned in item (1), for various instances of SFP in Hilbert spaces;
(5) Last but not least, we incorporate an appropriate numerical example for the demonstration of the framework as well as the applicability of the proposed algorithm for signal recovery problem.
2.
Preliminaries
Throughout the rest of the sections, the triplet (Ξ1,<⋅,⋅>,‖⋅‖) indicates the real Hilbert space with the conventional notations of the inner product and the norm and A1⊆Ξ1×Ξ1 denotes a set-valued operator with the usual definitions of dom(A1), gra(A1) and zer(A1). We use → (resp. ⇀) to represent the symbol for strong convergence (resp. weak convergence). The set of reals and natural numbers are symbolized as R and N, respectively.
Recall from the celebrated monograph [12] that the set-valued operator A1 is called monotone if ⟨r−s,u−v⟩≥0,∀(r,u),(s,v)∈gra(A1). In addition, A1 is coined as maximal monotone operator provided that the graph of A1 could not be properly contained in the graph of any other monotone operator. Let m>0, then the resolvent operator of A1 is defined as JA1m=(Id+mA1)−1, where Id denotes the identity operator. In this connection, JA1m is well-defined, single-valued and firmly nonexpansive operator.
Let T:H→H be an operator defined on a nonempty subset H of Ξ1. We use Fix(T)={p∈H|p=Tp} to denote the set of fixed points of the operator T. The metric projection operator ΠH:Ξ1→H associated with the nonempty closed convex subset H of Ξ1 is defined as ΠH(u)=argminv∈H‖u−v‖. It is well-known that the operator ΠH is firmly nonexpansive and satisfies ⟨u−ΠHu,ΠHu−v⟩≥0,∀u∈Ξ1,v∈H. Recall that the operator T is known as k-demicontractive [24] provided that k∈[0,1) such that
The class of k-demicontractive operators has been studied extensively in various instances of fixed point problems in Hilbert spaces. However, we are concerned with the fixed point problem of an infinite family of k-demicontractive operators in Hilbert spaces via the following construction of auxiliary operator Sk:
where 0≤βm≤1 and T′m=αx+(1+α)Tmx for all x∈H with Tm being k-demicontractive operator and α∈[k,1). It is well-known in the context of operator Sk that each T′m is nonexpansive and the limit limk→∞Qk,m exists. Moreover
This implies that Fix(S)=⋂∞k=1Fix(Sk) [28,34].
We now finally introduce the formalism of the proposed problem.
Let A1⊆Ξ1×Ξ1 and A2⊆Ξ2×Ξ2 be maximal monotone operators such that the domain of A1 is the subset of H and let JA1m and JA2m be the resolvents of A1 and A2, respectively, for m>0. Let h:Ξ1→Ξ2 be a bounded linear operator and let h∗ be the adjoint operator of h. Let Sk be the S-operator such that Γ:=Ω∩Fix(S)≠∅, where Ω:={ˆp∈A−11(0):hˆp∈A−12(0)} indicates the SCNPP [16]. We aim to find
The following crucial results are required in the sequel:
Lemma 2.1. [14] Let U:H→H be an operator defined on a nonempty closed convex subset H of a real Hilbert space Ξ1 and let (pk) be a sequence in H. If pk⇀p and if (Id−U)pk→0, then p∈Fix(U).
Lemma 2.2. Let μ,ν∈Ξ1 and θ∈R then
(i) ‖μ+ν‖2≤‖μ‖2+2⟨ν,μ+ν⟩;
(ii) ‖μ−ν‖2≤‖μ‖2−‖ν‖2−2⟨μ−ν,ν⟩;
(iii) ‖θμ+(1−θ)ν‖2=θ‖μ‖2+(1−θ)‖ν‖2−θ(1−θ)‖μ−ν‖2.
Lemma 2.3. [34] Let H be a nonempty closed and convex subset of a real Hilbert space Ξ1 and let T:H→H be a k-demicontractive operator with k∈[0,1). Let α∈[k,1) and set T′=(1−α)Id+αT, then T′:H→H is a nonexpansive operator such that Fix(T′)=Fix(T).
Lemma 2.4. [26] Let H be a nonempty closed convex subset of a real Hilbert space Ξ1. For every p,q,t∈Ξ1 and γ∈R, the set
is closed and convex.
Lemma 2.5. [34] Let H be a nonempty closed and convex subset of a real Hilbert space Ξ1 and let (T′m) be a sequence of nonexpansive operators such that ⋂∞k=1Fix(T′k)≠∅ and 0≤βm≤b<1. Then for a bounded subset K of H, we have
3.
Algorithm and convergence analysis
We start with the architecture of the algorithm for the construction of an optimal solution of the problem (2.1).
Theorem 3.1. The sequence (pk) generated by the Algorithm 1, under the following control conditions,
(C1) ∑∞k=1θk‖pk−pk−1‖<∞;
(C2) 0<a≤lim infk→∞αk≤lim supk→∞αk≤a∗;
(C3) lim infk→∞βk>0;
(C4) lim infk→∞mk>0;
converges strongly to an element p∗∈Γ.
Proof. Step 1. The Algorithm 1 is well-defined.
Recall that the set Γ is closed and convex whereas the closedness and the convexity of Hk+1, for each k≥1, follows from Lemma 2.4. Let p∗∈Γ, then recalling the Algorithm 1, we have
Also from Algorithm 1 and Lemma 2.3, we have
Further, we obtain
Recalling the nonexpansivity of JA1mk, we obtain
Denote λk=2δ⟨hˉwk−hp∗,(JA2mk−Id)hˉwk⟩ and recalling the firm nonexpansivity of JA2mk, we get
The estimate (3.3) implies after recalling the estimates (3.4) and (3.5)
The above estimate (3.6) indicates the inclusion Γ⊂Hk+1. Summarising the above stated facts, this infers that the Algorithm 1 is well-defined.
Step 2. The limit limk→∞‖pk−p1‖ exists.
Note that ‖pk+1−p1‖≤‖p−p1‖, for all p∈Hk+1 by employing the fact that pk+1=ΠHk+1p1. This infers that ‖pk+1−p1‖≤‖p∗−p1‖, for all p∗∈Γ⊂Hk+1 and consequently establishes the boundedness of (‖pk−p1‖). Also from pk=ΠHkp1, we have that
The above approximation infers that the sequence (‖pk−p1‖) is non-decreasing, therefore, we have
Step 3. Prove that q∈Γ.
The following crucial estimates are required to prove the claim:
By employing the limsup, and recalling the estimate (3.7), the above estimate implies that lim supk→∞‖pk+1−pk‖2=0. That is
As an easy inference of the approximates (3.8) and (3.9), we get
Since pk+1∈Hk+1, we have
Recalling the estimate (3.8) and the condition (C1), the above estimate implies that
Recalling the estimates (3.8), (3.11) and the following triangular inequality:
we get
Consider the estimate (3.6) in the following variation:
Letting k→∞ and recalling the conditions (C1)–(C2) and the estimate (3.12), we have
The estimate (3.13) also implies that
Recalling the estimates (3.8), (3.14) and the following triangular inequality:
we get
Now recalling the estimates (3.4), (3.5) and Lemma 2.2, we have
Rearranging the estimate (3.16), we have
Now recalling the conditions (C1), (C3), the estimate (3.9) and δ∈(0,2‖h‖2), the estimate (3.17) implies that
Recalling the estimates (3.4) and (3.5), we obtain
Denote ξk=JA1mk(ˉwk+δh∗(JA2mk−Id)hˉwk) and recalling the estimate (3.19), it follows that
That is
As a consequence, we have
The estimate (3.22), gives that
Recalling the estimate (3.18) and the condition (C3), we have
Reasoning as above, by recalling the definition of (ek), the condition (C1) and the estimate (3.24), we get
Note that the existence of a convergent subsequence (pkj) of (pk) such that pkj⇀q∈Ξ1 as j→∞ follows from the boundedness of (pk). This also infers that ξkj⇀q and ˉwkj⇀q as j→∞. To establish the claim, we first prove that q∈Ω.
Let (u,v)∈gra(A1). Since ξkj=JA1mkj(ˉwkj+δh∗(JA2mkj−Id)hˉwkj), therefore, we have
This implies that
Recalling the monotonicity of A1, we have
The above estimate infers that
Since ξkj⇀q, we obtain limj→∞⟨u−ξkj,v⟩=⟨u−q,v⟩. Now utilizing the estimates (3.18), (3.24) and (3.26), we have ⟨u−q,v⟩≥0. This implies that 0∈A1q.
Recalling the facts that (i) h is a bounded linear operator implies that hˉwkj⇀hq as j→∞, (ii) JA2mk is a nonexpansive operator with Id−JA2mk being demiclosed at the origin (Lemma 2.1), we also obtain that 0∈A2(hq). Hence q∈Ω.
We now show that q∈Fix(S). Observe that
Recalling the estimate (3.13) and Lemma 2.5, the above estimate implies that limk→∞‖ek−Sek‖=0. This together with the fact that ekj⇀q implies, with the help of Lemma 2.1, that q∈Fix(S)=⋂∞k=1Fix(Sk).
Step 4. Prove that pk→p∗=ΠΓp1.
Let p∗=ΠΓp1. As pk+1=ΠHk+1p1 and Γ⊂Hk+1, therefore, we have
Also
That is
This implies that limk→∞pk=q=p∗=ΠΓp1 and hence completes the proof.
If we take A2=0 and Ξ1=Ξ2, then the problem (2.1) reduces to find a point of the following problem:
Hence the following result is an easy consequence of the Theorem 3.1:
Corollary 3.1. Assume that Γ≠∅. Then the sequence (pk)
generated by (3.27), under the control conditions (C1)–(C4), converges strongly to an element p∗=ΠΓp1.
4.
Applications
In this section, we posit different frameworks, as an application of the framework established in Section 3.
4.1. Split feasibility problems
The classical SFP, essentially due to Censor and Elfving [18], aims to find ˆp∈ω:=H∩h−1(G)={ˉq∈H:hˉq∈G}, where H⊂Ξ1 and G⊂Ξ2 are nonempty, closed and convex subsets of Ξ1 and Ξ2, respectively. For the implementation of the Theorem 3.1, we recall the indicator operator of a nonempty, closed and convex subset H of Ξ1 as
It has been established that the subdifferential ∂ΦH satisfies the maximal monotonicity provided that the operator ΦH is proper, convex and lower semicontinuous. Also ∂ΦG=N(μ,H), where N(μ,H) is the normal cone of H at μ. Utilizing this fact, we conclude that the resolvent operator of ∂ΦH is the metric projection operator of Ξ1 onto H. We, therefore, arrive at the following variant of the problem (2.1):
Corollary 4.1. Assume that Γ=ω∩Fix(S)≠∅. Then the sequence (pk)
generated by (4.1), under the control conditions (C1)–(C4), converges strongly to an element p∗=ΠΓp1.
4.2. Split equilibrium problems
The equilibrium problem from [13] aims to compute a point p∗∈H such that
where f:H×H→R is a bifunction satisfying,
(A1)f(p∗,p∗)=0 for all p∗∈H;(A2)f is monotone, i.e., f(p∗,q∗)+f(q∗,p∗)≤0 for all p∗,q∗∈H;(A3) for each p∗,q∗,t∗∈H,lim supx→0f(xt∗+(1−x)p∗,q∗)≤f(p∗,q∗);(A4)for each p∗∈H,q∗↦h(p∗,q∗) is convex and lower semi-continuous.
The set EP(f) denotes the set of all solutions associated with the equilibrium problem (4.2). Recall the following auxiliary results for the equilibrium problem:
Lemma 4.1. [25] Let H be a nonempty closed convex subset of a real Hilbert space Ξ1 and let f:H×H→R be a bifunction satisfying (A1)–(A4). For s>0 and p∗∈Ξ1, there exists t∗∈H such that
Moreover, define an operator Ufs:Ξ1→H by
Lemma 4.2. [30] Let H be a nonempty closed convex subset of Ξ1 and let f:H×H→R be a bifunction satisfying (A1)–(A4). Let Af:Ξ1→2Ξ1 be a multivalued operator defined as:
Recall that the operator Af is a maximal monotone operator with domain D(Af)⊂C and EP(f)=A−1f0. Moreover, Ufs=(Id+sAf)−1 for s>0, i.e., Ufs is the resolvent of Af. We, therefore, arrive at the following variant of the problem (2.1):
Corollary 4.2. Assume that Γ=EP(f1)∩h−1(EP(f2))∩Fix(S)≠∅. Then the sequence (pk)
generated by (4.3), under the control conditions (C1)-(C4), converges strongly to an element p∗=ΠΓp1.
4.3. Split optimization problems
Let g:Ξ1→(−∞,∞] be a PCLS function, then the set of minimizer associated with g is defined as
Recall that the ∂g of PCLS function g is a maximal monotone operator and the corresponding resolvent operator of ∂g is called the proximity operator (see[22]). Hence argmin g=(∂g)−10. We, therefore, arrive at the following variant of the problem (2.1).
Corollary 4.3. Assume that Γ={x∈argming1:hx∈argming2}∩Fix(S)≠∅. Then the sequence (pk)
generated by (4.4), under the control conditions (C1)–(C4), converges strongly to an element p∗=ΠΓp1.
5.
Numerical experiment and results
In this section, we present an example that characterizes the performance of our algorithm.
Example 5.1. Let Ξ1=Ξ2=(R,⟨⋅,⋅⟩,|⋅|) where ⟨p,q⟩=pq. Consider the operators h,A1,A2:R→R are defined as h(p)=3p, A1p=2p and A2p=3p, respectively. It is evident from the definition that A1, A2 are maximal monotone operators such that Ω:={p∗∈A−110:hp∗∈A−120}=0 and h is a bounded linear operator on R with the adjoint operator h∗ such that ‖h‖=‖h∗‖=3. Let the sequence of operators Sk:R→R be defined by
Then Sk is an infinite family of 1−k2(1+k)2-demicontractive operators with ⋂∞k=1Fix(Sk)={0}. Hence Γ=Ω∩⋂∞k=1Fix(Sk)=0. We use the following initialization parameters for the computation of the Algorithm 1: θ=0.5, αk=1100k+1, βk=k100k+1, δ=18, L=3 and m = 0.02. Also observe that
Let Error=Ek=‖pk−pk−1‖<10−5 denote the stopping criteria. The performance of the Algorithm 1 (i.e., Algo.1, θk≠0) is compared with the non-inertial variant of the Algorithm 1 (i.e., Algo.1, θk=0) and Algo. 3.1[16]. For different choices of the initial inputs p0 and p1, the values of Algo.1 are summarized in the following table:
Choice A. Choose x0=(5), x1=(2),
Choice B. Choose x0=(4.2), x1=(1.5),
Choice C. Choose x0=(−7), x1=(4).
The error plotting Ek against Algorithm 1 and Algorithm 3.1[16], for each choices in Table 1, has shown in Figure 1.
It is evident from Figure 1 that Algorithm 1 outperforms its noninertial variant and Algorithm 3.1 [16] with respect to the computation of error, CPU time and the number of iterations.
5.1. Signal processing
The mathematical model of the signal recovery problem in an under-determined linear equation system is expressed as follows:
where ν∈RD denotes the original unknown signal to be recovered, ζ∈RP denotes the observed signal distort via the bounded linear matrix operator h:RD→RP,(P<D) and the noise ρ. One can define a natural convex constrained optimization-theoretic formulation of (5.1) via the following well known LASSO problem [33]:
The set of solutions of the ℓ1-minimization problem (5.1) is equivalent to the set of solutions of (5.2) under certain control conditions on the matrix h [15]. The ℓ1-norm based regularization problems are widely applicable in signal and image processing.
Set Γ=H∩h−1(G)≠∅ with H={ν|‖ν‖1≤c} and G={ζ}. The experiment is conducted under the matrix hD×P whose elements are generated from independently distributed normal distributions having 0 as mean and 1 as variance. The sparse vector ν, having t=spikes nonzero elements, is generated via uniform distribution in [−2,2]. The following iterative regularization method, often known as the Richardson method (or the Landweber method) [23], is generally employed to solve the problem (5.2):
where η, the step size, is assumed to be constant. The algorithm (5.3) converges for 0<η<2ϵ2max, where ϵmax is the maximum singular value of h. The initial points ν0,ν1 are chosen randomly. We use the mean squared error indicator to examine the performance of the algorithm for image restoration, i.e., Ek=1N‖νk−ν∗‖<10−4, where ν∗ is the approximation of the signal ν. The computation of the observed signal ζ is carried out by employing the Gaussian noise associated with the signal-to-noise ratio (SNR = 40). Also set mk=1.85‖h‖2, αk=k100k+1, βk=115k+1, δ=0.04, c=t−0.002, μ=0 and θk = 0.5.
Performance Test 1: Fix D=512, P=256 and spikes=15.
Performance Test 2: Fix D=1024, P=512 and spikes=35.
It is clear from the Figures 2 and 3 that the Algorithm 1 outperforms its variants and Algorithm 3.1 [16] for the signal recovery problem as well as exhibits fast convergence characteristic with regards to the error and number iterations.
6.
Conclusions
In this paper, we have posited a framework for the investigation of the SCNPP and the FPP associated with an infinite family of k-demicontractive operators in Hilbert spaces. The common optimal solution of the problem is then constructed via an inertial hybrid projection algorithm under the suitable set of constraints. We have incorporated an appropriate numerical example for the demonstration of the framework as well as for the applicability of our algorithm. We found that our algorithm outperforms its variants and Algorithm 3.1[16]. We have also discussed various instances of the proposed formalism and can pave a way for an important future research direction in the theories of SCNPP and FPP.
Acknowledgments
The authors Y. Arfat and P. Kumam acknowledge the financial support provided by the Center of Excellence in Theoretical and Computational Science (TaCS-CoE), KMUTT. Moreover, this research project is supported by Thailand Science Research and Innovation (TSRI) Basic Research Fund: Fiscal year 2021 under project number 64A306000005. The author Y. Arfat was supported by the Petchra Pra Jom Klao Ph.D Research Scholarship from King Mongkut's University of Technology Thonburi, Thailand (Grant No.16/2562). Finally, the corresponding author Supak Phiangsungnoen acknowledge the financial support provided by institute of Research and Development, Rajamangala University of Technology Rattanakosin (Fundamental Fund Grant No.FRB6620/25).
Availability of data and material
Data sharing not applicable to this article as no data-sets were generated or analysed during the current study.
Conflict of interest
The authors declare no conflict of interest.