We study a unified form of Tikhonov regularization for ill-posed problems with a general data similarity term. We discuss sufficient conditions on this generalized Tikhonov functional that guarantee existence and stability of solutions. Furthermore, we show that some particular cases of similarity functionals and regularization techniques can be cast into a unified theoretical framework. In particular, we consider the cases of p-power norm, Bregman distance and mutual information as examples of a data similarity term, and Tikhonov regularization of order one or using a Sobolev norm, total variation penalization and powers of semi-norms associated to closed operators as examples of a regularization term. Finally, we take up the case of the image registration problem.
Citation: El Mostafa Kalmoun, Fatimah Allami. On the existence and stability of minimizers for generalized Tikhonov functionals with general similarity data[J]. AIMS Mathematics, 2021, 6(3): 2764-2777. doi: 10.3934/math.2021169
[1] | Zhenping Li, Xiangtuan Xiong, Jun Li, Jiaqi Hou . A quasi-boundary method for solving an inverse diffraction problem. AIMS Mathematics, 2022, 7(6): 11070-11086. doi: 10.3934/math.2022618 |
[2] | M. J. Huntul, Kh. Khompysh, M. K. Shazyndayeva, M. K. Iqbal . An inverse source problem for a pseudoparabolic equation with memory. AIMS Mathematics, 2024, 9(6): 14186-14212. doi: 10.3934/math.2024689 |
[3] | Fethi Bouzeffour . Inversion formulas for space-fractional Bessel heat diffusion through Tikhonov regularization. AIMS Mathematics, 2024, 9(8): 20826-20842. doi: 10.3934/math.20241013 |
[4] | Daniel Gerth, Bernd Hofmann . Oversmoothing regularization with ℓ1-penalty term. AIMS Mathematics, 2019, 4(4): 1223-1247. doi: 10.3934/math.2019.4.1223 |
[5] | Hongwu Zhang, Xiaoju Zhang . Tikhonov-type regularization method for a sideways problem of the time-fractional diffusion equation. AIMS Mathematics, 2021, 6(1): 90-101. doi: 10.3934/math.2021007 |
[6] | Jiajia Zhao, Zuoliang Xu . Calibration of time-dependent volatility for European options under the fractional Vasicek model. AIMS Mathematics, 2022, 7(6): 11053-11069. doi: 10.3934/math.2022617 |
[7] | Phuong Nguyen Duc, Erkan Nane, Omid Nikan, Nguyen Anh Tuan . Approximation of the initial value for damped nonlinear hyperbolic equations with random Gaussian white noise on the measurements. AIMS Mathematics, 2022, 7(7): 12620-12634. doi: 10.3934/math.2022698 |
[8] | Valérie Gauthier-Umaña, Henryk Gzyl, Enrique ter Horst . Decoding as a linear ill-posed problem: The entropy minimization approach. AIMS Mathematics, 2025, 10(2): 4139-4152. doi: 10.3934/math.2025192 |
[9] | Lei Hu . A weighted online regularization for a fully nonparametric model with heteroscedasticity. AIMS Mathematics, 2023, 8(11): 26991-27008. doi: 10.3934/math.20231381 |
[10] | Dun-Gang Li, Fan Yang, Ping Fan, Xiao-Xiao Li, Can-Yun Huang . Landweber iterative regularization method for reconstructing the unknown source of the modified Helmholtz equation. AIMS Mathematics, 2021, 6(9): 10327-10342. doi: 10.3934/math.2021598 |
We study a unified form of Tikhonov regularization for ill-posed problems with a general data similarity term. We discuss sufficient conditions on this generalized Tikhonov functional that guarantee existence and stability of solutions. Furthermore, we show that some particular cases of similarity functionals and regularization techniques can be cast into a unified theoretical framework. In particular, we consider the cases of p-power norm, Bregman distance and mutual information as examples of a data similarity term, and Tikhonov regularization of order one or using a Sobolev norm, total variation penalization and powers of semi-norms associated to closed operators as examples of a regularization term. Finally, we take up the case of the image registration problem.
An inverse problem can be cast into the general form of finding the solution x of a given operator equation
F(x)=y, | (1.1) |
where y∈Y is a given data and F:Ω⊆X→Y is a general operator, which links two function spaces X and Y that are supposed to be normed spaces.
In recent years, inverse problems have become an area of active research, which is mainly due to the extension of their use to many fields of applications in science and engineering such as mathematical physics, signal processing, medical imaging, computer vision, geophysics, etc. For instance, in medical imaging, recent applications include computerized tomography, thermoacoustic imaging, electrical impedance tomography; see [18] and references therein.
The distinctive characteristic of these problems is their ill-posedness in the sense of Hadamard [11], which means the problem has either no solution, the solution may not be unique, or the solution does not depend continuously on the data. When the operator F is not surjective, the inverse problem might have no solution, so we instead seek a least residual norm solution. Furthermore, if the solution is not unique (this occurs when F is not injective) we look for a generalized solution called the best approximate to ensure uniqueness. Nevertheless, the main problem of ill-posedness is the instability of this best approximate solution. This can typically be avoided if a stabilization procedure or a regularization method is used.
One of the most popular stabilization procedure is due to Tikhonov [24]. In his method, Tikhonov approximated the solution of the original inverse problem by minimizing both the residual error and the solution norm to penalize large values in the desired solution. Thus the Tikhonov functional related to the inverse problem (1.1) is as follows
Jα(x):=‖F(x)−y‖2+α‖x‖2,∀x∈Ω, | (1.2) |
where α>0 is a regularization parameter to adjust the weight of the regularization term with respect to the data fitting term. This method is said to be of order zero, while another widely used version replaces the solution norm in the regularization term by the gradient norm. This method is said to be of order one and the energy functional takes the following form
Jα(x):=‖F(x)−y‖2+α‖∇x‖2,∀x∈Ω. | (1.3) |
In his pioneering work, Phillips [19] proposed in the context of integral equations to penalize the large oscillations in the solution by taking the L2 norm of the second derivative of the solution. This is why the method is also known as Tikhonov-Phillips regularization and the energy functional can be cast into the general form
Jα(x):=‖F(x)−y‖2+α‖Lx‖2,∀x∈Ω, | (1.4) |
where L is any differential operator of order greater or equal to zero. In some cases, if more regularity in the solution is needed, we can use the Sobolev norm instead of the L2 norm, which corresponds to a combination of several high order derivatives; see [10]. By the eighties the analysis of Tikhonov regularization for solving linear ill-posed problems was developed [7]. The setting to prove well-posedness (existence, uniqueness and stability) was Hilbert spaces settings for the solution space and the right-hand sides. Recently, the framework has been extended to Banach spaces in connection with Bregman distances; see [1,3,12,16,21].
On the other hand, the Tikhonov regularization was widely used in many imaging applications. For instance, the first order version (1.3) was applied to the optical flow problem, which gave rise to the well-known method of Horn and Schunck [13]. In many such applications the solution and data spaces were the L2 space, and the set of admissible solutions was the Sobolev space H1. In this case, the method was known to suffer from the over-smoothing effect that can lead to the loss of essential features in the solution. This is mainly due to the use of the L2 norm as the regularization term in the penalized functional. To overcome this drawback, Rudin, Osher and Fatemi [22] developed the total variation (TV) regularization technique in the context of image denoising and deblurring. The L2 norm was replaced by the bounded variation semi-norm, which leads to a non-differentiable but still convex regularization term. Despite this theoretical and numerical difficulty in the TV regularization, the method was highly successful in many applications, namely in image processing [4,5]. In [9], the authors have introduced a method that combines total variation and L2 regularization to reconstruct piecewise smooth signals. Hofmann et al [12] have treated the well-posedness of a generalized Tikhonov functional with a power norm similarity term that includes all the previous methods as special cases
JΨ,α(x)=‖F(x)−y‖p+αΨ(x) | (1.5) |
for 1≤p<∞ and Ψ:Ω⊂X→[0,∞] is an arbitrary function acting as the regularization procedure. Some recent imaging applications involving non-metric similarity terms has motivated the study of a unified form of the Tikhonov functional with a general similarity data term [8,20,23]
JΦ,Ψ,α(x)=Φ(F(x),y)+αΨ(x) | (1.6) |
where Φ:Y×Y→[0,∞) is a given similarity data functional measuring the error between F(x) and y.
In this paper, we provide weaker sufficient conditions for the existence and stability of the minimizer of the functional JΦ,Ψ,α. While assuming similar conditions on the regularization functional Ψ, our work differs from [17] in taking up a general similarity measure. The obtained results are applied to three known similarity measure; namely, the power norm, the Bregman distance and the mutual information. These functional metrics can be then combined with famous examples of regularization functionals like Tikhonov regularization of order one or using a Sobolev norm, total variation penalization and powers of semi-norms associated to closed operators. We illustrate as well how our results can be applied to the image registration problem, which is an important example of ill-posed inverse problems in computer vision.
The paper consists of four additional sections. In Section 2 we provide some preliminaries. In Section 3 and Section 4, we investigate the existence and stability of minimizers of the generalized Tikhonov functional (1.6). In Section 5 and Section 6, we present some examples of known similarity and regularization functionals. Finally, we consider the case of the image registration problem in Section 7, and conclude our work in Section 8.
We recall the necessary definitions needed in the sequel of this paper. We keep the notations of the previous section to be the same along this paper, and we take X and Y to be two normed spaces unless otherwise stated. We begin by general boundedness and coercivity concepts related to the general functional Ψ instead of the norm function.
Definition 1. A S⊂Ω is called Ψ-bounded if there exists a constant c≥0 such that Ψ(s)≤c for every s∈S.
Definition 2. We say that Ω is (weakly) Ψ-complete if each Ψ-bounded sequence of Ω admits a (weakly) convergent subsequence in Ω.
Note that the weak Ψ-completeness holds true for a closed set Ω in a reflexive Banach space X when Ψ is a norm defined on Ω being equivalent or stronger than the norm of X.
Definition 3. We say that a real functional f on Ω is Ψ-coercive if f(xn)→∞ for every sequence (xn)⊂Ω with Ψ(xn)→∞. Moreover, a sequence of real functional (fn) on Ω is Ψ-coercive if fn(xn)→∞ for every sequence (xn)⊂Ω with Ψ(xn)→∞.
The Ψ-coercivity of a real functional f means that f(x) grows infinitely when so does Ψ(x). An example of such a functional f is when it satisfies f(x)≥cΨ(x) for c>0. The concept of Ψ-coercivity, which will be used later to show the existence of global minimizers, is closely connected to the Ψ-boundedness of lower level sets.
Proposition 1. If f:Ω→[−∞,∞] is Ψ-coercive, then lower level sets of f are Ψ-bounded.
Proof. Suppose S={x∈Ω:f(x)≤α} is not Ψ-bounded for some real α. Then there is a sequence (xn)⊂S satisfying Ψ(xn)→∞. Since f(xn)≤α, then f(xn)↛∞. Hence f is not Ψ-coercive.
Now, we present a few continuity definitions.
Definition 4. The operator F:Ω→Y is called (weakly) sequentially continuous if for every sequence (xn)∈Ω such that xn(w)→x∈Ω, the sequence F(xn)(w)→F(x).
Definition 5. A functional g:Y→[−∞,∞] is said to be (weakly) lower semicontinuous if for every sequence (yn)∈Y such that yn(w)→y∈Y, we have g(y)≤lim infn→∞g(yn).
Definition 6. A functional f:Ω→[−∞,∞] is said to be (weakly) Ψ-subsequentially lower semicontinuous if each Ψ-bounded sequence (xn)⊂Ω with xn(w)→x∈Ω admits a subsequence (xnj)⊂(xn) such that f(x)≤lim infj→∞f(xnj).
Remark 1. When Ω is (weakly) Ψ-complete, it is clear that every (weakly) lower semicontinuous functional is (weakly) Ψ-subsequentially lower semicontinuous. That the converse is not always true can be seen from the following example. Define
f(x)={xif x<0x+1if x≥0. |
It is easy to check that f is not lower semicontinuous. In taking Ψ(x)=1/x2, we can see that f is continuous on Ψ-bounded set, and hence Ψ-subsequentially lower semicontinuous.
Definition 7. The functional Φ:Y×Y→[0,∞) is called Lipschitz continuous on bounded sets of Y if for every bounded set Z⊂Y, there exists a constant cz≥0 such that
|Φ(y1,y2)−Φ(z1,z2)|≤cz(‖y1−z1‖+‖y2−z2‖) |
for every (y1,y2),(z1,z2)∈Z×Z.
Finally, we recall the definition of uniform convergence of a sequence of functionals.
Definition 8. Let T and (Tn), n=1,2,… be functionals defined on Ω. We said that the sequence (Tn) converges uniformly to T on Ψ-bounded sets if for every c>0 and ε>0, there exists N=N(c,ε) such that for every n≥N and every x∈Ω with Ψ(x)≤c, we have |Tn(x)−T(x)|<ε.
We are now in position to give the main existence theorem of this paper.
Theorem 1. Suppose that
(H1): F is weakly sequentially continuous,
(H2): Φ is weakly lower semicontinuous in the first argument,
(H3): Ψ is proper and weakly Ψ-subsequentially lower semicontinuous,
(H4): Ω is weakly Ψ-complete.
Then the functional JΦ,Ψ,α in (1.6) has a global minimizer.
Proof. Let (xn)⊂Ω be a minimizing sequence of JΦ,Ψ,α; that is,
limn→∞JΦ,Ψ,α(xn)=infx∈ΩJΦ,Ψ,α(x)=Jmin. | (3.1) |
First, note that 0≤Jmin<∞. Also, the functional JΦ,Ψ,α is Ψ-coercive because for every sequence (zn)⊂Ω with limn→∞Ψ(zn)=∞, we have
JΦ,Ψ,α(zn)≥αΨ(zn)→∞. |
We claim that the sequence (xn) is Ψ-bounded. Otherwise, there would exist a subsequence (xnj)⊂(xn) such that Ψ(xnj)→∞, which yields JΦ,Ψ,α(xnj)→∞, owing to the Ψ-coercivity of JΦ,Ψ,α, but this contradicts (3.1). Thus, by (H4) there exist a subsequence of (xn), denoted the same, and ˉx∈Ω such that xnw→ˉx. Hence, from (H3), we see that there is another subsequence of (xn), denoted again the same, such that
Ψ(ˉx)≤lim infn→∞Ψ(xn). |
On the other hand, the combination of (H1) and (H2) tells us that
Φ(F(ˉx),y)≤lim infn→∞Φ(F(xn),y). |
Therefore
JΦ,Ψ,α(ˉx)=Φ(F(ˉx),y)+αΨ(ˉx)≤lim infn→∞Φ(F(xn),y)+αlim infn→∞Ψ(xn)≤lim infn→∞[Φ(F(xn),y)+αΨ(xn)]=lim infn→∞JΦ,Ψ,α(xn)=limn→∞JΦ,Ψ,α(xn)=Jmin, |
which means JΦ,Ψ,α(ˉx)=Jmin, and hence a global minimizer of JΦ,Ψ,α does exist.
Remark 2. When Ω is Ψ-complete, we can relax the continuity conditions on F,Φ and Ψ to hold true for the strong topology in the related spaces. More precisely, we can replace (H1)-(H4) by the following hypotheses, respectively:
(A1): F is sequentially continuous,
(A2): Φ is lower semicontinuous in the first argument,
(A3): Ψ is proper and Ψ-subsequentially lower semicontinuous,
(A4): Ω is Ψ-complete.
Remark 3. For a linear operator F, the global minimizer of JΦ,Ψ,α is unique when Φ(.,y) and Ψ are convex with at least one of them being strictly convex. For a nonlinear operator, uniqueness can not be guaranteed in the general case.
Remark 4. Theorem 1 recovers Theorem 2.5 in [17] as a special case by taking Φ(u,v)=‖u−v‖2 and assuming F to be linear.
The previous theorem is a general case for the existence of a minimizer of the Tikhonov functional (1.2).
Corollary 1. Suppose X is a reflexive Banach space and Ω is closed in X. Then the functional (1.2) has a global minimizer provided that F satisfies (H1). Furthermore, when F is linear, the minimizer is unique provided that F is injective or Ω is a Hilbert space.
Proof. For the existence, it suffices to see the functionals Φ(z,y)=‖z−y‖2 and Ψ=‖.‖2 satisfy assumptions (H2)-(H4) using the reflexivity of X and the weak lower semicontinuity of the norm.
For the uniqueness, when F is injective, the functional Φ(.,y) is strictly convex, and by convexity of Ψ(x)=‖x‖2, we get the strict convexity of Jα. For the case when (Ω,Ψ) is a Hilbert space, we use the strict convexity of the squared norm Ψ(x)=‖x‖2.
In order to study the stability of the generalized Tikhonov functional (1.6), we consider perturbations of the data (F,y) in the inverse problem (1.1) as follows. For n∈N, we let Fn:Ω→Y, yn∈Y and consider the perturbed functionals
J(n)Φ,Ψ,α(x)=Φ(Fn(x),yn)+αnΨ(x) | (4.1) |
where (αn) is a positive real sequence converging to α. In general, the minimization of (1.6) might be instable, meaning that minimizers of (4.1) might move away from those of (1.6) even if (Fn,yn) is enough close to (F,y). Let's first give a weak stability result for functionals minimization.
Proposition 2. Suppose Ψ satisfies Hypothesis (H4) of Theorem 1. Assume also that J is a weakly Ψ-subsequentially lower semicontinuous real functional on Ω, and (Jn) is a sequence of Ψ-coercive real functionals on Ω which converges uniformly to J on Ψ-bounded sets. Then each minimizer sequence of (Jn) has a weak cluster point that minimizes J.
Proof. Let (xn) be a sequence of global minimizers of Jn, and x be an arbitrary element of Ω. Since Jn(xn)≤Jn(x), it follows that
lim supn→∞Jn(xn)≤lim supn→∞Jn(ˉx)=J(ˉx)<∞. |
On the basis of the Ψ-coercivity of Jn, it's clear that Ψ(xn) is bounded above; hence it is bounded, which means the sequence (xn) is Ψ-bounded. In combining this with the hypothesis (H4), we see that there exist x∗∈Ω and a subsequence of (xn), denoted the same way, such that xnw→x∗. Since (xn) is Ψ-bounded and J is weakly Ψ-subsequentially lower semicontinuous, there is a subsequence, denoted again (xn), such that J(x∗)≤lim infn→∞J(xn). Recalling that (Jn) converges uniformly to J on Ψ-bounded sets, we get J(xn)−Jn(xn)→0. Consequently,
J(x∗)≤lim infn→∞J(xn)≤lim supn→∞J(xn)≤lim supn→∞(J(xn)−Jn(xn))+lim supn→∞Jn(xn)=lim supn→∞Jn(xn)≤J(ˉx), |
which holds for every x∈Ω. We conclude that x∗ is a global minimizer of J.
As a particular case, we get a weak stability result for the minimizer of the generalized Tikhonov functional (1.6).
Theorem 2. Suppose (Fn) satisfies (H1) such that (Fn) converges uniformly to F on Ψ-bounded sets, and let yn→y. In addition to the hypotheses of Theorem 1, assume also that Φ is Lipschitz continuous on Ψ-bounded sets. Then argminJ(n)Φ,Ψ,α≠∅, argminJΦ,Ψ,α≠∅ and each sequence in the former set has a weak cluster point in the latter one.
Proof. On the basis of Theorem 1, it's clear to see that argminJ(n)Φ,Ψ,α≠∅ and argminJΦ,Ψ,α≠∅. It is enough to check that the functionals JΦ,Ψ,α and J(n)Φ,Ψ,α satisfy the hypotheses of Proposition 2.
First, in applying similar arguments as in the proof of Theorem 1, we can easily show that JΦ,Ψ,α is weakly Ψ-subsequentially lower semicontinuous. For the Ψ-coercivity of (J(n)Φ,Ψ,α), we have for any sequence (xn) in Ω satisfying Ψ(xn)→∞,
J(n)Φ,Ψ,α(xn)≥αnΨ(xn)→∞. |
Finally, to show the uniform convergence of (J(n)Φ,Ψ,α) to JΦ,Ψ,α on Ψ-bounded sets, we let M be a Ψ-bounded set in Ω. Owing to the Lipschitz continuity of Φ on Ψ-bounded sets, we see that for every x∈M
|J(n)Φ,Ψ,α(x)−JΦ,Ψ,α(x)|=|Φ(Fn(x),yn)+αnΨ(x)−Φ(F(x),y)−αΨ(x)|≤|Φ(Fn(x),yn)−Φ(F(x),y)|+|(αn−α)||Ψ(x)|≤c(‖Fn(x)−F(x)‖+‖yn−y‖)+|(αn−α)||Ψ(x)| |
where c>0 is a Lipschitz constant. This yields (J(n)Φ,Ψ,α) converges uniformly to JΦ,Ψ,α on M because so does (Fn) to F, and taking into account yn→y and αn→α.
Remark 5. We observe that the weak stability result of Theorem 2 is satisfied if
1. The space X is a reflexive Banach space and Ω is closed in X.
2. F is weakly sequentially continuous.
3. The functionals Φ and Ψ are norms.
Taking into account Remark 2, the following strong stability result may be proved in much the same way as Theorem 2.
Theorem 3. Suppose (Fn) satisfies (A1) such that (Fn) converges uniformly to F on Ψ-bounded sets, and let yn→y. In addition to the hypotheses of Remark 2, assume also that Φ is Lipschitz continuous on Ψ-bounded sets. Then argminJ(n)Φ,Ψ,α≠∅, argminJΦ,Ψ,α≠∅ and each sequence in the former set has a cluster point in the latter one.
In this section, we present two examples of similarity measures that correspond to particular choices of the functional Φ in (1.6).
The most known example of a similarity measure is the power norm function
Φ(z,y)=‖z−y‖p,1≤p<∞. | (5.1) |
Letting p=2, we get the quadratic data term in the standard Tikhonov functional (1.2). The functional Φ satisfies Hypothesis (H2) of Theorem 1 by the weak lower semicontinuity of the norm.
For the weak stability result in Theorem 2, we need to show that Φ is Lipshcitz continuous on bounded sets assuming Ψ to be a norm on X. Let Z⊂Y be a bounded set. First, we have ||z||≤||y||+||z−y||. Using the p-duality map, we can show that
||z||p≤||y||p+p||z−y||||z||p−1. |
Then, by exchanging the role of z and y, we get
|||z||p−||y||p|≤pCp||z−y|| |
where C is the boundedness constant of Z. Hence,
|Φ(z1,y1)−Φ(z2,y2)|≤pCp||z1−y1−z2+y2||≤pCp(|‖z1−z2‖+‖y1−y2‖) |
for all z1,y1,z2,y2 in Z.
Bregman distance [2] was introduced in 1967 in the context of convex programming. Here we recall a generalized definition of this pseudo-distance related to convex functionals on a Banach space. Suppose Y is a Banach space and let f:Y→R∪{+∞} be a proper, convex and lower semicontinuous function. Let also v∗ be a subgradient of f at v∈Y; that is an element of the subdifferential
∂f(v)={w∈Y∗|f(v)+<w,u−v>≤f(u)∀u∈Y}. |
Since the function J is convex, finite and continuous at one point of U, then ΩB(f)={v:∂f(v)≠∅} the Bregman domain of f satisfies ΩB(f)=Y; see [6] (Proposition 5.2 in Chapter I). The Bregman distance is then defined for u,v∈Y by
Bv∗f(u,v):=f(u)−f(v)−<v∗,u−v>. |
The symmetric version of the Bregman distance is given by
Bsf(u,v):=Bu∗f(v,u)+Bv∗f(u,v)=<u∗−v∗,u−v>. |
We take Φ(u,v)=Bv∗f(u,v). By the definition of subgradients, the nonnegativity of Φ is clear. Moreover, the weak lower semicontinuity of Φ in the first argument results from the convexity and continuity of f.
For stability with respect to perturbation of right hand data y only, we need to assume a Lipschitz condition on f. Owing to the local boundedness of the subdifferential sets of f, the similarity functional Φ will be Lipschitz with respect to the first argument and this will be sufficient for the proof of Theorem 2. In case of perturbations in both the operator F and the data y, it is useful to use the symmetric version of the Bregman distance to get the joint Lipschitz condition of Φ.
In this section we give different known regularization terms that meet the conditions we have imposed on the functional Ψ in (1.6).
Instead of penalizing large components in the solution by choosing the penalizer Ψ(x)=‖x‖2 in the ordinary Tikhonov functional (1.2), we can penalize large deviations in the solution by penalizing the squared norm of the gradient.
Let U⊂Rd be open and bounded with a smooth boundary, X=L2(U) and Ω=H1(U). We take Ψ(x)=‖∇x‖2L2(U) in the functional (1.6). As the the L2 norm is weakly lower semicontinuous, the functional Ψ satisfies Hypothesis (H3). Note also that Ψ satisfies also Hypothesis (H4) in view of the Poincaré inequality and the reflexivity of the Sobolev space Ω.
Under the hypotheses (H1) and (H2), if the operator F is linear and the similarity measure Φ is convex in the first argument, then the minimizer is unique because Ψ is already strictly convex as the composition of a strictly convex function and a linear operator.
In case of seeking further regularity in the solution, we can apply a higher-order regularization by making use of the Sobolev norm. In particular, another different form of Tikhonov regularization arises by choosing the penalizer as Ψ(x)=‖x‖Wk,p where Wk,p is a Sobolev space [15].
Let U⊂Rd be open and bounded with a smooth boundary, X=Lp(U) and Ω=Wk,p(U) with 1<p<∞. To prove that Ψ satisfies Hypotheses (H3) and (H4), let us consider (xn)⊂Wk,p(U) to be a Ψ-bounded sequence such that xnw→x in Wk,p(U) and let α∈Zn be a multi-index such that |α|≤k. It follows that Dαxnw→Dαx in Lp(U). By the weak lower semicontinuity of the Lp norm, we have
‖Dαx‖Lp(U)≤lim infn→∞‖Dαxn‖Lp(U). |
Thus
Ψ(x)=‖x‖Wk,p(U)=(∑|α|≤k‖Dαx‖pLp(U))1/p≤lim infn→∞(∑|α|≤k‖Dαxn‖pLp(U))1/p=lim infn→∞‖xn‖Wk,p(U)=lim infn→∞Ψ(xn). |
This shows (H3). Moreover, since Wk,p(U) is compactly embedded in Lp(U), then there exist a subsequence (xnj)⊂(xn) and x∈Wk,p(U) such that xnjw−Lp→x. Hence (H4) holds true.
Once again, the uniqueness is guaranteed in case of the convexity of Φ(.,y) and linearity of F provided that Hypotheses (H1) and (H2) are satisfied because Ψ is strictly convex.
The second popular regularization technique after Tikhonov method is total variation penalization. The method was first introduced by Rudin et al [22] in the context of image denoising, and since then it achieved a highly success in many imaging applications [4,14]. We will follow the same arguments employed in [17] to show that this penalization term satisfies the conditions of Theorem 1 and Theorem 2.
Let U⊂Rd be convex and bounded with Lipschitz continuous boundary such that d≥2 and X=Lp(U) with 1≤p≤dd−1. We consider Ω to be the space of functions of bounded variation (BV) on U:
Ω=BV(U)={x∈L1(U):φ0(x)<∞} |
with
φ0(x)=supy∈V∫U(−x div y)ds |
and
V={y∈C10(U;Rd):|y(s)|≤1,∀s∈U}. |
Note that if x∈C1(U), then φ0(x)=∫U|∇x|ds. The BV norm is then defined for x∈Ω by ‖x‖BV(U)=‖x‖L1(Ω)+φ0(x), and we take Ψ(x)=‖x‖BV(U). Let (xn)⊂Ω be a Ψ-bounded sequence satisfying xnw−Lp→x∈Ω; hence unw−L1→u (since p≥1). Since the norm ‖.‖L1(U) is weakly lower semicontinuous, we get
‖x‖L1(Ω)≤lim infn→∞‖xn‖L1(Ω). | (6.1) |
On the other hand, from xnw−Lp→x, we can see that
∫U(−x div y)ds=limn→∞∫U(−xn div y)ds≤lim infn→∞∫U(−xn div y)ds. |
By taking supremum on both sides, we obtain
φ0(x)≤lim infn→∞φ0(xn). |
The combination of the last inequality with (6.1) tells us that
Ψ(x)=‖x‖BV(U)=‖x‖L1(U)+φ0(x)≤lim infn→∞‖xn‖L1(U)+lim infn→∞φ0(xn)≤lim infn→∞(‖xn‖L1(U)+φ0(xn))=lim infn→∞‖xn‖BV(U)=lim infn→∞Ψ(xn). |
This proves (H3). Furthermore, let (xn)⊂Ω be a Ψ-bounded sequence. As BV(U) is compactly embedded in Lp(U) for 1≤p≤dd−1, there is a subsequence (xnj) of (xn) such that xnjw−Lp→x, which confirms (H4).
If the operator F is linear and Φ(,.y) is strictly convex, then the minimizer is unique; otherwise we cannot guarantee uniqueness because the ‖.‖BV norm is not strictly convex.
We consider here a very general form of Tikhonov regularization [17], which corresponds to the choice of Ψ(x)=‖Lx‖q in the generalized Tikhonov functional (1.6) where L:Ω⊂X→Y is a closed linear operator with a weakly closed range R(L). Note that a closed and bounded below linear operator on a Banach space has a closed range. Here we suppose that X is a Hilbert space and Y is a reflexive Banach space. Let us denote the null space of L by N(L), the orthogonal space of a subspace M in X by M⊥, the orthogonal projection of X onto M by PM, and finally the Moore-Penrose generalized inverse of L by L†.
Let (xn)⊂Ω be a Ψ-bounded sequence such that xnw→x∈Ω. It is easily seen that the sequence (Lxn) is bounded in the weakly closed set R(L). Therefore it has a weakly convergent subsequence (Lxnj) to an element z of R(L) because Y is reflexive. Now, by the weak continuity of L†, we have PN(L)⊥(xnj)=L†(Lxnj)w→L†z. Thus PN(L)(xnj)=xnj−PN(L)⊥xnjw→x−L†z. Since N(L) is closed, we get x−L†z∈N(L); and therefore
0=L(x−L†z)=Lx−LL†z=Lx−P¯R(L)(z)=Lx−z. |
Hence Lx=z. It follows that
Ψ(x)=‖Lx‖q=‖z‖q≤lim infj→∞‖Lxnj‖q=lim infj→∞Ψ(xnj). |
Therefore Ψ satisfies (H3). Furthermore, since Ψ-bounded sequences in Ω are bounded in X which is reflexive, then Hypothesis (H4) holds true as well.
For a linear operator F and a convex functional Φ, the minimum is unique by use of the strict convexity of Ψ as the q-th power (q>1) of a strictly convex norm.
Remark 6. The hypotheses on L hold for most differential operators on Sobolev spaces.
In this section we take up the image registration problem, which is subject to intense research efforts due to its important applications in remote sensing, computer vision, medical imaging, etc. Given a reference image R:U→[0,1] and a template image T:U→R of the same object, defined on a bounded domain U∈Rd, the problem consists in finding a spatial displacement field x:U→U such that T(s+x(s))=R(s) for all s in U. This is a well-known example of ill-posed inverse problems, which can be cast into the form of a functional equation as
T(x)=R. |
Under a variational framework, this problem is solved by minimizing the generalized Tikhonov functional (1.6) for F=T, y=R, Ω and Y are given spaces of displacement fields and images, respectively.
The choice of the similarity measure Φ can be based either on the alignment of the geometrical structures or the gray levels of the two images. An extraordinary successful similarity metric in the context of multi-modal registration is the mutual information proposed by Viola and Wells [25]. In assuming ξx1:=T(x) and ξ2:=R are random variables in Y with respective probability densities pT and pR, the mutual information measures the Kullback-Leibler distance between the joint distribution pT,R(ξx1,ξ2) and the product distribution pT(ξx1)pR(ξ2)
D(T(x),R)=KL(pT,R(ξx1,ξ2),pT(ξx1)pR(ξ2)) |
where KL(p1,p2)=∫p1(s)logp1(s)p2(s)ds for any densities p1 and p2. This functional is known to be well-defined and positive. Also, if
Y={y∈L∞(U):‖y‖∞≤A} |
for a large positive constant A, then D(T(x),R)≤C where C is a constant that depends on A. It can be shown (see [23]) that Φ(u,v))=−D(u,v)+C is Lipschitz continuous on Y×Y in any Lq norm for q≥1, which means that Assumption (A2) is satisfied as well. On the other hand, the operator F=T is sequentially continuous when the space of displacement fields Ω is continuously embedded into Lq(U,Rd). A frequent case of such a space would be a Banach space that is compactly embedded into Lq(U,Rd). In combining the mutual information metric with any choice of a regularization functional from the previous section, we get the existence and stability of the image registration problem.
We have considered a unified form of Tikhonov regularization for ill-posed problems which involve a general similarity data term. Under weaker sufficient conditions, we have proven the existence and stability of global minima of the generalized Tikhonov functional. In some special cases, we have precised the additional conditions to ensure the uniqueness of the global minimum. We have also illustrated how known examples of similarity measures and regularization terms meet the existence and stability conditions we have imposed on the generalized Tikhonov functional. Finally, we have briefly demonstrated that our results can be applied to prove the existence and stability of solutions to the image registration problem.
The authors would like to thank two anonymous reviewers for their valuable suggestions and corrections.
[1] | S. W. Anzengruber, The Discrepancy Principle for Tikhonov Regularization in Banach spaces: Regularization pPoperties and Rates of Convergence. Südwestdeutscher Verlag für Hochschulschriften, 2012. |
[2] | L. M. Bregman, The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming, USSR Comput. Math. Math. Phys., 7 (1976), 200–217. |
[3] |
M. Burger, S. Osher, Convergence rates of convex variational regularization, Inverse Probl., 20 (2004), 1411. doi: 10.1088/0266-5611/20/5/005
![]() |
[4] | V. Caselles, A. Chambolle, M. Novaga, Total variation in imaging, Handb. Math. Methods Imaging, 2015 (2015), 1455–1499. |
[5] | T. Chan, S. Esedoglu, F. Park, A. Yip, Recent developments in total variation image restoration, Math. Models Comput. Vision, 17 (2005). |
[6] | I. Ekeland, R. Temam, Convex analysis and variational problems, 28 (1999). |
[7] | H. W. Engl, M. Hanke, A. Neubauer, Regularization of Inverse Problems, Springer Science Business Media, 375 (1996). |
[8] | J. Flemming, Theory and examples of variational regularization with non-metric fitting functionals, J. Inverse Ill-Posed Probl., 18 (2010), 677–699, |
[9] |
A. Gholami, S. M. Hosseini, A balanced combination of Tikhonov and total variation regularizations for reconstruction of piecewise-smooth signals, Signal Processing, 93 (2013), 1945–1960. doi: 10.1016/j.sigpro.2012.12.008
![]() |
[10] | C. Groetsch, The theory of Tikhonov regularization for Fredholm equations. 104p, Boston Pitman Publication, 1984. |
[11] | J. Hadamard, Lectures on Cauchy's problem in linear partial differential equations, yale univ, Press. New Haven, 1923. |
[12] |
B. Hofmann, B. Kaltenbacher, C. Poeschl, O. Scherzer, A convergence rates result for Tikhonov regularization in Banach spaces with non-smooth operators, Inverse Problems, 23 (2007), 987. doi: 10.1088/0266-5611/23/3/009
![]() |
[13] | B. K. Horn, B. G. Schunck, Determining optical flow, Artif. Intel., 17 (1981), 185–203. |
[14] |
E. M. Kalmoun, An investigation of smooth tv-like regularization in the context of the optical flow problem, J. Imaging, 4 (2018), 31. doi: 10.3390/jimaging4020031
![]() |
[15] | P. Kazemi, R. J. Renka, Tikhonov regularization using Sobolev metrics, Electron. J. Differ. Eq., 2014. |
[16] | S. Kindermann, Convex Tikhonov regularization in Banach spaces: New results on convergence rates, J. Inverse Ill-posed Probl., 24 (2016), 341–350. |
[17] |
G. Mazzieri, R. Spies, K. Temperini, Existence, uniqueness and stability of minimizers of generalized Tikhonov–Phillips functionals, J. Math. Anal. Appl., 396 (2012), 396–411. doi: 10.1016/j.jmaa.2012.06.039
![]() |
[18] | F. D. M. Neto, A. J. da Silva Neto, An Introduction to Inverse Problems with Applications, Springer Science & Business Media, 2012. |
[19] |
D. L. Phillips, A technique for the numerical solution of certain integral equations of the first kind, J. ACM (JACM), 9 (1962), 84–97. doi: 10.1145/321105.321114
![]() |
[20] | C. Pöschl, Tikhonov Regularization with General Residual Term, PhD thesis, University of Innsbruck, Austria, 2008. |
[21] |
E. Resmerita, Regularization of ill-posed problems in Banach spaces: Convergence rates, Inverse Probl., 21 (2005), 1303. doi: 10.1088/0266-5611/21/4/007
![]() |
[22] |
L. I. Rudin, S. Osher, E. Fatemi, Nonlinear total variation based noise removal algorithms, Physica D: Nonlinear Phenom., 60 (1992), 259–268. doi: 10.1016/0167-2789(92)90242-F
![]() |
[23] | S. Salzo, Variational Regularization for Image Registration: Theory and Algorithms, PhD thesis, Dipartimento di Informatica e Scienze dell'Informazione, 2012. |
[24] | A. N. Tikhonov, Solution of incorrectly formulated problems and the regularization method, Soviet Math., 4 (1963), 1035–1038. |
[25] |
P. Viola, W. M. Wells III, Alignment by maximization of mutual information, Int. J. Comput. Vision, 24 (1997), 137–154. doi: 10.1023/A:1007958904918
![]() |