1.
Introduction
In many imaging applications, images inevitably contain noise. Most of the literatures deal with the reconstruction of images corrupted by additive Gaussian noise, for instance [1,2,3,4,5,6]. However, in many engineering applications the noise has an impulsive characteristic, which is different from Gaussian noise and cannot be modeled by Gaussian noise. Based on [7], a type of impulsive degradation is given by Cauchy noise, which follows Cauchy distribution and appears frequently in radar and sonar applications, atmospheric and underwater acoustic images, wireless communication systems. For more details, we refer to [8,9]. Recently, much attention has been paid to dealing with Cauchy noise and several approaches have been proposed. In [10], Chang et al. employed recursive Markov random field models to reconstruct images corrupted by Cauchy noise. Based on non-Gaussian distributions, Loza et al. [11] proposed a statistical approach in the wavelet domain. By combining statistical methods with denoising techniques, Wan et al. [12] developed a segmentation approach for RGB images corrupted by Cauchy noise. Sciacchitano et al. [13] proposed a total variation (TV)-based variational method for reconstructing images corrupted by Cauchy noise. The variational model in [13] (called as SDZ model) is
where Ω is a bounded connected domain in R2, BV(Ω) is the space of functions of bounded variation, u∈ BV(Ω) (for more details, see (2.1)) represents the restored image and γ>0 is the scale parameter of Cauchy distribution. In (1.1), λ is a positive number, which controls the trade-off between the TV regularization term and the fidelity term, u0 is the image obtained by applying the median filter [14] to the noisy image f, and η>0 is a penalty parameter. If 8ηγ2≥1, the objective functional in (1.1) is strictly convex and its solution is unique. The term η||u−u0||22 in (1.1) results in the solution being close to the median filter result, but the median filter does not always perform well as to Cauchy noise removal. In order to avoid this, in [15], the authors developed the the alternating direction method of multipliers (ADMM) to solve the following non-convex variational model (called as MDH model) directly
where K represents a linear operator. As we know, solutions of variational problems with TV regularization have many desirable properties, such as the feature of preserving sharp edges. However, these solutions are always accompanied by blocking artifacts due to the property of BV space.
In order to overcome blocking artifacts, we will employ TGV as a regularization term. In [16], Bredies et al. proposed the concept of TGV, and they applied TGV to mathematical imaging problems to overcome blocking artifacts. For more details of TGV, we refer interested readers to [17,18]. In order to overcome the defect of the median filter result, based on the proximal algorithm idea, we will use the term ||u−z||22 to convexify the non-convex fidelity term ∫Ωlog(γ2+(u−f)2)dx. To simplify computing, for the TGV regularization term, we employ the proximal method. Based on these, we propose the following model
with the constraint u=z. Meanwhile, we compare the proposed model (1.3) with the following model
where u0 is the image obtained by applying the median filter [14] to the noisy image. According to Table 1, the numerical results show that the proposed model (1.3) is better than the model (1.4). Compared with previous reports, the main novelty of our proposed approach has been condensed into the following points:
1. Compared with the BV regularization term, we employ the TGV regularization term which preserves the image structure and we prove that the proposed model admits a unique solution.
2. Different from the constraint by applying the median filter, we use the constraint by applying the proximal approach and experiment results show better performance.
3. The previous literature used ADMM algorithm but we employ non-expansive operator and the fixed point algorithm such that the convergence of the proposed algorithm is more efficiently proved.
The next part is organized as below. We propose a new model and show the model has a unique solution in Section 2. In Section 3, we employ a minimization scheme to deal with the new model. We show the convergence of the proposed algorithm in Section 4. The performance of the new method is demonstrated by numerical results in Section 5. Some remarks are concluded in Section 6.
2.
Variational model
Similar to [13], we propose a new non-convex TGV model for denoising Cauchy noise. For completeness, firstly, a review of BV space and TGV space is given. For more details on TGV models and Cauchy noise removal, we refer to [19,20,21,22].
2.1. Preliminaries
For convenience, we introduce the following notations. The function u∈ BV(Ω) iff u∈L1(Ω) and its TV is finite, where TV of u is
The space BV(Ω) is a Banach space with the norm ||u|| BV(Ω)=||u||L1(Ω)+∫Ω|Du| [23,24].
Throughout the paper, we denote the dimension by d, which is typically 2 or 3. For convenience, Ckc(Ω,Symk(Rd)) expresses the space of compactly supported symmetric tensor field, where Symk(Rd) represents the symmetric tensors space on Rd, which can be written as [16]
The TGV of order k with positive weights α=(α0,α1,⋅⋅⋅,αk−1) is defined as [16]
When k=1,α=1, Sym1(Rd)=Rd, TGV1α(u)=TV(u). When k=2, Sym2(Rd) represents all symmetric Sd×d matrices as follows, for ξ∈Sym2(Rd),
for more details, we refer to [16]. In the following part, we mainly use second-order TGV as
where
and
Following the notation in [25], we define the discretized grid as
for some positive N1,N2∈N, where h denotes the grid width and we take h=1 for convenience. For convenience, we define U,W,Z as
For simplicity, the TGV2α functional will be discretized by finite differences with step-size 1. Based on [16], TGV2α(u) can be reformulated as
where w=(w1,w2)T∈W, ε(w)=12(∇w+∇Tw). The operators ∇ and ε, respectly, denote
∇: U→W, ∇u=(∂+xu∂+yu),
ε: W→Z, ε(w)=(∂+xw112(∂+yw1+∂+xw2)12(∂+yw1+∂+xw2)∂+yw2),
div:W→U, divw=∂−xw1+∂−yw2,
divh: Z→W, divhz=(∂−xz11+∂−yz12∂−xz21+∂−yz22).
For more details on the above discretion, we refer to [26].
By [27], the Cauchy distribution can be written as
where x represents a random variable which obeys the Cauchy distribution, μ represents the peak location, γ>0 is a scale parameter. The scale parameter is similar to the role of the variance. Here, we denote Cauchy distribution by C(μ,γ).
2.2. Proposed model
Following [13], we denote random variables by f,u,v and the respective instances by f,u,v. Denote the noisy image by f=u+v, where v follows the Cauchy noise. We assume that the v follows the Cauchy distribution with μ=0 and its density function is defined as follows
The MAP estimator of u is obtained by maximizing the conditional probability of u with given f. Based on Bayes' rule, we have
Equation (2.8) is equivalent to
where the term logP(f(x)|u(x)) presents the degradation process between f and u, and logP(u) denotes the prior information on u. For the Cauchy distribution C(0,γ) and each x∈Ω, we have
In order to overcome blocking artifacts, we employ the prior P(u)=exp(−2λTGV2α(u)). Then we obtain the TGV model for denoising as
where λ>0 is the regularization parameter.
Next, we show that problem (2.11) admits at least one solution.
Theorem 2.1. The problem (2.11) has at least one solution in BGV2α(Ω), if γ≥1,λ>0.
Proof. Clearly, if γ≥1, there exists a lower bound of the model (2.11). Assume that {uk}k∈N is a minimizing sequence for problem (2.11).
By contradiction, we show that {uk} is bounded in L2(Ω) and therefore bounded in L1(Ω). Assume that ||uk||2=+∞, so there exists a set E⊂Ω and measure(E)≠0, such that for any x∈E, uk(x)=+∞. With f∈L2(Ω), we have log(γ2+(uk−f)2)=+∞ for all x∈E, which contradicts to ∫Ωlog(γ2+(u−f)2)dx<+∞.
Noting that ||∇u||1 and ||ε(w)||1 are both bounded, we obtain that {uk} is a bounded sequence in BGV2α(Ω). According to Rellich-Kondrachov compactness theorem, there exists a function u∗∈L1(Ω) such that uk→u∗. Because TGV2α(u) is proper, semi-continuous and convex in BGV2α(Ω) [16], we obtain that liminfk→+∞TGV2α(uk)≥TGV2α(u∗). Meanwhile, according to the Fatou lemma, we can deduce that
which means that u∗ is the minimum point of (2.11), i.e., the problem (2.11) has at least one solution in BGV2α(Ω). Noting that the model (2.11) is strictly convex, based on the standard arguments in convex analysis[28,29], we obtain that the minimum u∗ is unique.
3.
The algorithm for solving (2.11)
In order to obtain a convergent algorithm, we employ the alternating minimization algorithm for the variational model (2.11). The model (2.11) can be discretized as
Remark 3.1. The proximal operator [30] proxf:Rn→Rn of f is defined as
The definition indicates that proxf(v) is the trade-off between minimizing f and being near to v. Based on this idea, we convexify the model (2.11) by adding a proximal term. The advantages are as follows:
(1) A strictly convex model is obtained due to the proximal term.
(2) The result of each iteration is near to the previous one.
In order to simplify the alternating minimization algorithm, we first introduce the following notations and definitions:
Now, we solve z-subproblem by (3.2). Based on [31], ∫Ω|Du| can be represented by
Equation (3.4) will make the calculation of the primal dual method very easy. Note that TGV2α(u)=infw∈W{α0||∇u−w||1+α1||ε(w)||1} (for more details, refer to [2,32]), where α0,α1 are positive constant parameters. Therefore the min-max problem of (3.2) can be reformulated as
where p,q are the dual variables associated with the sets given by
Similar to [25], the solution of the min-max problem (3.5) can be solved as follows
where the projection can be computed as
σ,τ are positive parameters such that στ≤1/12, and k,l represent iteration numbers.
The optimality condition for (3.3) is
Based on the proximal-operator idea, we can take v=uk such that the result of each iteration is near to the previous one. Multiplying both sides of (3.11) by γ2+(u−f)2, one can obtain that (3.11) is equivalent to
where
In order to solve (3.12), we need the following proposition.
Proposition 3.2. [33] A generic cubic equation with real coefficients
has at least one solution among the real numbers. Let
If there exists a unique real solution of (3.17), the discriminant, △=q3+r2 has to be positive. Furthermore, if △≥0, the only real root of (3.17) is given by
Since the problem (3.3) is strictly convex with respect to u, then there exists a unique real solution for (3.3) and it can be obtained by (3.19). Instead of the method presented above, the u subproblem (3.3) can be solved by the Newton method because the objective function in (3.3) is twice continuously differentiable.
The alternating minimization algorithm for Cauchy noise removal is given in Algorithm 1, where K represents the maximum iteration number.
4.
Convergence
In the following section, we prove the convergence of the proposed Algorithm 1.
Definition 4.1. ([34]). An operator Q:RN→RN is non-expansive, if for ∀y1,y2∈RN, there holds ||Q(y1)−Q(y2)||2≤||y1−y2||2.
Clearly, the identity map I(x)=x for all x is non-expansive. One can easily check that the product and the sum of two non-expansive operators are also non-expansive respectively. For any fixed v∈RN, the maps Q(y)=y+v and Q(y)=y−v are non-expansive.
Definition 4.2. ([34]). Given a non-expansive operator P, T=(1−β)I+βP, for some β∈(0,1), is said to be β-averaged non-expansive.
Definition 4.3. ([34]). An operator G:RN→RN is called firmly non-expansive, if for any x1,x2∈RN, there holds
Remark 4.4. An operator G is firmly non-expansive if and only if it is 12-averaged non-expansive.
Lemma 4.5. ([35]). Let φ be convex and lower semi-continuous, and β>0. Suppose ˆx is defined as follows
Then S is 12-averaged non-expansive.
Since TGV2α(u) is convex and lower semi-continuous, based on Lemma 4.5, it is obvious that S(u) is 12-averaged non-expansive. Note that
Let φ(u)=∑Ni=1log(γ2+(ui−f)2)+12||u−zk||22, we have
Noting that φ(u) is convex and by Lemma 4.5, we have that L(z) is 12-averaged non-expansive.
Lemma 4.6. ([36]) Let P1 and P2 be β1-averaged and β2-averaged non-expansive operators respectively.
By Lemma 4.5, we obtain that L∘S is 34-averaged non-expansive.
Definition 4.7. ([37]). A function ϕ:Rn→R is proper over a set X⊂Rn if ϕ(x)<+∞ for at least one x∈X and ϕ(x)>−∞ for all x∈X.
Definition 4.8. ([37]). A function ϕ:Rn→R is coercive over a set X⊂Rn if for every sequence {xk}∈X such that ||xk||→∞, we have limk→∞ϕ(xk)=∞.
The following Lemma 4.9 can be shown easily, and we omit its proof here.
Lemma 4.9. The functional E(z,u) in (3.1) is coercive.
Lemma 4.10. ([28]). Let ϕ:RN→R be a closed, proper and coercive function. Then the set of the minimizers of ϕ over RN is nonempty and compact.
Lemma 4.11. The set of the fixed points of L∘S is non-empty.
Proof. By Lemma 4.9, the objective function E(z,u) is coercive. Based on Lemma 4.10, the set of minimizers of E(z,u) is non-empty. Set (ˆz,ˆu) is a minimizer of E(z,u). Therefore we have
It indicates that
Thus we have ˆu=L∘S(ˆu).
According to the Krasnoselskii-Mann (KM) theorem [38], noting that L∘S is non-expansive and the set of the fixed points of L∘S is nonempty, one has that the sequence {ui} converges weakly to a fixed point of L∘S, for any initial point u0. Since E(z,u) is strictly convex and differentiable with u, one has that the minimizer of E(z,u) is unique. Clearly, the fixed points of L∘S are just the minimizers of E(z,u). Thus the sequence {uk} converges to the unique minimizer of E(z,u). Therefore, we have the following theorem.
Theorem 4.12. The sequence {uk} converges to the unique minimizer of E(z,u) as k→∞, for any initial point u0.
5.
Numerical simulations
In this section we provide numerical results to show the performance of the proposed method for image restoration problems under Cauchy noise. Here, we compare our method with existing models as follows: ROF model[1], the median filter[39], SDZ model[13] and MDH model[15]. For ROF model, we use the primal dual method proposed in [31]. For SDZ model and MDH model, we use the source codes of [13] and [15] respectively.
Considering the quality of the restoration results, we measure them by different evaluation metrics: The peak-signal-to-noise ratio (PSNR) value and the structural similarity (SSIM) value, which are defined as
where u0 denotes the original signal with mean ˉu0 and u is the denoised signal, M×N is the image size,
where μu,μu0,σ2u,σ2u0,σuu0 denote, respectively, mean, variance, co-variance of the image u and u0, c1 and c2 are small positive constants. To compare with different approaches easily, we use the same stopping criterion for all the algorithms, that is
or the maximum number of iterations.
Combined with relevant reports[13,15,16,17,25,26], we adjust each parameter one by one. For each image in Figure 1, we try our best to tune the parameters of the compared algorithms to obtain the highest PSNR and SSIM. Based on hundreds of experiments, we observe that τ is the key parameter to control the restoration quality and convergence speed. For the proposed model, ROF model, MDH model and the median filter, the grey level range is [0,255]. For SDZ model, the grey level range is normalized to [0, 1]. For the proposed model, the range of τ is [0.3, 0.7], σ=τ/12, the range of λ is [15,30], and the range of η is [0.9, 3]. For MDH model, the range of λ is [25,50]. For SDZ model, the range of λ is [2,9] and the range of η is [0.3, 3]. For ROF model, the range of λ is [1,8].
According to the images in Figures 2–4, images visual quality of our method is better than others. Compared with TV-based methods, block-effects can be more significantly reduced by our method. The reason is that the solution of kernel space of second-order TGV is first-order polynomial, but not the piecewise constant function in BV space. In Tables 2 and 3, we list PSNR, SSIM values of numerical results. Clearly, PSNR, SSIM values of our method are better than others. According to the zoomed images in Figures 2–4, our method enhances the image quality and reduces noise more significantly while there is much more noise residual in images of other methods. According to the images in Figure 4, the structure around the eye is better preserved in the proposed method than others.
6.
Conclusions
Based on the Moreau envelop[30] idea and TGV regularization, we propose a new approach to Cauchy noise removal. We show that the solution of the proposed model is unique. In order to solve the new model, an alternating minimization method is employed and its convergence is proved. Numerical results demonstrate that the images quality of our method is better than that of some earlier restoration methods.
Acknowledgments
We really appreciate authors of [13] and [15] for their codes.
Conflict of interest
We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.