
Let u,v be two analytic functions on the open unit disk D in the complex plane, φ an analytic self-map of D, and m,n nonnegative integers such that m<n. In this paper, we consider the generalized Stević-Sharma type operator Tm,nu,v,φf(z)=u(z)f(m)(φ(z))+v(z)f(n)(φ(z)) acting from the derivative Hardy spaces into Zygmund-type spaces, and investigate its boundedness, essential norm and compactness.
Citation: Zhitao Guo, Jianyong Mu. Generalized Stević-Sharma type operators from derivative Hardy spaces into Zygmund-type spaces[J]. AIMS Mathematics, 2023, 8(2): 3920-3939. doi: 10.3934/math.2023196
[1] | Miyoun Jung . A variational image denoising model under mixed Cauchy and Gaussian noise. AIMS Mathematics, 2022, 7(11): 19696-19726. doi: 10.3934/math.20221080 |
[2] | Myeongmin Kang, Miyoun Jung . Nonconvex fractional order total variation based image denoising model under mixed stripe and Gaussian noise. AIMS Mathematics, 2024, 9(8): 21094-21124. doi: 10.3934/math.20241025 |
[3] | Donghong Zhao, Ruiying Huang, Li Feng . Proximity algorithms for the L1L2/TVα image denoising model. AIMS Mathematics, 2024, 9(6): 16643-16665. doi: 10.3934/math.2024807 |
[4] | Mingying Pan, Xiangchu Feng . Application of Fisher information to CMOS noise estimation. AIMS Mathematics, 2023, 8(6): 14522-14540. doi: 10.3934/math.2023742 |
[5] | Abdelilah Hakim, Anouar Ben-Loghfyry . A total variable-order variation model for image denoising. AIMS Mathematics, 2019, 4(5): 1320-1335. doi: 10.3934/math.2019.5.1320 |
[6] | Yating Zhu, Zixun Zeng, Zhong Chen, Deqiang Zhou, Jian Zou . Performance analysis of the convex non-convex total variation denoising model. AIMS Mathematics, 2024, 9(10): 29031-29052. doi: 10.3934/math.20241409 |
[7] | Yuzi Jin, Soobin Kwak, Seokjun Ham, Junseok Kim . A fast and efficient numerical algorithm for image segmentation and denoising. AIMS Mathematics, 2024, 9(2): 5015-5027. doi: 10.3934/math.2024243 |
[8] | Simone Costa, Marco Pavone . Orthogonal and oriented Fano planes, triangular embeddings of K7, and geometrical representations of the Frobenius group F21. AIMS Mathematics, 2024, 9(12): 35274-35292. doi: 10.3934/math.20241676 |
[9] | Lufeng Bai . A new approach for Cauchy noise removal. AIMS Mathematics, 2021, 6(9): 10296-10312. doi: 10.3934/math.2021596 |
[10] | Ming Shi, Ibrar Hussain . Improved region-based active contour segmentation through divergence and convolution techniques. AIMS Mathematics, 2025, 10(1): 654-671. doi: 10.3934/math.2025029 |
Let u,v be two analytic functions on the open unit disk D in the complex plane, φ an analytic self-map of D, and m,n nonnegative integers such that m<n. In this paper, we consider the generalized Stević-Sharma type operator Tm,nu,v,φf(z)=u(z)f(m)(φ(z))+v(z)f(n)(φ(z)) acting from the derivative Hardy spaces into Zygmund-type spaces, and investigate its boundedness, essential norm and compactness.
Images are inevitably deteriorated by noise that is generated during the image acquisition process. Thus, image denoising, which aims to retrieve a clean image by removing noise while conserving important structures and details, is a fundamental issue in image processing. This work focuses on color image denoising in the presence of multiplicative noise. Multiplicative noise usually occurs under coherent image systems such as synthetic aperture radar (SAR) [1], ultrasound imaging [2], and laser imaging [3]. Owing to the coherent nature of these image acquisition procedures, most of the information in an original image may be lost when corrupted by multiplicative noise. Hence, removing multiplicative noise is a more difficult task than removing additive Gaussian noise.
Let Ω⊂R2 be an open and bounded domain with a Lipschitz boundary. The observed image f:Ω→R corrupted by multiplicative noise can be formulated as
f=u⋅η, | (1.1) |
where u:Ω→R is the original clean image, and η denotes the multiplicative noise that follows a gamma distribution whose density function [4] is given by
g(η)=LLΓ(L)ηL−1e−Lη,forη≥0, | (1.2) |
where L>0 is an integer representing the noise level, and Γ is the gamma function. Thus, the mean of η is 1, and its standard deviation is 1/√L. This kind of noise commonly appears in SAR images.
For the removal of multiplicative gamma noise, a variety of filtering-based models [5,6,7,8,9] and variational models [10,11,12,13,14,15,16,17,18,19,20,21,22] have been developed. Among them, SAR-BM3D [9] is one of the most widely known filtering-based methods, which integrates nonlocal principles with wavelet representation. Particularly, it uses a similarity measure to collect 3-dimensional (3D) groups of similar image patches and applies wavelet decomposition to compute 3D blocks to achieve denoised images. This algorithm performs well, but is computationally very expensive and produces undesirable artifacts in cases of heavy multiplicative noise. Within a variational framework, Aubert and Aujol (AA) [10] proposed a new model combining a data-fidelity term derived from the gamma distribution and the total variation (TV) regularization. TV [25] has been widely used due to its edge preserving capacity and convexity. Nevertheless, the data-fidelity term of this model is not convex, thus the resulting computed solution may not be optimal from a global perspective. Moreover, this model highly depends on the initialization. To resolve these problems, previous studies [11,16,19] have suggested convex data-fidelity terms, by using a logarithmic transformation or adding a penalty term. In particular, Lu et al. [19] developed a convex data-fidelity term that is suitable for removing severe multiplicative noise, which integrates the ideas included in the works of [11,16]. All the above-mentioned models are proposed for grayscale image restoration. Recently, several studies [23,24] have been proposed to eliminate multiplicative noise in color images by utilizing the data-fidelity terms introduced in [10,11]. This work makes use of the data-fidelity term presented in [19] to better deal with heavy multiplicative noise in color images.
For color image restoration, TV regularization has been extended and studied in various works [24,26,27,28,29,30,31,32,33,34,35,36,37,38]. Especially, based on the view of color edge detection, Jia et al. [38] introduced a saturation-value-based TV (SVTV) regularization in the HSV color space. They reported in [38] that SVTV is superior to the existing RGB-based color TV methods in several image processing problems by considerably reducing artificial color values and producing well-preserved details and edges. Wang et al. [23] also adopted SVTV for the removal of multiplicative noise and blur in color images. However, SVTV creates a staircase effect; it produces some staircase shapes in smooth transition regions similar to other TV-based regularizations. To alleviate this shortcoming, higher-order derivatives-based SV regularizers were suggested in [24], which were utilized to remove multiplicative noise in color images. These regularizers benefited from both SV-based regularization and higher-order regularization. Despite the excellent performance of SV-based regularization in color image restoration, local derivatives-based regularization is limited due to the smearing of details and repeated patterns.
To overcome this drawback of local regularizers, various methods exploiting the nonlocal properties of images have been developed for Gaussian noise problems. First, based on the nonlocal self-similarity property that expresses the global repetitiveness of textures and structures in images, a nonlocal means filter was proposed in [39], which enhanced the local smoothing filters. As an extension, Dabov et al. [40] proposed a denoising algorithm, called BM3D, using block matching and 3D transform domain collaborative filtering, through which SAR-BM3D was developed. In addition, a series of nonlocal regularizations have been developed for inverse problems [41,42,43,44,45]. These nonlocal methods provide superior results in comparison to local methods, but many image details and structures cannot be restored due to inaccurate nonlocal weights. The patch-based sparse representation methods have also attracted much attention. Elad and Aharon [46] suggested a denoising algorithm, called K-SVD, by assuming that each patch of an image is accurately represented by a few elements from a dictionary that is learned from natural images. This also led to a variational model in [15] for removing multiplicative noise, which combines K-SVD and TV regularization. Regardless of the phenomenal results of K-SVD, it ignores self-similarity as each patch is considered independently in dictionary learning and sparse coding. In addition, the dictionary learning process is computationally expensive. To achieve better performance, sparsity and self-similarity were combined in several works [47,48,49], while low-rank-based approaches were also proposed in [22,50,51,52]. The aforementioned sparse representation methods consider an image patch as the basic unit of sparse representation.
Unlike these sparse representation methods, Zhang et al. [53] proposed a group-based sparse representation (GSR) model that uses an image group as its basic processing unit. The GSR model characterizes the local sparsity of patches and nonlocal self-similarity of repeated patterns between patches. This also involves an efficient self-adaptive group dictionary learning method, rather than dictionary learning from images. The model demonstrated excellent performance in image deblurring, inpainting, and compressive sensing. It was also utilized in diverse applications such as image deblocking [54], low-light image enhancement [55], SAR image despeckling [56], and Cauchy noise removal [57]. In several works [58,59,60], the GSR framework was extended, by integrating it with the low-rank approach or patch-based sparse representation. Despite the outstanding performance of these GSR-based models, for color image processing, they convert the image from the RGB space to the YCbCr space and only extract the Y (luminance) channel. Although the Y channel includes most of the important features of the image, such as edges and textures, the Y channel alone might not be sufficient for image processing when all RGB channels are independently corrupted by noise. Recently, in [61], a GSR model utilizing all channel information was proposed and applied to the color image inpainting problem. In this model, an image group corresponding to each RGB channel is constructed, and the final image group is formed from these three image groups. That is, this approach calculates the patch distance between two grayscale patches in each channel, rather than using the patch distance between two color patches. In our work, we exploit a multi-channel-based GSR approach for the removal of multiplicative noise in color images, but unlike the approach in [61], we compute the patch distance between two color image patches to form a color image group. Moreover, we also adopt the SVTV regularization to mitigate certain artifacts caused by patches.
The main contribution of our work is to propose a GSR-based model for restoring color images deteriorated by multiplicative gamma noise. The proposed GSR approach is based on multi-color channels, which is different from existing approaches. The model also involves a convex data-fidelity term and SVTV regularization. The proposed model sufficiently denoises color images in smooth regions while maintaining textures, details, and edges, even in the presence of severe multiplicative noise. Moreover, unlike existing methods, it does not generate artificial color values or artifacts. To numerically solve the proposed model, an efficient iterative algorithm is offered based on the alternating direction method of multiplies.
The outline of the paper is as follows. Section 2 reviews several variational models for removing multiplicative gamma noise in grayscale images, as well as SVTV regularization and a GSR approach. In Section 3, we present the proposed model and optimization algorithm. Section 4 presents the experimental results of the proposed model, with comparisons to existing methods. Finally, in Section 5, we summarize and conclude our work with some discussions.
In this section, we review several multiplicative noise removal models for grayscale images, in addition to SVTV regularization and a GSR model.
First, based on maximum a posterior estimation and Bayes' rule, Aubert and Aujol (AA) [10] proposed a TV model for denoising images in the presence of multiplicative gamma noise, with the assumption f>0:
minu>0∫Ω(logu+fu)dx+μ∫Ω|∇u|dx, | (2.1) |
where μ>0 is a parameter that manages the smoothness of the image u, and |∇u|=√(∂x1u)2+(∂x2u)2 with x=(x1,x2). TV regularization [25] has been widely utilized owing to its convexity and edge conservation ability. Nevertheless, since the data-fidelity term in model (2.1) is not convex, classical optimization algorithms may not furnish a global optimal solution and are highly reliant on initialization.
To rectify these problems, Shi and Osher (SO) [11] proposed the following convex model using the logarithmic transformation, w=logu:
minw∫Ω(w+fe−w)dx+μ∫Ω|∇w|dx. | (2.2) |
Thus, after solving this minimization problem to obtain w, the denoised image u is achieved as u=ew. The denoising performance of this model surpasses that of the AA model (2.1) and its denoising results do not depend on initial guesses.
As another approach to build up a convex data-fidelity term, Dong and Zeng [16] inserted a penalty term to the data-fidelity term in model (2.1):
minu>0∫Ω[logu+fu+α(√uf−1)2]dx+μ∫Ω|∇u|dx, | (2.3) |
where α>0 is a parameter. The objective function in (2.3) is convex if α≥2√69. This model is based on the observation that the value of E((√u/f−1)2) is close to 0 with a large L, and its value is still rather small even when L is small. However, E(√u/f) is always larger than 1 and is close to 1 when L is large enough, i.e., the noise level is relatively small.
To deal with heavy multiplicative noise, Lu et al. [19] substituted 1 in the penalty term with a parameter β≥1. In addition, they adopted the logarithmic transformation, w=log(u), since exponent-like models provide better denoising performance than logarithm-like models:
minw∫Ω[w+fe−w+α(√ewf−β)2]dx+μ∫Ω|∇w|dx, | (2.4) |
where β depends on the noise level L. The objective function in (2.4) is strictly convex if αβ4≤409627. The numerical results in [19] validated that the data-fitting term in (2.4) is suitable for eliminating heavy multiplicative noise.
The HSV color model is a color system similar to human perception. Hue refers to the color portion, saturation refers to the gray level of the color, and value refers to the brightness of the color and varies depending on saturation. The important contours and textures of an image lie primarily in its saturation and brightness components.
For a color image u=(ur,ug,ub), the saturation (S) and value (V) components can be expressed in ur, ug, and ub as
S(x)=13‖Cu(x)T‖2,V(x)=1√3|ur(x)+ug(x)+ub(x)|,withC=[2−1−1−12−1−1−12],u(x)T=[ur(x)ug(x)ub(x)], | (2.5) |
where ‖⋅‖2 is the vector Euclidean norm.
Based on the observation that the saturation and value components contain most of the important edges and textures of an image, Jia et al. [38] proposed a saturation-value total variation (SVTV) for color images as follows:
SV-TV(u)=sup(ψ1,ψ2)∈K2,ψ3∈K1{∫Ω1√2(ur(x)−ug(x))div(ψ1(x))+1√6(ur(x)+ug(x)−2ub(x))div(ψ2(x))+γ√3(ur(x)+ug(x)+ub(x))div(ψ3(x))}dx, | (2.6) |
where Kn=C1(Ω,B2n) stands for the set of continuously differentiable and bounded functions from the compact support in Ω to B2n, with Bn representing the closed united ball with a positive integer n, and the parameter γ>0 is the weight of the value component in the SVTV term. The convexity, lower semi-continuity, and compactness properties of SVTV were proven. Assuming u∈W1,2(Ω), SVTV can be written as
SV-TV(u)=∫Ω√∑j=1,2|∂xju(x)|2s+γ√∑j=1,2|∂xju(x)|2vdx, | (2.7) |
where |∂xju(x)|s and |∂xju(x)|v (j=1,2) are defined as
|∂xju(x)|s=13‖C∂xju(x)T‖2,|∂xju(x)|v=1√3|∂xjur(x)+∂xjug(x)+∂xjub(x)|, | (2.8) |
with ∂xju(x)T=(∂xjur(x),∂xjug(x),∂xjub(x))T. The experimental results in [38] demonstrate the superior restoration performance of SVTV over existing TV-based methods, by reducing the color artifacts induced by noise and better conserving the details.
This subsection discusses the GSR model in [53] proposed for grayscale images.
Given a grayscale image x∈RN, it is divided into n overlapping patches xk of size √P×√P (k=1,2,...,n), with stride s. For each patch xk, c most similar patches are searched in a local window of size L×L, and the selected patches are stacked into a matrix xGk=[xGk,1,xGk,2,...,xGk,c]∈RP×c as columns. The matrix xGk is called a group and can be defined by a linear operator RGk as
xGk=RGk(x). | (2.9) |
The transpose RTGk can convert a group into the k-th position in the image padded with zeros elsewhere.
The GSR model assumes that each group xGk can be accurately represented by a few atoms of a self-adaptive learning dictionary DGk=[dGk,1,dGk,2,...,dGk,m], where each atom dGk,i∈RP×c is a matrix of the same size as the group xGk and m=min(P,c). The sparse coding process of each group xGk over DGk is to find a sparse vector αGk=[αGk,1,αGk,2,...,αGk,m] such that
xGk=DGkαGk:=m∑i=1αGk,idGk,i. | (2.10) |
Then, the image x is reconstructed from the sparse codes {αGk} as
x=DG∘αG:=n∑k=1RTGk(DGkαGk)./n∑k=1RTGk(1P×c), | (2.11) |
where DG and αG denote the concatenation of all DGk and αGk, respectively, ./ is the element-wise division of vectors, and 1P×c∈RP×c is a matrix of ones.
Assuming that a degraded image is given by y=Hx+n, where H is a linear degradation operator and n is noise, the GSR model for image restoration can be formulated as
ˆαG:=argminαG12‖HDG∘αG−y‖22+λ‖αG‖0, | (2.12) |
where λ>0 is a regularization parameter, and ‖αG‖0 indicates the number of non-zero elements in αG. By selecting an appropriate self-adaptive group dictionary, the minimization problem (2.12) can be explicitly solved. Then, the reconstructed image is obtained as ˆx=DG∘ˆαG.
In this section, we present a GSR model to restore color images degraded by multiplicative gamma noise and an optimization algorithm for solving the proposed model.
Let us assume that a noisy color image f∈R3N is given by
fch=uch⋅ηch>0,ch=r,g,b, | (3.1) |
where u∈R3N is the clean image, and ηch∈RN represents the gamma noise in the ch-channel that follows the distribution in (1.2).
The problem (3.1) is an ill-posed inverse problem, so we formulate a minimization problem to reconstruct u from the given data f. First, following the process in (2.1)–(2.4) and by letting w=log(u), we consider the following convex data-fidelity term to handle heavy multiplicative gamma noise:
F(w)=⟨w+fe−w+α(√ewf−β)2,1⟩, | (3.2) |
where ⟨⋅,⋅⟩ is the inner product, 1∈R3N, and α, β>0 are parameters satisfying αβ4≤409627.
Next, we introduce our GSR approach to reconstruct color images corrupted by multiplicative gamma noise. For a color image x∈R3N, we extract a color image patch xk=[x1k,x2k,x3k]T∈R3P, where xchk is a patch of size √P×√P in the ch-channel (ch=1,2,3). Then the group xGk consisting of c most similar patches is given by
xGk=[x1Gk,1x1Gk,2⋯x1Gk,cx2Gk,1x2Gk,2⋯x2Gk,cx3Gk,1x3Gk,2⋯x3Gk,c]∈R3P×c, | (3.3) |
and the self-adaptive learning dictionary DGk is as follows
DGk=[dGk,1dGk,2⋯dGk,m]∈R(3P×c)×m, | (3.4) |
where dGk,i=[d1Gk,i,d2Gk,i,d3Gk,i]T∈R3P×c (i=1,⋯,m) with m=min(3P,c).
To search for similar patches in the image w, we first take the logarithmic transformation in (3.1), which yields
ˆf=w+ˆη, | (3.5) |
where ˆf=log(f), and ˆη=log(η). In Figure 1, we plot the histograms of noise ˆη in the RGB or HSV color spaces when the noise level L=10, 5, and 3. The figure shows that the histograms of ˆη in both color spaces at all noise levels are similar to normal distributions. Especially, the noise distribution in the HSV color space fits better with a normal distribution. Based on this observation, we extract the color patches in the HSV color space and compute the Euclidean distance between them. Specifically, after transforming the RGB-based color image w to the HSV-based image ˜w, we extract two color image patches, ˜wk1 and ˜wk2∈R3P, from the transformed image ˜w and compute the similarity measure between the two patches using the following distance function:
dist(˜wk1,˜wk2)=‖˜wk1−˜wk2‖22. | (3.6) |
Thus, the group ˜wGk and dictionary ˜DGk will be attained in the HSV color space, which is discussed in the subsequent subsection.
Lastly, despite the outstanding restoration performance of the GSR-based approaches, these patch-based methods also cause some ringing or block artifacts in the restored images. To alleviate these artifacts, we also utilize the local regularizer, SVTV, defined in (2.7). Consequently, by combining the data-fidelity term in (3.2) with GSR and SVTV regularization, we propose the following minimization problem for the elimination of multiplicative gamma noise in color images:
minw,αGF(w)+λ‖αG‖0+μSV-TV(w),subject to:w=DG∘αG, | (3.7) |
where λ and μ>0 are regularization parameters.
Since problem (3.7) is nonconvex and nonsmooth, it is challenging to solve. Therefore, we utilize the alternating direction method of multipliers (ADMM) [62,63] that are proposed for solving convex or nonconvex linearly constrained problems.
First, we utilize the variable splitting technique. That is, we introduce an auxiliary variable z to replace w in the SVTV term. Then, problem (3.7) can be transformed into the following constrained problem:
minw,αG,zF(w)+λ‖αG‖0+μSV-TV(z),subject to:w=DG∘αG,z=w. | (3.8) |
The augmented Lagrangian function (ALF) for problem (3.8) is given by
Lτ(w,αG,z;p,q)=F(w)+λ‖αG‖0+μSV-TV(z)−⟨p,w−DG∘αG⟩+τ2‖w−DG∘αG‖22−⟨q,z−w⟩+τ2‖z−w‖22, | (3.9) |
where p∈R3N and q∈R3N are the Lagrangian multipliers, and τ>0 is a penalty parameter.
The ADMM applied to problem (3.8) results in the following iterative algorithm:
{(αℓ+1G,zℓ+1):=argminαG,zLτ(wℓ,αG,z;pℓ,qℓ),wℓ+1:=argminwLτ(w,αℓ+1G,zℓ+1;pℓ,qℓ),pℓ+1=pℓ−τ(wℓ+1−DG∘αℓ+1G),qℓ+1=qℓ−τ(zℓ+1−wℓ+1). | (3.10) |
In the subsequent paragraphs, we solve two subproblems in (3.10).
First, we solve the (αG,z)-subproblem in (3.10), which is formulated as
minαG,zλ‖αG‖0+τ2‖DG∘αG−r‖22+τ2‖z−s‖22+μSV-TV(z), | (3.11) |
where r=wℓ−pℓ/τ, and s=wℓ+qℓ/τ.
As perviously mentioned, we measure the Euclidean distance in the HSV color space to find similar patches. Moreover, to handle the SVTV term defined in the RGB color space, we convert the RGB-based model (3.11) to a HSV-based model using the following orthogonal matrix ˉP, that is derived the dual-form of SVTV in (2.6) and has a relation with C in (2.5) as follows:
ˉP=[1√2−1√201√61√6−2√61√31√31√3],C=ˉPTAˉPwithA=[300030000]. | (3.12) |
If we define ˜u(x)=ˉPu(x)T for x∈Ω, then the S and V components can also be expressed as S(x)=‖(˜u1(x),˜u2(x))T‖2 and V(x)=|˜u3(x)| respectively. In addition, |∂xju(x)|s and |∂xju(x)|v in (2.8) can be expressed as
|∂xju(x)|s=‖13ˉPTAˉPˉPT∂xj˜u(x)T‖2=‖(∂xj˜u1(x),∂xj˜u2(x))T‖2,|∂xju(x)|v=|∂xj˜u3(x)|, | (3.13) |
where the second equality of the first term is because the transpose of an orthogonal matrix is also orthogonal, ˉPˉPT is the identity matrix, and the norm of a vector is invariant under multiplication by an orthogonal matrix.
Now, using the following orthogonal matrix
P=[1√2I−1√2I01√6I1√6I−2√6I1√3I1√3I1√3I], | (3.14) |
where I∈RN is the identity matrix, we define ˜DG=PDG, ˜r=Pr, ˜s=Ps, and v=(v1,v2,v3)T=Pz. Therefore, problem (3.11) can be converted to
minαG,vλ‖αG‖0+τ2‖˜DG∘αG−˜r‖22+τ2‖v−˜s‖22+μ(‖∇ˉv‖1+γ‖∇v3‖1), | (3.15) |
where ˉv=(v1,v2)T and ∇ˉv=(∇v1,∇v2)T. In this problem, the variables αG and v are independent of each other, so problem (3.15) can be separated into two problems as
{αℓ+1G:=argminαGτ2‖˜DG∘αG−˜r‖22+λ‖αG‖0,vℓ+1:=argminvτ2‖v−˜s‖22+μ(‖∇ˉv‖1+γ‖∇v3‖1). | (3.16) |
Solving the αG-subproblem in (3.16). The αG-subproblem in (3.16) can be reformulated as
minαG12‖˜x−˜r‖22+λτ‖αG‖0, | (3.17) |
where ˜x=˜DG∘αG. The following proposition indicates that the error term in (3.17) can be well approximated by the sum of the error terms of groups under certain assumptions:
Proposition 1. Let ˜x, ˜r∈R3N, ˜xGk, ˜rGk∈R3P×c, and assume that each element of e=˜x−˜r is independent and follows a distribution with zero mean and variance σ2. Then, for any ε>0, a relationship between ‖˜x−˜r‖22 and ∑nk=1‖˜xGk−˜rGk‖22 holds as follows:
limN→∞K→∞Pr{|13N‖˜x−˜r‖22−1Kn∑k=1‖˜xGk−˜rGk‖22|<ε}=1, |
where Pr(⋅) represents the probability, and K=3P×c×n.
Proof. It follows from the law of large numbers, see [53].
According to Proposition 1 and using ‖αG‖0=∑nk=1‖αGk‖0, problem (3.17) is equivalent to
argminαGn∑k=1(12‖˜xGk−˜rGk‖22+˜τ‖αGk‖0), | (3.18) |
where ˜τ=λK3τN. This problem can be efficiently minimized by solving nαGk-subproblems, one of which is formulated as
argminαGk12‖˜xGk−˜rGk‖22+˜τ‖αGk‖0=12‖˜DGkαGk−˜rGk‖22+˜τ‖αGk‖0, | (3.19) |
where ˜DGk is the self-adaptive learned dictionary from ˜rGk as follows.
First, the singular value decomposition is applied to ˜rGk:
˜rGk=UGkΣGkVTGk=m∑i=1γ˜rGk,i(uGk,ivTGk,i), | (3.20) |
where ΣGk=diag(γ˜rGk) is a diagonal matrix with the elements of γ˜rGk=[γ˜rGk,1,γ˜rGk,2,...,γ˜rGk,m] on its main diagonal, and uGk,i and vGk,i are the columns of UGk and VGk, respectively. Then, the atoms in ˜DGk is defined as ˜dGk,i=uGk,ivTGk,i, thus the dictionary ˜DGk for ˜xGk is given by
˜DGk=[˜dGk,1,˜dGk,2,...,˜dGk,m]. | (3.21) |
Using ˜rGk=˜DGkγ˜rGk and due to the unitary property of ˜DGk, the αGk-subproblem in (3.19) is equivalent to
argminαGk12‖αGk−γ˜rGk‖22+˜τ‖αGk‖0. | (3.22) |
This problem has a closed-form solution as
ˆαGk=hard(γ˜rGk,√2˜τ)=γ˜rGk⊙1(|γ˜rGk|−√2˜τ), | (3.23) |
where hard(⋅) represents the hard thresholding operator, ⊙ represents the element-wise multiplication, and 1(y) is the indicator function that has a value of 1 if y>0 and 0 otherwise.
Solving the v-subproblem in (3.16). Next, we solve the v-subproblem in (3.16). To solve this convex problem, we again adopt the ADMM in [62]. First, to handle the non-differentiable terms, we introduce two auxiliary variables d1 and d2 to substitute ∇ˉv and ∇v3, respectively. Then, the unconstrained v-problem in (3.16) is converted into the following constrained minimization problem:
minv,d1,d2τ2‖v−˜s‖22+μ‖d1‖1+μγ‖d2‖1,subject to:d1=∇ˉv,d2=∇v3. | (3.24) |
The ALF of problem (3.24) is given by
Lξ(v,d1,d2;b1,b2)=τ2‖v−˜s‖22+μ‖d1‖1+μγ‖d2‖1−⟨b1,d1−∇ˉv⟩+ξ2‖d1−∇ˉv‖22−⟨b2,d2−∇v3⟩+ξ2‖d2−∇v3‖22, | (3.25) |
where b1∈R3N×4 and b2∈R3N×2 are the Lagrangian multipliers, and ξ>0 is a penalty parameter.
The ADMM applied to problem (3.24) brings the following iterative algorithm:
{vt+1:=argminvLξ(v,dt1,dt2,bt1,bt2),(dt+11,dt+12):=argmind1,d2Lξ(vt+1,d1,d2,bt1,bt2),bt+11=bt1−ξ(dt+11−∇ˉvt+1),bt+12=bt2−ξ(dt+12−∇v3t+1). | (3.26) |
The constraint in (3.24) can be written as (d1,d2)T=B(ˉv,v3)T with B=[∇00∇]. Since the matrix B has a full column-rank, we can obtain the following convergence results, according to Theorem 1 in [64]: If the sequence {(vt,dt1,dt2,bt1,bt2)} is generated by ADMM in (3.26), then this sequence has a single limit point (v∗,d∗1,d∗2,b∗1,b∗2), and this limit point is a saddle point of the augmented Lagrangian Lξ. That is, {(vt,dt1,dt2)} has a single limit point (v∗,d∗1,d∗2), and this limit point solves problem (3.24).
The v-subproblem in (3.26) can be rewritten as
minvτ2‖v−˜s‖22+ξ2‖dt1−∇ˉv−bt1/ξ‖22+ξ2‖dt2−∇v3−bt2/ξ‖22. | (3.27) |
From the first-order optimality condition, vt+1 must satisfy the following normal equations:
(τ+ξ∇T∇)vi=τsi+ξ∇T(dt1,i−bt1,i/ξ),i=1,2,(τ+ξ∇T∇)v3=τs3+ξ∇T(dt2−bt2/ξ), | (3.28) |
where ∇T=−div with div being the discrete divergence operator defined as div(s1,s2)=∂x1s1+∂x2s2. Assuming the periodic boundary condition, the above equations can be explicitly solved using the 2-dimensional discrete Fourier transform (DFT).
The (d1,d2)-subproblem in (3.26) can be separated into two subproblems for each variable since the variables are independent of each other. Specifically, the di-subproblems can be written as
dt+11:=argmind1μ‖d1‖1+ξ2‖d1−(∇ˉvt+1+bt1/ξ)‖22,dt+12:=argmind2μγ‖d2‖1+ξ2‖d2−(∇v3t+1+bt2/ξ)‖22. | (3.29) |
Then, the solutions dt+1i can be obtained as follows
dt+11=shrink(∇ˉvt+1+bt1/ξ,μ/ξ),dt+12=shrink(∇vt+13+bt2/ξ,μγ/ξ), | (3.30) |
where shrink(⋅) is the soft-thresholding operator defined as
shrink(a,b)=a|a|⋅max(|a|−b,0). | (3.31) |
Now, we solve the w-subproblem in (3.10), which can be written as
minwF(w)+τ2‖w−xℓ+1−pℓ/τ‖22+τ2‖zℓ+1−w−qℓ/τ‖22, | (3.32) |
where xℓ+1=PT˜xℓ+1 with ˜xℓ+1=(˜DG∘ˆαG)ℓ+1, and zℓ+1=PTvℓ+1. This minimization problem is convex, but it is difficult to obtain an explicit form of solution due to the complex data-fidelity term. Fortunately, the problem can be efficiently solved by using Newton's method as follows:
wt+1=wt−ˆF′(wt)+τ(wt−xℓ+1−pℓ/τ)+τ(wt+qℓ/τ−zℓ+1)ˆF″(wt)+2τ, | (3.33) |
where ˆF′(w) and ˆF″(w) are given by
ˆF′(w)=1−fe−w+α(ewf−β√ewf),ˆF″(w)=fe−w+α(ewf−β2√ewf). | (3.34) |
In practice, only a few iteration numbers are required for convergence.
Consequently, the algorithm for solving model (3.7) is summarized in Algorithm 1.
Algorithm 1 Solving the proposed model (3.7) |
1: Input: choose the parameters α,β,λ,μ,γ,τ,ξ>0, patch size P, similar patch number c, window size L, stride s, maximum iteration numbers Mout, Mv, Mw, tolerance values tolout, tolv, tolw.
2: Initialization: set w0=log(f), p0=q0=0, b0i=0 (i=1,2). 3: repeat 4: Let ˜r=P(wℓ−pℓ/τ) and ˜s=P(wℓ+qℓ/τ). 5: Compute ˜xℓ+1=(˜DG∘ˆαG)ℓ+1 using ˜DGk in (3.21) and ˆαGk in (3.23). 6: Compute vℓ+1 by iterating for t=0,1,2,⋯,Mv: 7: compute vt+1 by solving (3.28) using DFT, 8: compute dt+1i (i=1,2) as in (3.30), 9: compute bt+1i (i=1,2) as in (3.26), 10: Let xℓ+1=PT˜xℓ+1 and zℓ+1=PTvℓ+1. 11: Compute wℓ+1 by iterating for t=0,1,2,⋯,Mw: 12: wt+1=wt−ˆF′(wt)+τ(wt−xℓ+1−pℓ/τ)+τ(wt+qℓ/τ−zℓ+1)ˆF″(wt)+2τ, with ˆF′ and ˆF″ in (3.34), 13: pℓ+1=pℓ−τ(wℓ+1−xℓ+1), 14: qℓ+1=qℓ−τ(zℓ+1−wℓ+1), 15: until a stopping criterion is satisfied. 16: Output: restored image u=ew. |
This section presents the numerical results of the proposed model, with comparisons to other existing models including L2-SVTV in [38] and AA-SVTV in [23], that combine SVTV with the L2 data-fidelity term or the data-fidelity term in (2.1), in addition to SO-SVNTVTV2 in [24] that integrates the data-fidelity term in (2.2) with a SV-based nonconvex higher-order regularization. We also compare our model with SAR-BM3D which is applied to each RGB channel. For a fair comparison, we adopt the variable splitting approach and ADMM for solving the aforementioned existing variational models. All numerical experiments were implemented using MATLAB R2023a on a 64-bit Windows 10 desktop PC with an Intel CPU at 4.2GHz and 64GB RAM. All numerical results can be found in the material at the following link: https://url.kr/svqunr.
We test each model on 16 natural images and five SAR images, provided in Figure 2. The range of intensity values in the original images is assumed to be [0,255]. In the experiments, all test images are corrupted by multiplicative gamma noise with noise levels of L=10, 5, or 3.
To evaluate the effectiveness of the models, we calculate the peak-signal-to-noise-ratio (PSNR) value defined as
PSNR(u,u∗)=10log10(3N×2552‖u−u∗‖22), |
where u and u∗ denote the restored and original images, respectively, and N is the image size. We also compute the structural similarity (SSIM) index [65], which is a perception-based measure that considers the structure of the objects in the visual scene. Specifically, we compute the average of the SSIM values calculated from each RGB channel. For all models, we adjust the parameters to achieve the optimal restored images, taking into account both the visual quality and the PSNR and SSIM values.
The stopping criterion for the proposed model is as follows:
‖hiter−hiter−1‖2‖hiter‖2≤tolork>M, |
where h=w or v, and tol>0 and M>0 are the given tolerance and maximum iteration number, respectively. We set tol=2×10−3 and M=30 for the outer iteration, and tol=10−3 and M=10 for the v-subproblem and Newton's method for w.
The parameter settings of the proposed model are as follows. First, in the data-fidelity term, parameter α, which controls the effect of the penalty term, is fixed at 0.1. The value of β inside the penalty term is closely related to the noise level L and is chosen as 10.1113+0.1109L2+1 as suggested in [19]. The parameter γ in the SVTV term, which controls the weight of the value component, is set to 0.6, while parameter μ, which controls the effect of SVTV regularization, is fixed at 0.05 throughout the experiment. For the GSR representation, the image patch size, √P×√P, is set as 6×6 for natural images and 5×5 for SAR images. The number of similar patches, c, is set as 80, so the size of a group is 108×80 for natural images and 75×80 for SAR images. The size of the search window is 20×20. The overlapping width between adjacent patches is s=4 pixels. The parameters τ and λ also affect the GSR representation; the larger the τ or λ, the smoother the restored image becomes. However, the larger the τ, the longer the ADMM algorithm takes, so we fix parameter τ as 3. The parameter λ mainly affects the quality of restored images, so the optimal values of λ are chosen differently depending on the noise levels and images. For instance, λ∈{0.08,0.1,0.11} when L=10, λ∈{0.13,0.15,0.17} when L=5, and λ∈{0.22,0.24,0.26} when L=3. The specific value of λ is given in each figure. The ADMM parameter ξ is fixed at 1. For other variational models, we set γ=0.4 for L2-SVTV and γ=0.6 for AA-SVTV and SO-SVNTVTV2. Specific values of the regularization parameters for all variational models are provided in each figure in the supplementary file at the above link.
In Figures 3–6, we present the denoising results of the proposed model tested on natural color images, with comparisons to other models.
First, Figure 3 presents the denoising results tested on the Hill image with different noise levels including L=10, 5, and 3. The results show that L2-SVTV does not sufficiently remove the noise in some bright areas, such as the sky and white wall areas. This results in erroneous color values in those areas and makes the entire image dark. It also smooths out dark regions such as the roof and right tree areas, which can be clearly seen in the zoomed-in image in Figure 4. Contrarily, AA-SVTV adequately smooths bright regions while retaining noise in the dark regions. In addition, it induces certain artifacts such as black dots, especially in the dark regions. This becomes worse as the noise level increases, as shown in the zoomed-in image in Figure 4. Moreover, as the noise level increases, the entire color tone of the restored image of AA-SVTV is inclined to be distorted. For instance, the color tone of the red roof is degraded compared to those of SO-SVNTVTV2 and our model. SO-SVNTVTV2 suitably eliminates noise in both bright and dark regions, with fewer faulty color values or artifacts than L2-SVTV and AA-SVTV. These indicate the advantages of the SO data-fidelity term when removing multiplicative gamma noise. Additionally, SO-SVNTVTV2 alleviates the staircase effect caused by TV-based models. However, our model improves SO-SVNTVTV2 by preserving more details and generating more naturally restored images with no staircase artifacts and fewer artificial colors. This also leads to the higher PSNR and SSIM values of our model. In contrast, SAR-BM3D also sufficiently conserves edges and fine features, but it produces artificial color values and undesirable artifacts near edges, which deteriorates the denoising performance of SAR-BM3D on color images. These findings demonstrate the effectiveness of the proposed model for restoring color images polluted by multiplicative gamma noise.
In Figure 5, we present the denoising results of all models when L=5 and 3. Similar to the previous results, L2-SVTV oversmooths the dark regions, such as the hair and textured background in the Barbara image, the stems of flowers in the Building image and the body of statue in the Statues image. It also yields improper color values in some bright regions. This can be observed in the scarf in the Barbara image, the wall in the Building image and the sky area in the Statues image. AA-SVTV retains some noise or black dots in the aforementioned dark regions. This can also be seen in the sky area in the Building image, as shown in the zoomed-in image in Figure 6. As the noise level increases, the denoising performance of L2-SVTV and AA-SVTV significantly deteriorates. L2-SVTV generates the degraded color tones of restored images, while AA-SVTV supplies the black dots in the entire image region, as observed in the Peppers image. SO-SVNTVTV2 not only properly denoises the image in both the dark and bright regions, but also mitigates the staircase artifacts that appeared in the restored images from L2-SVTV and AA-SVTV. However, SO-SVNTVTV2 fails to preserve fine textures and details, as seen in the scarf in the Barbara image and the window areas in the Building image. In contrast, these textural regions are well conserved in SAR-BM3D and our model. This leads to the higher PSNR values of SAR-BM3D and our model than SO-SVNTVTV2 in the Barbara and Building images. Nevertheless, SAR-BM3D produces faulty color values, especially near edges, while SO-SVNTVTV2 and our model do not. Especially when L=3, SAR-BM3D produces artificial color values in the entire image region and certain artifacts induced by patches. These can be more obviously seen in the entire image of Peppers and the zoomed-in image of Statues in Figure 6. Consequently, our model provides the cleanest and clearest restored images among the models, while retaining textures, details, and sharp edges. This also corresponds to the highest PSNR and SSIM values. These examples also confirm the superior denoising performance of the proposed model over existing models, even in the presence of heavy multiplicative gamma noise.
In Figure 7, we show the effect of our multi-channel-based GSR approach by using the proposed model (3.7) with two different GSR approaches. Specifically, our GSR approach selects similar color image patches by computing the distance between two color patches and constructing an image group using these selected color patches. The other GSR approach chooses similar image patches in each color channel by measuring the patch distance in each channel. For comparable results, we use P=8×8 and c=80 for this decoupled GSR approach. In the Barbara image, the decoupled GSR approach slightly better preserves textures such as the scarf parts, which leads to higher SSIM values than our coupled GSR approach. However, the decoupled GSR approach furnishes artificial color values in all the examples, unlike our GSR approach. Thus, our GSR approach provides higher PSNR and SSIM values in most examples. These demonstrate the efficacy of our GSR approach for the removal of multiplicative noise in color images.
Figure 8 displays the effect of the data-fidelity term in the proposed model. In particular, we show the effect of the penalty term, (√ew/f−β)2, included in the data-fidelity term, since the effectiveness of the SO data-fidelity term has already been shown in the previous examples compared to the L2 or AA data-fidelity terms. That is, we compare our model with the model (3.7) without this penalty term. Although the restored images look almost identical, the use of the penalty term produces reconstructed images with higher PSNR and SSIM values. Therefore, these results validate the superiority of our data-fidelity term, compared to the SO data-fidelity term. For more examples, please refer to the supplementary file at the above link.
Figure 9 illustrates the effect of the SVTV term in the proposed model. Specifically, we compare our model with the model (3.7) without the SVTV term. The GSR model (3.7) without the SVTV term generates certain ringing artifacts near edges and block artifacts in homogeneous regions, which are usually induced by patch-based methods. However, the SVTV term alleviates or diminishes these artifacts by sufficiently smoothing these areas, as shown in the images in the bottom row. These show the effectiveness of the SVTV term in the proposed model.
Figure 10 presents the denoising results of all models tested on real SAR images, when L=3. The SAR images consist of many dark regions such as trees and shadows, while including some bright regions such as buildings and ships. Similar to the natural color image cases, L2-SVTV smooths out dark areas but does not adequately remove noise in bright areas, leading to inappropriate color values. In addition, the color tone of the restored image of L2-SVTV is significantly degraded compared to other models. In contrast, AA-SVTV sufficiently denoises the bright regions but supplies some black artifacts in the homogeneous regions including the grass or ground, and dark regions such as building shadows and tree areas. On the other hand, the denoising performance of SAR-BM3D on color SAR images is not as good as on natural color images. SAR-BM3D performed somewhat well on natural color images including large repeated textures or various color values, such as the Barbara, Building, and Peppers images, as shown in Figure 5. However, color SAR images do not usually contain large repeated textures or various color values. Instead, they include many small objects and edges. In color SAR images, SAR-BM3D sufficiently eliminates noise in both dark and bright regions and preserves details well, but it produces many erroneous color values in these small objects and edges. Thus, its PSNR and SSIM values are much lower than those of SO-SVNTVTV2 and our model at all noise levels, unlike the case of natural color images. SO-SVNTVTV2 conducts well on color SAR images, by appropriately denoising both dark and bright areas in addition to retaining fine details and edges. Additionally, it does not generate faulty color values or artifacts compared to SAR-BM3D or AA-SVTV. The restored images of SO-SVNTVTV2 and our model are visually very similar, but our model further improves SO-SVNTVTV2 by conserving more details and edges, thereby yielding cleaner and sharper restored images. In all cases, this results in higher PSNR and SSIM values for our model. Therefore, these examples establish the considerable denoising performance of the proposed model, even on real SAR images.
Figure 11 displays plots of the relative errors of wℓ and PSNR values of uℓ=ewℓ via the outer iteration number ℓ for the proposed model, when L=10, 5, and 3. The figure shows that as ℓ increases, the relative error ultimately decreases, while the PSNR value increases and converges to a certain value. In addition, as the noise level L increases, a higher number of outer iterations is required for convergence. These plots verify the numerical convergence of the proposed algorithm, even though the theoretical convergence of the algorithm is still a remaining issue.
Lastly, in Tables 1–3, we present the PSNR and SSIM values of all models, when L=10, 5, and 3, respectively. The results show that the PSNR and SSIM values of AA-SVTV decrease significantly at noise level L=3. Furthermore, SAR-BM3D provides similar PSNR and SSIM values as SO-SVNTVTV2 for natural color images, but its PSNR and SSIM values are considerably diminished for color SAR images. In all cases, the proposed model yields the highest PSNR and SSIM values. These measurements also ensure the outstanding denoising performance of the proposed model compared to other models.
Image | Data f | L2-SVTV | AA-SVTV | SO-SVNTVTV2 | SAR-BM3D | Proposed |
Airplane | 12.69/0.1970 | 23.93/0.7176 | 24.83/0.7965 | 25.06/0.7985 | 25.34/0.7419 | 26.83/0.8246 |
Baboon | 15.30/0.4449 | 21.55/0.5299 | 22.39/0.6325 | 22.47/0.6135 | 22.83/0.6569 | 23.54/0.6702 |
Barbara | 16.00/0.3362 | 25.16/0.6897 | 26.25/0.7587 | 26.48/0.7784 | 27.66/0.8110 | 29.41/0.8699 |
Bears | 18.12/0.3207 | 27.83/0.6846 | 28.38/0.7433 | 28.85/0.7496 | 28.18/0.7242 | 29.34/0.7690 |
Boat | 15.16/0.3794 | 24.42/0.7217 | 26.16/0.8042 | 26.44/0.8047 | 25.80/0.7766 | 27.52/0.8279 |
Building | 17.30/0.3882 | 25.76/0.7974 | 25.55/0.7840 | 26.68/0.8357 | 26.88/0.8337 | 28.68/0.8737 |
Butterfly | 16.55/0.4424 | 23.24/0.8235 | 24.07/0.8508 | 25.31/0.8841 | 24.09/0.8558 | 26.85/0.9149 |
Castle | 15.74/0.2609 | 25.85/0.7728 | 26.89/0.7946 | 27.50/0.8421 | 27.04/0.8179 | 28.72/0.8665 |
Cathedral | 15.37/0.2984 | 23.69/0.7055 | 24.74/0.7798 | 25.21/0.7744 | 25.16/0.7547 | 26.79/0.7999 |
Flower | 19.31/0.3042 | 30.82/0.8958 | 30.92/0.8462 | 32.12/0.9150 | 31.53/0.9004 | 32.85/0.9253 |
Hill | 17.06/0.4076 | 25.64/0.6292 | 26.83/0.7389 | 27.05/0.7323 | 26.96/0.7249 | 27.62/0.7401 |
House | 14.62/0.1839 | 26.97/0.7481 | 28.03/0.7816 | 28.41/0.7904 | 29.76/0.7776 | 30.64/0.8087 |
Lena | 15.10/0.2795 | 25.09/0.7358 | 26.96/0.8072 | 27.05/0.8109 | 27.41/0.7968 | 29.26/0.8606 |
Peppers | 15.88/0.3609 | 25.04/0.7566 | 25.75/0.8066 | 26.51/0.8182 | 27.44/0.8176 | 28.64/0.8526 |
Policemen | 16.76/0.4615 | 23.41/0.7372 | 24.92/0.8372 | 25.70/0.8576 | 25.09/0.8469 | 27.48/0.9086 |
Statues | 17.39/0.5465 | 23.84/0.6094 | 24.88/0.7734 | 25.23/0.7525 | 24.18/0.7156 | 26.10/0.7933 |
Average | 16.15/0.3139 | 25.14/0.6209 | 26.10/0.6813 | 26.63/0.6895 | 26.58/0.6789 | 28.14/0.7202 |
Remote1 | 20.05/0.5335 | 26.79/0.7756 | 27.03/0.8134 | 27.83/0.8411 | 26.19/0.8004 | 28.81/0.8689 |
Remote2 | 21.31/0.5074 | 28.81/0.8124 | 29.07/0.8497 | 30.23/0.8822 | 28.47/0.8481 | 31.00/0.8984 |
Remote3 | 19.89/0.6262 | 26.55/0.7933 | 27.06/0.8624 | 27.92/0.8690 | 26.05/0.8295 | 28.83/0.8949 |
Remote4 | 19.68/0.5646 | 26.47/0.7679 | 27.28/0.8453 | 28.32/0.8583 | 26.80/0.8252 | 29.31/0.8797 |
Remote5 | 18.50/0.4210 | 26.98/0.7480 | 27.46/0.7952 | 28.04/0.8052 | 26.73/0.7587 | 28.82/0.8255 |
Average | 19.87/0.5305 | 27.12/0.7794 | 27.58/0.8332 | 28.47/0.8512 | 26.85/0.8124 | 29.35/0.8735 |
* Bold values indicate the best denoising performance. |
Image | Data f | L2-SVTV | AA-SVTV | SO-SVNTVTV2 | SAR-BM3D | Proposed |
Airplane | 9.66/0.1375 | 22.54/0.6631 | 23.35/0.7415 | 23.59/0.7518 | 23.70/0.6709 | 25.10/0.7781 |
Baboon | 12.34/0.3214 | 20.69/0.4526 | 21.15/0.5497 | 21.18/0.5063 | 21.46/0.5419 | 22.48/0.6147 |
Barbara | 12.96/0.2297 | 24.27/0.6355 | 24.64/0.6809 | 24.81/0.6954 | 25.97/0.7455 | 27.66/0.8271 |
Bears | 15.12/0.2077 | 26.71/0.6307 | 26.74/0.6634 | 27.54/0.6836 | 26.92/0.6572 | 27.98/0.7070 |
Boat | 12.13/0.3016 | 23.08/0.6592 | 24.53/0.7570 | 24.73/0.7573 | 24.35/0.7196 | 25.98/0.7875 |
Building | 14.33/0.2730 | 23.89/0.7217 | 23.78/0.7022 | 24.92/0.7786 | 25.14/0.7756 | 27.02/0.8386 |
Butterfly | 13.61/0.3445 | 21.47/0.7615 | 22.02/0.7914 | 23.42/0.8385 | 22.63/0.8071 | 24.83/0.8792 |
Castle | 12.73/0.1732 | 24.55/0.7319 | 25.12/0.7526 | 25.93/0.8017 | 25.49/0.7594 | 27.19/0.8347 |
Cathedral | 12.37/0.2070 | 22.74/0.6591 | 23.06/0.7142 | 23.48/0.7161 | 23.56/0.6864 | 25.19/0.7587 |
Flower | 16.31/0.2081 | 29.17/0.8699 | 29.46/0.8416 | 30.76/0.8954 | 30.10/0.8699 | 31.50/0.9085 |
Hill | 14.04/0.2879 | 24.43/0.5644 | 25.31/0.6597 | 25.76/0.6658 | 25.63/0.6557 | 26.67/0.7059 |
House | 11.60/0.1177 | 25.48/0.7153 | 26.28/0.7451 | 26.89/0.7645 | 27.85/0.7279 | 28.82/0.7811 |
Lena | 12.14/0.1993 | 24.17/0.6917 | 25.33/0.7709 | 25.62/0.7711 | 25.84/0.7363 | 27.67/0.8234 |
Peppers | 12.84/0.2734 | 23.70/0.7107 | 24.28/0.7513 | 24.66/0.7839 | 25.83/0.7727 | 26.98/0.8239 |
Policemen | 13.72/0.3619 | 21.82/0.6658 | 23.09/0.7760 | 23.78/0.7982 | 23.60/0.7856 | 25.60/0.8615 |
Statues | 14.35/0.4083 | 22.54/0.5237 | 23.24/0.6823 | 23.55/0.6469 | 22.92/0.6234 | 24.52/0.7188 |
Average | 13.14/0.2147 | 23.83/0.5733 | 24.46/0.6304 | 25.04/0.6399 | 25.06/0.6220 | 26.57/0.6832 |
Remote1 | 17.05/0.3980 | 24.92/0.6857 | 25.04/0.7425 | 26.11/0.7800 | 24.71/0.7295 | 26.99/0.8147 |
Remote2 | 18.29/0.3773 | 27.05/0.7483 | 27.14/0.8030 | 28.50/0.8380 | 27.02/0.7956 | 29.27/0.8577 |
Remote3 | 16.89/0.4868 | 24.60/0.7005 | 25.00/0.7987 | 26.11/0.8114 | 24.64/0.7644 | 27.06/0.8507 |
Remote4 | 16.67/0.4314 | 25.03/0.7070 | 25.25/0.7809 | 26.52/0.8024 | 25.32/0.7649 | 27.60/0.8397 |
Remote5 | 15.49/0.2986 | 25.66/0.6906 | 25.66/0.7244 | 26.51/0.7471 | 25.29/0.6906 | 27.25/0.7688 |
Average | 16.88/0.3984 | 25.45/0.7064 | 25.62/0.7699 | 26.75/0.7958 | 25.40/0.7490 | 27.63/0.8263 |
* Bold values indicate the best denoising performance. |
Image | Data f | L2-SVTV | AA-SVTV | SO-SVNTVTV2 | SAR-BM3D | Proposed |
Airplane | 7.41/0.1020 | 21.86/0.5995 | 22.11/0.6777 | 22.73/0.7236 | 22.71/0.6098 | 24.08/0.7308 |
Baboon | 10.12/0.2379 | 20.06/0.4003 | 20.03/0.4381 | 20.33/0.4231 | 20.60/0.4604 | 21.13/0.4823 |
Barbara | 10.80/0.1654 | 23.51/0.5944 | 23.44/0.6154 | 24.08/0.6503 | 24.75/0.6823 | 26.55/0.7818 |
Bears | 12.87/0.1438 | 26.02/0.5990 | 25.72/0.6054 | 26.72/0.6331 | 26.05/0.6092 | 27.06/0.6449 |
Boat | 9.91/0.2481 | 22.37/0.6059 | 22.99/0.7008 | 23.72/0.7197 | 23.39/0.6704 | 24.95/0.7456 |
Building | 12.07/0.1987 | 22.91/0.6753 | 22.62/0.6418 | 23.65/0.7241 | 23.86/0.7200 | 25.86/0.7960 |
Butterfly | 11.39/0.2756 | 20.46/0.7176 | 20.76/0.7504 | 22.17/0.8021 | 21.51/0.7607 | 23.49/0.8444 |
Castle | 10.46/0.1229 | 23.81/0.6931 | 23.84/0.7015 | 24.69/0.7628 | 24.25/0.7081 | 25.94/0.7863 |
Cathedral | 10.12/0.1505 | 21.88/0.6228 | 22.02/0.6584 | 22.52/0.6740 | 22.54/0.6341 | 24.10/0.7122 |
Flower | 14.10/0.1508 | 28.35/0.8524 | 28.32/0.8059 | 29.80/0.8817 | 28.95/0.8430 | 30.52/0.8927 |
Hill | 11.83/0.2107 | 23.90/0.5335 | 24.08/0.5894 | 24.91/0.6182 | 24.65/0.5982 | 25.54/0.6374 |
House | 9.37/0.0813 | 24.51/0.6767 | 24.76/0.7051 | 25.65/0.7388 | 26.30/0.6781 | 27.56/0.7530 |
Lena | 9.92/0.1492 | 23.36/0.6428 | 23.91/0.7232 | 24.52/0.7344 | 24.57/0.6810 | 26.52/0.7863 |
Peppers | 10.70/0.2187 | 22.57/0.6746 | 22.90/0.7158 | 23.78/0.7554 | 24.62/0.7324 | 25.88/0.7977 |
Policemen | 11.52/0.2917 | 21.29/0.6291 | 22.01/0.7199 | 22.62/0.7486 | 22.58/0.7318 | 24.39/0.8188 |
Statues | 12.15/0.3132 | 21.79/0.4702 | 22.11/0.6057 | 22.84/0.6073 | 22.11/0.5667 | 23.46/0.6569 |
Average | 10.92/0.1616 | 23.04/0.5371 | 23.23/0.5789 | 24.05/0.6044 | 23.97/0.5753 | 25.44/0.6398 |
Remote1 | 14.86/0.3045 | 23.81/0.6245 | 23.63/0.6777 | 24.94/0.7286 | 23.68/0.6698 | 25.81/0.7702 |
Remote2 | 16.06/0.2884 | 26.00/0.7058 | 25.86/0.7528 | 27.42/0.8020 | 26.02/0.7537 | 28.13/0.8240 |
Remote3 | 14.68/0.3837 | 23.42/0.6331 | 23.56/0.7425 | 24.93/0.7627 | 23.63/0.7103 | 25.84/0.8089 |
Remote4 | 14.42/0.3371 | 23.89/0.6455 | 23.88/0.7225 | 25.33/0.7546 | 24.24/0.7141 | 26.36/0.8014 |
Remote5 | 13.27/0.2206 | 24.57/0.6308 | 24.33/0.6582 | 25.46/0.6967 | 24.21/0.6326 | 26.25/0.7308 |
Average | 14.66/0.3069 | 24.34/0.6479 | 24.25/0.7107 | 25.62/0.7489 | 24.36/0.6961 | 26.48/0.7870 |
* Bold values indicate the best denoising performance. |
This paper proposes a multi-color channel-based GSR model for the removal of multiplicative gamma noise in color images. The proposed GSR comprises an image group with similar color patches, by measuring the distance between two color patches. This prevented artificial color values produced by separate color channel-based approaches. Furthermore, GSR, which exploits the sparsity and nonlocal similarity of patches, aided in the conservation of fine structures and repeated textures. Additionally, it adequately denoised smooth regions without the staircase artifacts caused by TV-based models, which led to the natural visual quality of restored images. The model also includes a convex data-fidelity term and SVTV regularization. The data-fidelity term facilitated the removal of heavy multiplicative noise, while SVTV helped alleviate some ringing or block artifacts induced by patches, thus enhancing the denoising results. We adopted the ADMM to solve the proposed nonconvex model and its subproblem. These approaches yielded an efficacious iterative algorithm. The experimental results validated the superior denoising performance of the proposed model over other existing models, when it comes to both visual inspection and image quality measurements. However, despite the numerical convergence of the proposed algorithm, the theoretical convergence of the proposed algorithm still remains an issue. Furthermore, the GSR approach requires more computational time than local regularization-based methods due to searching for similar patches and the singular value decomposition of groups. Thus, faster implementation also needs to be investigated in future work.
The author declares they have not used Artificial Intelligence (AI) tools in the creation of this article.
The author was supported by the Hankuk University of Foreign Studies Research Fund and the National Research Foundation of Korea (2021R1F1A1048111).
The author declares no conflicts of interest in this paper.
[1] | P. L. Duren, Theory of Hp spaces, New York: Academic Press, 1970. |
[2] | M. D. Contreras, A. G. Hernández-Díaz, Weighted composition operators on spaces with derivative in a Hardy space, J. Operator Theory, 52 (2004), 173–184. |
[3] |
N. Hu, Weighted composition operators from derivative Hardy spaces into n-th weighted-type spaces, J. Math., 2021 (2021), 4398397. http://dx.doi.org/10.1155/2021/4398397 doi: 10.1155/2021/4398397
![]() |
[4] | B. D. MacCluer, Composition operators on Sp, Houston J. Math., 13 (1987), 245–254. |
[5] | R. C. Roan, Composition operators on the space of functions with Hp-derivative, Houston J. Math., 4 (1978), 423–438. |
[6] |
E. Abbasi, X. Zhu, Product-type operators from the Bloch space into Zygmund-type spaces, Bull. Iran. Math. Soc., 48 (2022), 385–400. http://dx.doi.org/10.1007/s41980-020-00523-1 doi: 10.1007/s41980-020-00523-1
![]() |
[7] |
K. Esmaeili, M. Lindström, Weighted composition operators between Zygmund type spaces and their essential norms, Integr. Equ. Oper. Theory, 75 (2013), 473–490. http://dx.doi.org/10.1007/s00020-013-2038-4 doi: 10.1007/s00020-013-2038-4
![]() |
[8] |
Z. Guo, L. Liu, Product-type operators from Hardy spaces to Bloch-type spaces and Zygmund-type spaces, Numer. Funct. Anal. Optim., 43 (2022), 1240–1264. http://dx.doi.org/10.1080/01630563.2022.2097693 doi: 10.1080/01630563.2022.2097693
![]() |
[9] |
Z. Guo, L. Liu, Y. Shu, On Stević-Sharma operator from the mixed norm spaces to Zygmund-type spaces, Math. Inequal. Appl., 24 (2021), 445–461. http://dx.doi.org/10.7153/mia-2021-24-31 doi: 10.7153/mia-2021-24-31
![]() |
[10] |
S. Li, S. Stević, Volterra-type operators on Zygmund spaces, J. Inequal. Appl., 2007 (2007), 1–10. http://dx.doi.org/10.1155/2007/32124 doi: 10.1155/2007/32124
![]() |
[11] |
S. Li, S. Stević, Generalized composition operators on Zygmund spaces and Bloch type spaces, J. Math. Anal. Appl., 338 (2008), 1282–1295. http://dx.doi.org/10.1016/j.jmaa.2007.06.013 doi: 10.1016/j.jmaa.2007.06.013
![]() |
[12] |
S. Li, S. Stević, Products of Volterra type operator and composition operator from H∞ and Bloch spaces to Zygmund spaces, J. Math. Anal. Appl., 345 (2008), 40–52. http://dx.doi.org/10.1016/j.jmaa.2008.03.063 doi: 10.1016/j.jmaa.2008.03.063
![]() |
[13] |
S. Li, S. Stević, Weighted composition operators from Zygmund spaces into Bloch spaces, Appl. Math. Comput., 206 (2008), 825–831. http://dx.doi.org/10.1016/j.amc.2008.10.006 doi: 10.1016/j.amc.2008.10.006
![]() |
[14] |
S. Li, S. Stević, Products of composition and differentiation operators from Zygmund spaces to Bloch spaces and Bers spaces, Appl. Math. Comput., 217 (2010), 3144–3154. http://dx.doi.org/10.1016/j.amc.2010.08.047 doi: 10.1016/j.amc.2010.08.047
![]() |
[15] |
F. Zhang, Y. Liu, On a Stević-Sharma operator from Hardy spaces to Zygmund-type spaces on the unit disk, Complex Anal. Oper. Theory., 12 (2018), 81–100. http://dx.doi.org/10.1007/s11785-016-0578-8 doi: 10.1007/s11785-016-0578-8
![]() |
[16] |
X. Zhu, E. Abbasi, A. Ebrahimi, Product-type operators on the Zygmund space, Iran. J. Sci. Technol. Trans. Sci., 45 (2021), 1689–1697. http://dx.doi.org/10.1007/s40995-021-01138-9 doi: 10.1007/s40995-021-01138-9
![]() |
[17] | C. C. Cowen, B. D. MacCluer, Composition operators on spaces of analytic functions, Boca Raton: CRC Press, 1995. |
[18] |
R. A. Hibschweiler, N. Portnoy, Composition followed by differentiation between Bergman and Hardy spaces, Rocky Mountain J. Math., 35 (2005), 843–855. http://dx.doi.org/10.1216/rmjm/1181069709 doi: 10.1216/rmjm/1181069709
![]() |
[19] |
S. Li, S. Stević, Composition followed by differentiation from mixed-norm spaces to α-Bloch spaces, Sbornik: Math., 199 (2008), 1847–1857. http://dx.doi.org/10.1070/SM2008v199n12ABEH003983 doi: 10.1070/SM2008v199n12ABEH003983
![]() |
[20] |
S. Ohno, Products of differentiation and composition on Bloch spaces, Bull. Korean Math. Soc., 46 (2009), 1135–1140. http://dx.doi.org/10.4134/BKMS.2009.46.6.1135 doi: 10.4134/BKMS.2009.46.6.1135
![]() |
[21] |
S. Stević, Norm and essential norm of composition followed by differentiation from α-Bloch spaces to H∞μ, Appl. Math. Comput., 207 (2009), 225–229. http://dx.doi.org/10.1016/j.amc.2008.10.032 doi: 10.1016/j.amc.2008.10.032
![]() |
[22] |
S. Stević, A. K. Sharma, A. Bhat, Essential norm of products of multiplication composition and differentiation operators on weighted Bergman spaces, Appl. Math. Comput., 218 (2011), 2386–2397. http://dx.doi.org/10.1016/j.amc.2011.06.055 doi: 10.1016/j.amc.2011.06.055
![]() |
[23] |
S. Stević, A. K. Sharma, A. Bhat, Products of multiplication composition and differentiation operators on weighted Bergman space, Appl. Math. Comput., 217 (2011), 8115–8125. http://dx.doi.org/10.1016/j.amc.2011.03.014 doi: 10.1016/j.amc.2011.03.014
![]() |
[24] |
Y. Liu, Y. Yu, On a Stević-Sharma operator from Hardy spaces to the logarithmic Bloch spaces, J. Inequal. Appl., 22 (2015), 1–19. http://dx.doi.org/10.1186/s13660-015-0547-1 doi: 10.1186/s13660-015-0547-1
![]() |
[25] |
S. Wang, M. Wang, X. Guo, Differences of Stević-Sharma operators, Banach J. Math. Anal., 14 (2020), 1019–1054. http://dx.doi.org/10.1007/s43037-019-00051-z doi: 10.1007/s43037-019-00051-z
![]() |
[26] |
M. S. Al Ghafri, J. S. Manhas, On Stević-Sharma operators from weighted Bergman spaces to weighted-type spaces, Math. Inequal. Appl., 23 (2020), 1051–1077. http://dx.doi.org/10.7153/mia-2020-23-81 doi: 10.7153/mia-2020-23-81
![]() |
[27] |
Z. Guo, On Stević-Sharma operator from weighted Bergman-Orlicz spaces to Bloch-type spaces, Math. Inequal. Appl., 25 (2022), 91–107. http://dx.doi.org/10.7153/mia-2022-25-07 doi: 10.7153/mia-2022-25-07
![]() |
[28] |
Z. Guo, Y. Shu, On Stević-Sharma operators from Hardy spaces to Stević weighted spaces, Math. Inequal. Appl., 23 (2020), 217–229. http://dx.doi.org/10.7153/mia-2020-23-17 doi: 10.7153/mia-2020-23-17
![]() |
[29] |
X. Zhu, E. Abbasi, A. Ebrahimi, A class of operator-related composition operators from the Besov spaces into the Bloch space, Bull. Iran. Math. Soc., 47 (2021), 171–184. http://dx.doi.org/10.1007/s41980-020-00374-w doi: 10.1007/s41980-020-00374-w
![]() |
[30] |
E. Abbasi, Y. Liu, M. Hassanlou, Generalized Stević-Sharma type operators from Hardy spaces into nth weighted type spaces, Turkish J. Math., 45 (2021), 1543–1554. http://dx.doi.org/10.3906/mat-2011-67 doi: 10.3906/mat-2011-67
![]() |
[31] |
S. Stević, Composition operators from the weighted Bergman space to the nth weighted spaces on the unit disc, Discrete Dyn. Nat. Soc., 2009 (2009), 1–12. http://dx.doi.org/10.1155/2009/742019 doi: 10.1155/2009/742019
![]() |
[32] |
S. Stević, Weighted differentiation composition operators from the mixed-norm space to the nth weighted-type space on the unit disk, Abstr. Appl. Anal., 2010 (2010), 1–16. http://dx.doi.org/10.1155/2010/246287 doi: 10.1155/2010/246287
![]() |
[33] |
S. Stević, A. K. Sharma, R. Krishan, Boundedness and compactness of a new product-type operator from a general space to Bloch-type spaces, J. Inequal. Appl., 2016 (2016), 1–32. http://dx.doi.org/10.1186/s13660-016-1159-0 doi: 10.1186/s13660-016-1159-0
![]() |
[34] |
E. Abbasi, The product-type operators from Hardy spaces into nth weighted-type spaces, Abstr. Appl. Anal., 2021 (2021), 1–8. http://dx.doi.org/10.1155/2021/5556275 doi: 10.1155/2021/5556275
![]() |
[35] |
Z. Guo, X. Zhao, On a Stević-Sharma type operator from Qk(p,q) spaces to Bloch-type spaces, Oper. Matrices, 16 (2022), 563–580. http://dx.doi.org/10.7153/oam-2022-16-43 doi: 10.7153/oam-2022-16-43
![]() |
[36] |
S. Stević, Composition operators from the Hardy space to the nth weighted-type space on the unit disk and the half-plane, Appl. Math. Comput., 215 (2010), 3950–3955. http://dx.doi.org/10.1016/j.amc.2009.11.043 doi: 10.1016/j.amc.2009.11.043
![]() |
[37] |
F. Colonna, M. Tjani, Weighted composition operators from Banach spaces of analytic functions into Bloch-type spaces, Probl. Recent Methods Oper. Theory, 687 (2017), 75–95. http://dx.doi.org/10.1090/conm/687/13790 doi: 10.1090/conm/687/13790
![]() |
[38] |
P. Galindo, M. Lindström, S. Stević, Essential norm of operators into weighted-type spaces on the unit ball, Abstr. Appl. Anal., 2011 (2011), 1–14. http://dx.doi.org/10.1155/2011/939873 doi: 10.1155/2011/939873
![]() |
[39] |
S. Li, S. Stević, Generalized weighted composition operators from α-Bloch spaces into weighted-type spaces, J. Inequal. Appl., 2015 (2015), 1–12. http://dx.doi.org/10.1186/s13660-015-0770-9 doi: 10.1186/s13660-015-0770-9
![]() |
Image | Data f | L2-SVTV | AA-SVTV | SO-SVNTVTV2 | SAR-BM3D | Proposed |
Airplane | 12.69/0.1970 | 23.93/0.7176 | 24.83/0.7965 | 25.06/0.7985 | 25.34/0.7419 | 26.83/0.8246 |
Baboon | 15.30/0.4449 | 21.55/0.5299 | 22.39/0.6325 | 22.47/0.6135 | 22.83/0.6569 | 23.54/0.6702 |
Barbara | 16.00/0.3362 | 25.16/0.6897 | 26.25/0.7587 | 26.48/0.7784 | 27.66/0.8110 | 29.41/0.8699 |
Bears | 18.12/0.3207 | 27.83/0.6846 | 28.38/0.7433 | 28.85/0.7496 | 28.18/0.7242 | 29.34/0.7690 |
Boat | 15.16/0.3794 | 24.42/0.7217 | 26.16/0.8042 | 26.44/0.8047 | 25.80/0.7766 | 27.52/0.8279 |
Building | 17.30/0.3882 | 25.76/0.7974 | 25.55/0.7840 | 26.68/0.8357 | 26.88/0.8337 | 28.68/0.8737 |
Butterfly | 16.55/0.4424 | 23.24/0.8235 | 24.07/0.8508 | 25.31/0.8841 | 24.09/0.8558 | 26.85/0.9149 |
Castle | 15.74/0.2609 | 25.85/0.7728 | 26.89/0.7946 | 27.50/0.8421 | 27.04/0.8179 | 28.72/0.8665 |
Cathedral | 15.37/0.2984 | 23.69/0.7055 | 24.74/0.7798 | 25.21/0.7744 | 25.16/0.7547 | 26.79/0.7999 |
Flower | 19.31/0.3042 | 30.82/0.8958 | 30.92/0.8462 | 32.12/0.9150 | 31.53/0.9004 | 32.85/0.9253 |
Hill | 17.06/0.4076 | 25.64/0.6292 | 26.83/0.7389 | 27.05/0.7323 | 26.96/0.7249 | 27.62/0.7401 |
House | 14.62/0.1839 | 26.97/0.7481 | 28.03/0.7816 | 28.41/0.7904 | 29.76/0.7776 | 30.64/0.8087 |
Lena | 15.10/0.2795 | 25.09/0.7358 | 26.96/0.8072 | 27.05/0.8109 | 27.41/0.7968 | 29.26/0.8606 |
Peppers | 15.88/0.3609 | 25.04/0.7566 | 25.75/0.8066 | 26.51/0.8182 | 27.44/0.8176 | 28.64/0.8526 |
Policemen | 16.76/0.4615 | 23.41/0.7372 | 24.92/0.8372 | 25.70/0.8576 | 25.09/0.8469 | 27.48/0.9086 |
Statues | 17.39/0.5465 | 23.84/0.6094 | 24.88/0.7734 | 25.23/0.7525 | 24.18/0.7156 | 26.10/0.7933 |
Average | 16.15/0.3139 | 25.14/0.6209 | 26.10/0.6813 | 26.63/0.6895 | 26.58/0.6789 | 28.14/0.7202 |
Remote1 | 20.05/0.5335 | 26.79/0.7756 | 27.03/0.8134 | 27.83/0.8411 | 26.19/0.8004 | 28.81/0.8689 |
Remote2 | 21.31/0.5074 | 28.81/0.8124 | 29.07/0.8497 | 30.23/0.8822 | 28.47/0.8481 | 31.00/0.8984 |
Remote3 | 19.89/0.6262 | 26.55/0.7933 | 27.06/0.8624 | 27.92/0.8690 | 26.05/0.8295 | 28.83/0.8949 |
Remote4 | 19.68/0.5646 | 26.47/0.7679 | 27.28/0.8453 | 28.32/0.8583 | 26.80/0.8252 | 29.31/0.8797 |
Remote5 | 18.50/0.4210 | 26.98/0.7480 | 27.46/0.7952 | 28.04/0.8052 | 26.73/0.7587 | 28.82/0.8255 |
Average | 19.87/0.5305 | 27.12/0.7794 | 27.58/0.8332 | 28.47/0.8512 | 26.85/0.8124 | 29.35/0.8735 |
* Bold values indicate the best denoising performance. |
Image | Data f | L2-SVTV | AA-SVTV | SO-SVNTVTV2 | SAR-BM3D | Proposed |
Airplane | 9.66/0.1375 | 22.54/0.6631 | 23.35/0.7415 | 23.59/0.7518 | 23.70/0.6709 | 25.10/0.7781 |
Baboon | 12.34/0.3214 | 20.69/0.4526 | 21.15/0.5497 | 21.18/0.5063 | 21.46/0.5419 | 22.48/0.6147 |
Barbara | 12.96/0.2297 | 24.27/0.6355 | 24.64/0.6809 | 24.81/0.6954 | 25.97/0.7455 | 27.66/0.8271 |
Bears | 15.12/0.2077 | 26.71/0.6307 | 26.74/0.6634 | 27.54/0.6836 | 26.92/0.6572 | 27.98/0.7070 |
Boat | 12.13/0.3016 | 23.08/0.6592 | 24.53/0.7570 | 24.73/0.7573 | 24.35/0.7196 | 25.98/0.7875 |
Building | 14.33/0.2730 | 23.89/0.7217 | 23.78/0.7022 | 24.92/0.7786 | 25.14/0.7756 | 27.02/0.8386 |
Butterfly | 13.61/0.3445 | 21.47/0.7615 | 22.02/0.7914 | 23.42/0.8385 | 22.63/0.8071 | 24.83/0.8792 |
Castle | 12.73/0.1732 | 24.55/0.7319 | 25.12/0.7526 | 25.93/0.8017 | 25.49/0.7594 | 27.19/0.8347 |
Cathedral | 12.37/0.2070 | 22.74/0.6591 | 23.06/0.7142 | 23.48/0.7161 | 23.56/0.6864 | 25.19/0.7587 |
Flower | 16.31/0.2081 | 29.17/0.8699 | 29.46/0.8416 | 30.76/0.8954 | 30.10/0.8699 | 31.50/0.9085 |
Hill | 14.04/0.2879 | 24.43/0.5644 | 25.31/0.6597 | 25.76/0.6658 | 25.63/0.6557 | 26.67/0.7059 |
House | 11.60/0.1177 | 25.48/0.7153 | 26.28/0.7451 | 26.89/0.7645 | 27.85/0.7279 | 28.82/0.7811 |
Lena | 12.14/0.1993 | 24.17/0.6917 | 25.33/0.7709 | 25.62/0.7711 | 25.84/0.7363 | 27.67/0.8234 |
Peppers | 12.84/0.2734 | 23.70/0.7107 | 24.28/0.7513 | 24.66/0.7839 | 25.83/0.7727 | 26.98/0.8239 |
Policemen | 13.72/0.3619 | 21.82/0.6658 | 23.09/0.7760 | 23.78/0.7982 | 23.60/0.7856 | 25.60/0.8615 |
Statues | 14.35/0.4083 | 22.54/0.5237 | 23.24/0.6823 | 23.55/0.6469 | 22.92/0.6234 | 24.52/0.7188 |
Average | 13.14/0.2147 | 23.83/0.5733 | 24.46/0.6304 | 25.04/0.6399 | 25.06/0.6220 | 26.57/0.6832 |
Remote1 | 17.05/0.3980 | 24.92/0.6857 | 25.04/0.7425 | 26.11/0.7800 | 24.71/0.7295 | 26.99/0.8147 |
Remote2 | 18.29/0.3773 | 27.05/0.7483 | 27.14/0.8030 | 28.50/0.8380 | 27.02/0.7956 | 29.27/0.8577 |
Remote3 | 16.89/0.4868 | 24.60/0.7005 | 25.00/0.7987 | 26.11/0.8114 | 24.64/0.7644 | 27.06/0.8507 |
Remote4 | 16.67/0.4314 | 25.03/0.7070 | 25.25/0.7809 | 26.52/0.8024 | 25.32/0.7649 | 27.60/0.8397 |
Remote5 | 15.49/0.2986 | 25.66/0.6906 | 25.66/0.7244 | 26.51/0.7471 | 25.29/0.6906 | 27.25/0.7688 |
Average | 16.88/0.3984 | 25.45/0.7064 | 25.62/0.7699 | 26.75/0.7958 | 25.40/0.7490 | 27.63/0.8263 |
* Bold values indicate the best denoising performance. |
Image | Data f | L2-SVTV | AA-SVTV | SO-SVNTVTV2 | SAR-BM3D | Proposed |
Airplane | 7.41/0.1020 | 21.86/0.5995 | 22.11/0.6777 | 22.73/0.7236 | 22.71/0.6098 | 24.08/0.7308 |
Baboon | 10.12/0.2379 | 20.06/0.4003 | 20.03/0.4381 | 20.33/0.4231 | 20.60/0.4604 | 21.13/0.4823 |
Barbara | 10.80/0.1654 | 23.51/0.5944 | 23.44/0.6154 | 24.08/0.6503 | 24.75/0.6823 | 26.55/0.7818 |
Bears | 12.87/0.1438 | 26.02/0.5990 | 25.72/0.6054 | 26.72/0.6331 | 26.05/0.6092 | 27.06/0.6449 |
Boat | 9.91/0.2481 | 22.37/0.6059 | 22.99/0.7008 | 23.72/0.7197 | 23.39/0.6704 | 24.95/0.7456 |
Building | 12.07/0.1987 | 22.91/0.6753 | 22.62/0.6418 | 23.65/0.7241 | 23.86/0.7200 | 25.86/0.7960 |
Butterfly | 11.39/0.2756 | 20.46/0.7176 | 20.76/0.7504 | 22.17/0.8021 | 21.51/0.7607 | 23.49/0.8444 |
Castle | 10.46/0.1229 | 23.81/0.6931 | 23.84/0.7015 | 24.69/0.7628 | 24.25/0.7081 | 25.94/0.7863 |
Cathedral | 10.12/0.1505 | 21.88/0.6228 | 22.02/0.6584 | 22.52/0.6740 | 22.54/0.6341 | 24.10/0.7122 |
Flower | 14.10/0.1508 | 28.35/0.8524 | 28.32/0.8059 | 29.80/0.8817 | 28.95/0.8430 | 30.52/0.8927 |
Hill | 11.83/0.2107 | 23.90/0.5335 | 24.08/0.5894 | 24.91/0.6182 | 24.65/0.5982 | 25.54/0.6374 |
House | 9.37/0.0813 | 24.51/0.6767 | 24.76/0.7051 | 25.65/0.7388 | 26.30/0.6781 | 27.56/0.7530 |
Lena | 9.92/0.1492 | 23.36/0.6428 | 23.91/0.7232 | 24.52/0.7344 | 24.57/0.6810 | 26.52/0.7863 |
Peppers | 10.70/0.2187 | 22.57/0.6746 | 22.90/0.7158 | 23.78/0.7554 | 24.62/0.7324 | 25.88/0.7977 |
Policemen | 11.52/0.2917 | 21.29/0.6291 | 22.01/0.7199 | 22.62/0.7486 | 22.58/0.7318 | 24.39/0.8188 |
Statues | 12.15/0.3132 | 21.79/0.4702 | 22.11/0.6057 | 22.84/0.6073 | 22.11/0.5667 | 23.46/0.6569 |
Average | 10.92/0.1616 | 23.04/0.5371 | 23.23/0.5789 | 24.05/0.6044 | 23.97/0.5753 | 25.44/0.6398 |
Remote1 | 14.86/0.3045 | 23.81/0.6245 | 23.63/0.6777 | 24.94/0.7286 | 23.68/0.6698 | 25.81/0.7702 |
Remote2 | 16.06/0.2884 | 26.00/0.7058 | 25.86/0.7528 | 27.42/0.8020 | 26.02/0.7537 | 28.13/0.8240 |
Remote3 | 14.68/0.3837 | 23.42/0.6331 | 23.56/0.7425 | 24.93/0.7627 | 23.63/0.7103 | 25.84/0.8089 |
Remote4 | 14.42/0.3371 | 23.89/0.6455 | 23.88/0.7225 | 25.33/0.7546 | 24.24/0.7141 | 26.36/0.8014 |
Remote5 | 13.27/0.2206 | 24.57/0.6308 | 24.33/0.6582 | 25.46/0.6967 | 24.21/0.6326 | 26.25/0.7308 |
Average | 14.66/0.3069 | 24.34/0.6479 | 24.25/0.7107 | 25.62/0.7489 | 24.36/0.6961 | 26.48/0.7870 |
* Bold values indicate the best denoising performance. |
Image | Data f | L2-SVTV | AA-SVTV | SO-SVNTVTV2 | SAR-BM3D | Proposed |
Airplane | 12.69/0.1970 | 23.93/0.7176 | 24.83/0.7965 | 25.06/0.7985 | 25.34/0.7419 | 26.83/0.8246 |
Baboon | 15.30/0.4449 | 21.55/0.5299 | 22.39/0.6325 | 22.47/0.6135 | 22.83/0.6569 | 23.54/0.6702 |
Barbara | 16.00/0.3362 | 25.16/0.6897 | 26.25/0.7587 | 26.48/0.7784 | 27.66/0.8110 | 29.41/0.8699 |
Bears | 18.12/0.3207 | 27.83/0.6846 | 28.38/0.7433 | 28.85/0.7496 | 28.18/0.7242 | 29.34/0.7690 |
Boat | 15.16/0.3794 | 24.42/0.7217 | 26.16/0.8042 | 26.44/0.8047 | 25.80/0.7766 | 27.52/0.8279 |
Building | 17.30/0.3882 | 25.76/0.7974 | 25.55/0.7840 | 26.68/0.8357 | 26.88/0.8337 | 28.68/0.8737 |
Butterfly | 16.55/0.4424 | 23.24/0.8235 | 24.07/0.8508 | 25.31/0.8841 | 24.09/0.8558 | 26.85/0.9149 |
Castle | 15.74/0.2609 | 25.85/0.7728 | 26.89/0.7946 | 27.50/0.8421 | 27.04/0.8179 | 28.72/0.8665 |
Cathedral | 15.37/0.2984 | 23.69/0.7055 | 24.74/0.7798 | 25.21/0.7744 | 25.16/0.7547 | 26.79/0.7999 |
Flower | 19.31/0.3042 | 30.82/0.8958 | 30.92/0.8462 | 32.12/0.9150 | 31.53/0.9004 | 32.85/0.9253 |
Hill | 17.06/0.4076 | 25.64/0.6292 | 26.83/0.7389 | 27.05/0.7323 | 26.96/0.7249 | 27.62/0.7401 |
House | 14.62/0.1839 | 26.97/0.7481 | 28.03/0.7816 | 28.41/0.7904 | 29.76/0.7776 | 30.64/0.8087 |
Lena | 15.10/0.2795 | 25.09/0.7358 | 26.96/0.8072 | 27.05/0.8109 | 27.41/0.7968 | 29.26/0.8606 |
Peppers | 15.88/0.3609 | 25.04/0.7566 | 25.75/0.8066 | 26.51/0.8182 | 27.44/0.8176 | 28.64/0.8526 |
Policemen | 16.76/0.4615 | 23.41/0.7372 | 24.92/0.8372 | 25.70/0.8576 | 25.09/0.8469 | 27.48/0.9086 |
Statues | 17.39/0.5465 | 23.84/0.6094 | 24.88/0.7734 | 25.23/0.7525 | 24.18/0.7156 | 26.10/0.7933 |
Average | 16.15/0.3139 | 25.14/0.6209 | 26.10/0.6813 | 26.63/0.6895 | 26.58/0.6789 | 28.14/0.7202 |
Remote1 | 20.05/0.5335 | 26.79/0.7756 | 27.03/0.8134 | 27.83/0.8411 | 26.19/0.8004 | 28.81/0.8689 |
Remote2 | 21.31/0.5074 | 28.81/0.8124 | 29.07/0.8497 | 30.23/0.8822 | 28.47/0.8481 | 31.00/0.8984 |
Remote3 | 19.89/0.6262 | 26.55/0.7933 | 27.06/0.8624 | 27.92/0.8690 | 26.05/0.8295 | 28.83/0.8949 |
Remote4 | 19.68/0.5646 | 26.47/0.7679 | 27.28/0.8453 | 28.32/0.8583 | 26.80/0.8252 | 29.31/0.8797 |
Remote5 | 18.50/0.4210 | 26.98/0.7480 | 27.46/0.7952 | 28.04/0.8052 | 26.73/0.7587 | 28.82/0.8255 |
Average | 19.87/0.5305 | 27.12/0.7794 | 27.58/0.8332 | 28.47/0.8512 | 26.85/0.8124 | 29.35/0.8735 |
* Bold values indicate the best denoising performance. |
Image | Data f | L2-SVTV | AA-SVTV | SO-SVNTVTV2 | SAR-BM3D | Proposed |
Airplane | 9.66/0.1375 | 22.54/0.6631 | 23.35/0.7415 | 23.59/0.7518 | 23.70/0.6709 | 25.10/0.7781 |
Baboon | 12.34/0.3214 | 20.69/0.4526 | 21.15/0.5497 | 21.18/0.5063 | 21.46/0.5419 | 22.48/0.6147 |
Barbara | 12.96/0.2297 | 24.27/0.6355 | 24.64/0.6809 | 24.81/0.6954 | 25.97/0.7455 | 27.66/0.8271 |
Bears | 15.12/0.2077 | 26.71/0.6307 | 26.74/0.6634 | 27.54/0.6836 | 26.92/0.6572 | 27.98/0.7070 |
Boat | 12.13/0.3016 | 23.08/0.6592 | 24.53/0.7570 | 24.73/0.7573 | 24.35/0.7196 | 25.98/0.7875 |
Building | 14.33/0.2730 | 23.89/0.7217 | 23.78/0.7022 | 24.92/0.7786 | 25.14/0.7756 | 27.02/0.8386 |
Butterfly | 13.61/0.3445 | 21.47/0.7615 | 22.02/0.7914 | 23.42/0.8385 | 22.63/0.8071 | 24.83/0.8792 |
Castle | 12.73/0.1732 | 24.55/0.7319 | 25.12/0.7526 | 25.93/0.8017 | 25.49/0.7594 | 27.19/0.8347 |
Cathedral | 12.37/0.2070 | 22.74/0.6591 | 23.06/0.7142 | 23.48/0.7161 | 23.56/0.6864 | 25.19/0.7587 |
Flower | 16.31/0.2081 | 29.17/0.8699 | 29.46/0.8416 | 30.76/0.8954 | 30.10/0.8699 | 31.50/0.9085 |
Hill | 14.04/0.2879 | 24.43/0.5644 | 25.31/0.6597 | 25.76/0.6658 | 25.63/0.6557 | 26.67/0.7059 |
House | 11.60/0.1177 | 25.48/0.7153 | 26.28/0.7451 | 26.89/0.7645 | 27.85/0.7279 | 28.82/0.7811 |
Lena | 12.14/0.1993 | 24.17/0.6917 | 25.33/0.7709 | 25.62/0.7711 | 25.84/0.7363 | 27.67/0.8234 |
Peppers | 12.84/0.2734 | 23.70/0.7107 | 24.28/0.7513 | 24.66/0.7839 | 25.83/0.7727 | 26.98/0.8239 |
Policemen | 13.72/0.3619 | 21.82/0.6658 | 23.09/0.7760 | 23.78/0.7982 | 23.60/0.7856 | 25.60/0.8615 |
Statues | 14.35/0.4083 | 22.54/0.5237 | 23.24/0.6823 | 23.55/0.6469 | 22.92/0.6234 | 24.52/0.7188 |
Average | 13.14/0.2147 | 23.83/0.5733 | 24.46/0.6304 | 25.04/0.6399 | 25.06/0.6220 | 26.57/0.6832 |
Remote1 | 17.05/0.3980 | 24.92/0.6857 | 25.04/0.7425 | 26.11/0.7800 | 24.71/0.7295 | 26.99/0.8147 |
Remote2 | 18.29/0.3773 | 27.05/0.7483 | 27.14/0.8030 | 28.50/0.8380 | 27.02/0.7956 | 29.27/0.8577 |
Remote3 | 16.89/0.4868 | 24.60/0.7005 | 25.00/0.7987 | 26.11/0.8114 | 24.64/0.7644 | 27.06/0.8507 |
Remote4 | 16.67/0.4314 | 25.03/0.7070 | 25.25/0.7809 | 26.52/0.8024 | 25.32/0.7649 | 27.60/0.8397 |
Remote5 | 15.49/0.2986 | 25.66/0.6906 | 25.66/0.7244 | 26.51/0.7471 | 25.29/0.6906 | 27.25/0.7688 |
Average | 16.88/0.3984 | 25.45/0.7064 | 25.62/0.7699 | 26.75/0.7958 | 25.40/0.7490 | 27.63/0.8263 |
* Bold values indicate the best denoising performance. |
Image | Data f | L2-SVTV | AA-SVTV | SO-SVNTVTV2 | SAR-BM3D | Proposed |
Airplane | 7.41/0.1020 | 21.86/0.5995 | 22.11/0.6777 | 22.73/0.7236 | 22.71/0.6098 | 24.08/0.7308 |
Baboon | 10.12/0.2379 | 20.06/0.4003 | 20.03/0.4381 | 20.33/0.4231 | 20.60/0.4604 | 21.13/0.4823 |
Barbara | 10.80/0.1654 | 23.51/0.5944 | 23.44/0.6154 | 24.08/0.6503 | 24.75/0.6823 | 26.55/0.7818 |
Bears | 12.87/0.1438 | 26.02/0.5990 | 25.72/0.6054 | 26.72/0.6331 | 26.05/0.6092 | 27.06/0.6449 |
Boat | 9.91/0.2481 | 22.37/0.6059 | 22.99/0.7008 | 23.72/0.7197 | 23.39/0.6704 | 24.95/0.7456 |
Building | 12.07/0.1987 | 22.91/0.6753 | 22.62/0.6418 | 23.65/0.7241 | 23.86/0.7200 | 25.86/0.7960 |
Butterfly | 11.39/0.2756 | 20.46/0.7176 | 20.76/0.7504 | 22.17/0.8021 | 21.51/0.7607 | 23.49/0.8444 |
Castle | 10.46/0.1229 | 23.81/0.6931 | 23.84/0.7015 | 24.69/0.7628 | 24.25/0.7081 | 25.94/0.7863 |
Cathedral | 10.12/0.1505 | 21.88/0.6228 | 22.02/0.6584 | 22.52/0.6740 | 22.54/0.6341 | 24.10/0.7122 |
Flower | 14.10/0.1508 | 28.35/0.8524 | 28.32/0.8059 | 29.80/0.8817 | 28.95/0.8430 | 30.52/0.8927 |
Hill | 11.83/0.2107 | 23.90/0.5335 | 24.08/0.5894 | 24.91/0.6182 | 24.65/0.5982 | 25.54/0.6374 |
House | 9.37/0.0813 | 24.51/0.6767 | 24.76/0.7051 | 25.65/0.7388 | 26.30/0.6781 | 27.56/0.7530 |
Lena | 9.92/0.1492 | 23.36/0.6428 | 23.91/0.7232 | 24.52/0.7344 | 24.57/0.6810 | 26.52/0.7863 |
Peppers | 10.70/0.2187 | 22.57/0.6746 | 22.90/0.7158 | 23.78/0.7554 | 24.62/0.7324 | 25.88/0.7977 |
Policemen | 11.52/0.2917 | 21.29/0.6291 | 22.01/0.7199 | 22.62/0.7486 | 22.58/0.7318 | 24.39/0.8188 |
Statues | 12.15/0.3132 | 21.79/0.4702 | 22.11/0.6057 | 22.84/0.6073 | 22.11/0.5667 | 23.46/0.6569 |
Average | 10.92/0.1616 | 23.04/0.5371 | 23.23/0.5789 | 24.05/0.6044 | 23.97/0.5753 | 25.44/0.6398 |
Remote1 | 14.86/0.3045 | 23.81/0.6245 | 23.63/0.6777 | 24.94/0.7286 | 23.68/0.6698 | 25.81/0.7702 |
Remote2 | 16.06/0.2884 | 26.00/0.7058 | 25.86/0.7528 | 27.42/0.8020 | 26.02/0.7537 | 28.13/0.8240 |
Remote3 | 14.68/0.3837 | 23.42/0.6331 | 23.56/0.7425 | 24.93/0.7627 | 23.63/0.7103 | 25.84/0.8089 |
Remote4 | 14.42/0.3371 | 23.89/0.6455 | 23.88/0.7225 | 25.33/0.7546 | 24.24/0.7141 | 26.36/0.8014 |
Remote5 | 13.27/0.2206 | 24.57/0.6308 | 24.33/0.6582 | 25.46/0.6967 | 24.21/0.6326 | 26.25/0.7308 |
Average | 14.66/0.3069 | 24.34/0.6479 | 24.25/0.7107 | 25.62/0.7489 | 24.36/0.6961 | 26.48/0.7870 |
* Bold values indicate the best denoising performance. |