Loading [MathJax]/jax/output/SVG/jax.js
Research article

An MTL1TV non-convex regularization model for MR Image reconstruction using the alternating direction method of multipliers

  • Received: 20 November 2023 Revised: 24 February 2024 Accepted: 09 May 2024 Published: 28 May 2024
  • The acquisition time of magnetic resonance imaging (MRI) is relatively long. To achieve high-quality and fast reconstruction of magnetic resonance (MR) images, we proposed a non-convex regularization model for MR image reconstruction with the modified transformed l1 total variation (MTL1TV) regularization term. We addressed this new model using the alternating direction method of multipliers (ADMM). To evaluate the proposed MTL1TV model, we performed numerical experiments on several MR images. The numerical results showed that the proposed model gives reconstructed images of improved quality compared with those obtained from state of the art models. The results indicated that the proposed model can effectively reconstruct MR images.

    Citation: Xuexiao You, Ning Cao, Wei Wang. An MTL1TV non-convex regularization model for MR Image reconstruction using the alternating direction method of multipliers[J]. Electronic Research Archive, 2024, 32(5): 3433-3456. doi: 10.3934/era.2024159

    Related Papers:

    [1] Jye Ying Sia, Yong Kheng Goh, How Hui Liew, Yun Fah Chang . Constructing hidden differential equations using a data-driven approach with the alternating direction method of multipliers (ADMM). Electronic Research Archive, 2025, 33(2): 890-906. doi: 10.3934/era.2025040
    [2] Talaat Abdelhamid, F. Khayat, H. Zayeni, Rongliang Chen . Levenberg-Marquardt method for identifying Young's modulus of the elasticity imaging inverse problem. Electronic Research Archive, 2022, 30(4): 1532-1557. doi: 10.3934/era.2022079
    [3] Ruini Zhao . Nanocrystalline SEM image restoration based on fractional-order TV and nuclear norm. Electronic Research Archive, 2024, 32(8): 4954-4968. doi: 10.3934/era.2024228
    [4] Ranran Li, Hao Liu . On global randomized block Kaczmarz method for image reconstruction. Electronic Research Archive, 2022, 30(4): 1442-1453. doi: 10.3934/era.2022075
    [5] Wenya Shi, Xinpeng Yan, Zhan Huan . Faster free pseudoinverse greedy block Kaczmarz method for image recovery. Electronic Research Archive, 2024, 32(6): 3973-3988. doi: 10.3934/era.2024178
    [6] Jiange Liu, Yu Chen, Xin Dai, Li Cao, Qingwu Li . MFCEN: A lightweight multi-scale feature cooperative enhancement network for single-image super-resolution. Electronic Research Archive, 2024, 32(10): 5783-5803. doi: 10.3934/era.2024267
    [7] Haijun Wang, Wenli Zheng, Yaowei Wang, Tengfei Yang, Kaibing Zhang, Youlin Shang . Single hyperspectral image super-resolution using a progressive upsampling deep prior network. Electronic Research Archive, 2024, 32(7): 4517-4542. doi: 10.3934/era.2024205
    [8] Shuaiqun Wang, Huiqiu Chen, Wei Kong, Xinqi Wu, Yafei Qian, Kai Wei . A modified FGL sparse canonical correlation analysis for the identification of Alzheimer's disease biomarkers. Electronic Research Archive, 2023, 31(2): 882-903. doi: 10.3934/era.2023044
    [9] Zhengyu Liu, Yufei Bao, Changhai Wang, Xiaoxiao Chen, Qing Liu . A fast matrix completion method based on truncated L2,1 norm minimization. Electronic Research Archive, 2024, 32(3): 2099-2119. doi: 10.3934/era.2024095
    [10] Hyungyeong Jung, Sunghwan Moon . Reconstruction of the initial function from the solution of the fractional wave equation measured in two geometric settings. Electronic Research Archive, 2022, 30(12): 4436-4446. doi: 10.3934/era.2022225
  • The acquisition time of magnetic resonance imaging (MRI) is relatively long. To achieve high-quality and fast reconstruction of magnetic resonance (MR) images, we proposed a non-convex regularization model for MR image reconstruction with the modified transformed l1 total variation (MTL1TV) regularization term. We addressed this new model using the alternating direction method of multipliers (ADMM). To evaluate the proposed MTL1TV model, we performed numerical experiments on several MR images. The numerical results showed that the proposed model gives reconstructed images of improved quality compared with those obtained from state of the art models. The results indicated that the proposed model can effectively reconstruct MR images.



    Magnetic resonance imaging (MRI) is an approach broadly used for clinical diagnosis because of its ability to effectively reflect changes in human organs [1,2,3]. However, a long-standing challenge is that the acquisition time of MRI is relatively long. To address this problem, researchers have proposed numerous techniques, such as sparse sampling [4,5] and parallel imaging [6,7]. After the theory of compressed sensing (CS) [8] was proposed, many scholars applied it to recover magnetic resonance (MR) images from highly undersampled measurements [9,10]. Research on MR image compression sensing has thus become increasingly important.

    In the MRI domain, the data acquisition of MRI can be modeled as follows:

    y=Ax+ε (1.1)

    where xRn denotes the desired MRI data, yRm is the observed undersampled k-space MRI data, and the matrix A=RFRm×n(m<n). R represents the undersampling operator, such as the radial sampling operator and Cartesian sampling operator. F is the discrete Fourier transform (DFT) operator, and εRm is the additive Gaussian noise. The objective of MR image reconstruction is to recover x from y. Consequently, the reconstruction problem of (1) can be formulated as

    minxλφ(x)+12yAx22 (1.2)

    in which yAx22 is the data fidelity term, λ>0 is the regularization parameter, and φ(x) is some regularizing functional exploiting image prior knowledge. In CS-MRI, φ(x) is the standard total variation (TV) norm (or l1-based regularization) [11], and the corresponding reconstruction model can be expressed as

    minxλxTV+12yAx22 (1.3)

    where xTV=Dx1 (D is the gradient operator). Although standard TV regularization is convex and widely used, the l1-norm regularized model induces bias for large coefficients [12] and, hence, lacks the oracle property. To overcome the shortcomings of l1-norm regularization, researchers have proposed some non-convex regularization methods based on the capped-L1 (CaP) [13], smoothly clipped absolute deviation (SCAD) [14], minimax concave (MC) penalty [15,16,17,18], arctangent penalty (ATAN) [19], logarithm function (Log) [20], transformed l1 (TL1) [21,22], fraction function penalty [23], and non-convex graph total variation (GTV) regularization [23]. For instance, Liu et al. [24] introduced the minimax concave total variation (MCTV) penalty as a regularization term for MR brain image reconstruction. Further, Luo et al. [25] used the arctangent function as the non-convex TV regularize term (AtanTV) for MR image reconstruction. The same team [26] introduced the non-convex MR image reconstruction model via the SCAD penalty function. Lu et al. [27] also used the non-convex Cauchy total variation (CauchyTV) as the regularization term for MR image reconstruction. These studies have shown that non-convex penalties in image restoration usually have better performance than the l1-norm (or standard TV).

    The contributions of this study are as follows. First, we slightly modify the transformed l1 (TL1) function and propose a modified TL1 penalty function (MTL1). Second, inspired by the previous work, we use the MTL1 as the regularization term (MTL1TV) to construct a non-convex regularization model for MR image reconstruction. Third, the proposed model can be solved by the alternating direction method of multipliers (ADMM) [28,29,30]. Finally, numerical experiments on several MR images showed that compared with the traditional TV and state of the art models, the performance of the MTL1TV model was significantly improved. These results confirm the effectiveness of the proposed model.

    The remainder of this paper is organized as follows. In Section 2, we formally introduce the MTL1 function and present a non-convex MTL1TV model, which is based on the MTL1 function. In Section 3, we utilize the ADMM algorithm to address the proposed non-convex model. The convergence result is provided in Section 4. The experimental results are then presented in Section 5. Finally, we conclude the paper in Section 6.

    In this section, we provide the definition of the the modified TL1 penalty function (MTL1) and give an explicit expression of the proximity operator of MTL1. Then, we use MTL1 to define MTL1TV regularization and structure a non-convex MTL1TV model for MR image reconstruction.

    Definition 1. The modified transformed l1 function (MTL1) is defined as

    ϕa(x)=a|x|a+|x|, (2.1)

    where parameter a>0. It is obtained by slightly modifying the following TL1 penalty function [20]:

    pa(x)=(a+1)|x|a+|x|. (2.2)

    Figure 1 shows that the MTL1 is a good alternative to the l1 norm. The larger the value of parameter a is, the closer the behavior of MTL1 is to the l1 norm.

    Figure 1.  |x| and MTL1 function plots with a={0.2,0.5,1,5,10}.

    The proximal operator of ϕa(x) at tR is defined as

    proxϕλ(t)=argminxR{θλ,t(x):=λϕa(x)+12(xt)2}. (2.3)

    Similar to Theorem Ⅲ.1 of [20], we can derive that for the MTL1 function, there exists a closed-formed representation of the optimal solution to (2.3).

    Theorem 1. The optimal solution x=argminθλ,t(x) is a threshold function defined as

    x={0,|t|δλ;gλ(t),|t|>δλ. (2.4)

    where

    gλ(t)=sgn(t)(23(a+|t|)cos(φ(t)3)2a3+|t|3), (2.5)

    with φ(t)=arccos(127λa22(a+|t|)3), and the threshold value δλ satisfies

    δλ={λλa2;2λaa2λ>a2.

    To prove Theorem 1, we first give the following lemmas.

    Lemma 1. Define three parameters

    t1=327λa24a,t2=λ,t3=2λaa2

    for any parameters λ>0 and a>0, then the inequalities t1t3t2 hold. Furthermore, they are equal to a2 when λ=a2.

    Lemma 2. For any given t, the roots of the following two cubic polynomials of x satisfy the following properties:

    1) If t>t1, then the cubic polynomial

    x(a+x)2t(a+x)2+λa2 (2.6)

    has three distinct real roots, and the largest root x0 is given by x0=gλ(t), |gλ(t)||t|.

    2) If t<t1, then the cubic polynomial

    x(ax)2t(ax)2λa2 (2.7)

    has also three distinct real roots, and the smallest root x0 is given by x0=gλ(t).

    The proof of Lemma 2 is essentially the same as Lemma Ⅲ.1 in [20] and Lemma 8 in [22], with only minor changes required due to the different values of t1, t2, and t3 and different cubic polynomials. The proof of Theorem 1 is similar to Theorem Ⅲ.1 in [20] and Lemma 9 in [22]. The detailed proofs of Lemma 2 and Theorem 1 can be found in Appendixes A and B, respectively. Based on the Theorem 1, the following corollary holds.

    Corollary 1. For any a2λ, the function θλ,t(x) defined in (2.6) is strictly convex.

    The proof of Corollary 1 can be found in Appendix C.

    The graph presented in Figure 2(a) shows the plots of the function θλ,t(x)=λϕa(x)+12(xt)2 with t=0, λ=5, and a=15,50. It can be seen in Figure 2(a) that the function θλ,t(x) is strictly convex for a2λ. However, when λ=5, and a=0.5,1, the graph presented in Figure 2(b) shows that the function θλ,t(x) is non-convex.

    Figure 2.  The behavior of the function θλ,t(x) with a, λ=5, and t=0.

    From Theorem 1, the convexity of the function θλ,t(x) can be ensured by constraining the parameter a as a2λ, which means that there exists the unique global minimizer to θλ,t(x) when a2λ. The unique global minimizer to θλ,t(x) can be expressed in the following theorem.

    According to (2.1), we consider another function ψa(x), which is induced by ϕa(x):

    ψa(x):=|x|ϕa(x). (2.8)

    Figure 3 shows the curves of the functions |x|, ϕa(x), ψa(x) and x2a(a=10). Clearly, the function ψa(x) is a convex, continuously differentiable function that satisfies 0ψa(x)|x| and ψa(x)x2a for all xR.

    Figure 3.  Plots of the |x|, ϕa(x), ψa(x), and x2a(a=10) functions.

    Then, the MTL1 function can be written as

    ϕa(x)=|x|ψa(x). (2.9)

    To establish our MTL1TV MRI model, we define a multivariate penalty Φa:RnR based on the above penalty ϕa. For this purpose, we first define a multivariate generalization form of the corresponding function ψa in (2.8).

    Definition 2. Let ψa be correspondingly defined in (2.8); then Ψa:RnR is defined as

    Ψa(v)=ni=1ψa(vi),vRn. (2.10)

    Remark 1. The function Ψa satisfies 0Ψa(v)v1 and Ψa(v)1av22 for all vRn.

    Then, the multivariate penalty of ϕa is given by

    Φa(v)=ni=1(|vi|ψa(vi))=v1Ψa(v),vRn (2.11)

    Finally, by replacing v with gradient Dx in (2.11) (D is the first-order difference operator/matrix), we obtain the definition of MTL1TV below.

    Definition 3 (MTL1TV). The MTL1TV regularizer: xMTL1TV:RnR,

    xMTL1TV:=Φa(Dx)=Dx1Ψa(Dx). (2.12)

    Now, the MR image reconstruction model can be formulated as

    minxλxMTL1TV+12yAx22, (2.13)

    where λ>0 is called the regularization parameter. Because the MTL1 penalty function is non-convex, the model in (2.13) is non-convex. However, the global convexity of the objective function can be controlled by adjusting parameter a, as shown in Theorem 2 below.

    Theorem 2. Let λ>0, a>0. Define F:RnR as

    Fa(x)=12yAx22+λxMTL1TV, (2.14)

    If a2λ and DTDATA, then Fa(x) is a convex function.

    Proof. Substituting (2.12) into (2.14), we have

    Fa(x)=12yAx22+λxMTL1TV=12yAx22+λ(Dx1Ψa(Dx))=12[y222yTAx+Ax22]+λDx1λΨa(Dx)=[12y22yTAx+λDx1]+[12Ax22λΨa(Dx)].

    Note that 12y22 does not depend on x and that yTAx is linear in x. Hence, the function Fa(x) is convex if F2(x) is convex, where F2(x) is defined as

    F2(x)=12Ax22λΨa(Dx)=12(Ax222λaDx22)+λ(1aDx22Ψa(Dx))=12xT(ATA2λaDTD)x+λ(1aDx22Ψa(Dx)).

    The first term is convex if ATA2λaDTD. Since a2λ and are DTDATA given, it follows that the first term is convex. From Remark 1, the last term is convex. Hence, F2(x) is convex. Therefore, Fa(x) is convex.

    In this section, the optimization algorithm is presented in detail. From Theorem 2, (2.13) is a convex optimization problem under certain conditions, which can be solved via the convex optimization algorithms. Hence, the ADMM is used to effectively solve the model in (2.13). First, according to the definition of xMTL1TV, we rewrite (2.13) as follows:

    minxλΦa(Dx)+12yAx22. (3.1)

    Next, we introduce the auxiliary variable z with the constraint z=Dx. Then, the optimization problem (3.1) is rewritten as

    minxλΦa(z)+12yAx22,s.t.z=Dx. (3.2)

    Under the ADMM framework, the augmented Lagrangian function of (3.2) is

    L(x,z,w,β)=λΦa(z)+12yAx22w,zDx+β2zDx22 (3.3)
    =λΦa(z)+12yAx22+β2zDxwβ22w222β, (3.4)

    where wRn is the Lagrange multiplier, and β>0 is a penalty parameter. Then, we invoke the ADMM by iterating the variable updates in (3.5) to (3.7):

    xk+1=argminxL(x,zk,wk,βk)=argminx{12yAx22+wk,Dx+βk2zkDx22}, (3.5)
    zk+1=argminzL(xk+1,z,wk,βk)=argminz{λΦa(z)wk,z+βk2zDxk+122}, (3.6)
    wk+1=wk+βk(Dxk+1zk+1),βk+1=θβk. (3.7)

    Now, we give the detailed steps for solving the subproblems in (3.5) and (3.6) alternatively.

    Step 1. Update xk+1 with zk, and keep wk fixed. Let the gradient of this objective function be zero, and obtain the following equation:

    (βDTD+ATA)xk+1=βDTzk+ATyDTwk. (3.8)

    The matrix A=RF, where F is the Fourier operator with the property FT=F1. Since DTD is a cyclic matrix, it can be diagonalized by Fourier transform. Therefore, xk+1 can be solved by using two Fourier transforms:

    xk+1=(βDTD+ATA)1(βDTzk+ATyDTwk). (3.9)

    Step 2. Update zk+1 with xk+1, and keep wk fixed. From (3.6), we have

    zk+1=argminz{λΦa(z)+β2z(Dxk+1+wkβ)22}.

    According to Theorem 1, zk+1 is given by the following expression:

    zk+1=proxΦaλ/β(Dxk+1+wkβ). (3.10)

    Based on the above analysis, the specific algorithm for solving (3.1) can be summarized as Algorithm 1.

    Algorithm 1 ADMM for solving the MTL1TV model
    1: Input: a>0, β>0, λ>0, θ>1, k=0, u0=(x0,z0,w0), maxiter
    2: while not converged, do
    3:  Update xk+1 via (3.9).
    4:  Update zk+1 via (3.10).
    5:  Update wk+1 via (3.7).
    6:  Set uk+1=(xk+1,zk+1,wk+1).
    7:  k=k+1.
    8: end while

    In this section, we discuss the convergence of the proposed alternating minimization algorithm. Through this paper, we assume that ker(A)ker(D)=0. The assumption is very reasonable for the imaging problem [31,32]. In Algorithm 1, the dominant computation is the steps to solve the two minimization subproblems (3.5) and (3.6). Inspired by [31,32], we have the following convergence result for Algorithm 1.

    Assume that (x,z,λ) is a stationary point that satisfies the first order optimality conditions of (3.6):

    0=AT(Axy)+DTw,0λΦa(z)w,0=Duz. (4.1)

    Obviously, one can easily find that x satisfies

    0λΦa(z)+AT(Axy), (4.2)

    which is the first order necessary condition of (2.13) about stationary point. Otherwise, by Algorithm 1, each iteration step about the subproblems (3.5)–(3.7) follows:

    0=AT(Axk+1y)+DTwk+βkDT(Dxk+1zk), (4.3)
    0λΦa(zk+1)wkβk(Dxk+1zk+1), (4.4)
    wk+1=wk+βk(Dxk+1zk+1). (4.5)

    In the following, we will prove the sequence {xk,zk,λk} generated by the Algorithm 1 has a limit point (x,z,λ) that satisfies (4.1). The following two lemmas state that the metric L(xk,zk,wk,βk) and the sequences xk and {zk} are bounded.

    Lemma 3. Let L(xk,zk,wk,βk) be the sequence generated by Algorithm 1, then {wk} and {L(xk,zk,wk,βk)} are bounded.

    Proof. First, we prove the Lagrange multiplier wkRn is bounded and wknλ. Let dλϕa(x), then, if x>0, d=λa2(a+x)2(0,λ). If x<0, d=λa2(ax)2(λ,0). If x=0, d[λ,λ]. Therefore, |d|λ. On the other hand, combine (4.4) and (4.5), and we have wk+1λΦa(zk+1). Thus, by |d|λ, we have wk+1λ. Because wk+1 is finite dimensional, it implies that {wk} is bounded and wknλ<. Then, we prove {L(xk,zk,wk,βk)} is bounded. By (4.5), we notice that

    Dxk+1zk+1=wk+1wkβk. (4.6)

    From the above formula and (3.7), we have

    L(xk+1,zk+1,wk+1,βk+1)L(xk+1,zk+1,wk+1,βk)=βk+12zk+1Dxk+122βk2zk+1Dxk+122=βk+1βk2(βk)2wk+1wk22, (4.7)
    L(xk+1,zk+1,wk+1,βk)L(xk+1,zk+1,wk,βk)=wkwk+1,zk+1Dxk+1=1βkwk+1wk22. (4.8)

    By the definitions of zk+1 and xk+1, we know that

    L(xk+1,zk+1,wk,βk)L(xk+1,zk,wk,βk)0, (4.9)
    L(xk+1,zk,wk,βk)L(xk,zk,wk,βk)0. (4.10)

    Summing (4.7)–(4.10), we have

    L(xk+1,zk+1,wk+1,βk+1)L(xk,zk,wk,βk)βk+1+βk(2βk)2wk+1wk222(θ+1)nλ2β0θk, (4.11)

    Summing up the above inequality from k=0, we obtain

    L(xk+1,zk+1,wk+1,βk+1)L(x0,z0,w0,β0)2n(θ+1)λ2(11θk+1)β0(11θ), (4.12)

    Because θ>1, let k, and we have

    limkL(xk+1,zk+1,wk+1,βk+1)<L(x0,z0,w0,β0)+2nθ(θ+1)λ2β0(θ1)<, (4.13)

    On the other hand, from limkβk=, and the fact that {wk+1} is bounded, we know

    limkDxk+1zk+122=limkwk+1wkβk22=0, (4.14)

    and then

    limkL(xk+1,zk+1,wk+1,βk+1)limkwk+1,Dxk+1zk+1=0. (4.15)

    By (4.13) and (4.15), we know

    0limkL(xk+1,zk+1,wk+1,βk+1)<. (4.16)

    So, {L(xk+1,zk+1,wk+1,βk+1)} is bounded.

    Lemma 4. Let {(xk,zk)} be the sequence generated by Algorithm 1, then the sequences {xk} and {zk} are bounded.

    Proof. From (4.9) and (4.10), we obtain that L(xk+1,zk+1,wk,βk) is upper bounded. Next, we know that L(xk,zk,wk,βk) is strongly convex about the variable x, and the following inequality holds:

    L(xk+1,zk,wk,βk)L(xk,zk,wk,βk)ck2xk+1xk22. (4.17)

    Summing (4.7)–(4.9) and (4.17), we have

    L(xk+1,zk+1,wk+1,βk+1)L(xk,zk,wk,βk)ck2xk+1xk22+2(θ+1)nλ2β0θk. (4.18)

    Summing the above inequality from k=0, we have

    L(xk+1,zk+1,wk+1,βk+1)L(x0,z0,w0,β0)2(θ+1)nλ2β0(θ1)0ikci2xi+1xi22, (4.19)

    which together with (4.16) yields

    limkxk+1xk22=0. (4.20)

    Otherwise, if limkxk+1xk220, then 0ikci2xi+1xi22=+; thus,

    limkL(xk+1,zk+1,wk+1,βk+1)=,

    and (4.16), a contradiction. By (4.20), we know that {xk} is a Cauchy sequence and, thus, is convergent and bounded.

    Furthermore, we show that {zk} is bounded. By (3.4) and (4.12), we have

    λΦa(zk)+12yAxk22=L(zk,xk,wk1,βk1)+wk1222βk1βk12zkDxkwk1βk122=L(zk,xk,wk1,βk1)+12βk1(wk122wk22).

    Because limkβk= and {wk} are bounded, we also know that {zk} and {Axk} are bounded.

    Now, we are ready for proving the following convergence result.

    Theorem 3. Let {xk,zk,wk} be the sequence generated by Algorithm 1, then, there exists a subsequence {xkj,zkj,wkj}, which converges to a stationary point (x,z,w) and satisfies (4.1).

    Proof. By Lemmas 3 and 4, the sequence {xk,zk,wk} is bounded, so there exists a subsequence {xkj,zkj,wkj} that converges to a cluster point (x,z,w). From the lower semi-continuity of L(x,z,w,β),

    lim infjL(xkj+1,zkj+1,wkj,βkj)L(x,z,w,β)

    and as zkj+1 is a minimizer of function L(xkj+1,zkj+1,wkj,βkj) with respect to z,

    lim supjL(xkj+1,zkj+1,wkj,βkj)L(x,z,w,β).

    By the above two inequalities, we get

    limjλΦa(zkj)=λΦa(z).

    One can immediately verify that

    0=AT(Axy)+DTw,0λΦa(z)w,0=Duz. (4.21)

    Therefore, (x,z,w) is a Karush-Kuhn-Tucker point of the problem (3.2).

    Remark 2. From Corollary 1, if a2λ, the two subproblems (3.5) and (3.6) are continuous, coercive, and strictly convex. Consequently, every subproblem has a global solution. Additionally, from Theorem 2, we know that the objective function is a convex function, which implies the stationary point is the global minimum point.

    In this section, we present experiments conducted to verify the performance of the MTL1TV method in MATLAB 2019a on a laptop with 32 GB RAM and an Intel Core i7-12700H CPU at 2.70 GHz. We compared our proposed method with the standard TV [11], MCTV [24], and transformed total variation (TTV) [33] methods. For fair performance comparison, the parameters used in comparing algorithms have been appropriately tuned to yield a better reconstruction effect. We used three metrics to assess the quality and accuracy of image reconstruction: the relative error (RE), peak signal-to-noise ratio (PSNR), and structural similarity index (SSIM). These metrics are defined as follows:

    PSNR=10lg2552xkx2,RE=xkx2x2.

    where x, xk are the original image and reconstructed image, respectively, and

    SSIM=(2μxμxk+C1)(2σxxk+C2)(μ2x+μ2xk+C1)(σ2x+σ2xk+C2),

    where μx and μxk are the mean of x and xk, k is the number of iterations in the algorithm, σ2 and σ2xk denote the variance of x and xk, and σxxk denotes the covariance of x and xk. The positive constants C1 and C2 are stabilization constants. Generally, the MR image reconstruction is better if it has a lower RE and a higher PSNR and SSIM.

    The sampling templates and the experimental data are shown in Figure 4. The test data (a)–(c) was three typical MR images: Shepp-Logan, Brain, and Brain 2; (c) and (d) are two different brain MR images. The sampling templates (d)–(f) were radial, random, and Cartesian sampling, respectively. The size of sampling masks and MRI data was 256×256. In the experiments, the stopping criterion was usually as follows:

    xk+1xk2xk+12104,

    or the maximum number of iteration 200.

    Figure 4.  Experimental data and sampling patterns. (a) Shepp-Logan Phantom; (b) Brain; (c) Brain 2; (d) Radial sampling; (e) Random sampling; (f) Cartesian sampling.

    The sampling masks and MR data shown in Figure 4 are used to evaluate our proposed MTL1TV method. Table 1 displays the PSNR, SSIM, RE, and CPU time required to reconstruct various MR images using different methods with varying sampling templates. As observed in Table 1, the time required to reconstruct an MR image using TV is shorter than other methods, but its performance is inferior. Furthermore, when compared with TV, MCTV, and TTV, it is evident that MTL1TV exhibits higher PSNR and SSIM values with lower RE. This suggests that MTL1TV offers improved MR image reconstruction quality, simultaneously maintaining a reasonable runtime as shown in Table 1.

    Table 1.  Numerical results of the four algorithms.
    Test image Template Method RE (%) PSNR (dB) SSIM CPU time (s)
    Shepp-Logan Radial sampling 3% TV 15.71 28.2474 0.6794 1.649427
    MCTV 5.44 37.4617 0.7120 9.922404
    TTV 4.36 39.3778 0.7731 2.089988
    MTL1TV 2.74 43.4180 0.8824 2.081267
    Random sampling 30% TV 0.76 54.5952 0.9805 1.644969
    MCTV 0.59 56.7759 0.9858 9.789923
    TTV 0.09 72.6855 0.9996 2.157613
    MTL1TV 0.05 78.7386 0.9999 2.144944
    Cartesian sampling 34% TV 0.70 55.2138 0.9849 1.711586
    MCTV 0.53 57.6397 0.9904 9.709789
    TTV 0.12 70.5970 0.9999 2.839062
    MTL1TV 0.04 79.7220 0.9999 2.107535
    Brain Radial sampling 3% TV 22.28 23.4557 0.4981 1.648866
    MCTV 21.36 23.8205 0.6275 9.888602
    TTV 20.88 24.0185 0.7004 2.198074
    MTL1TV 19.78 24.4883 0.7877 2.186878
    Random sampling 30% TV 3.69 38.5007 0.8673 1.645874
    MCTV 2.84 41.3783 0.9171 10.657604
    TTV 1.23 48.6261 0.9924 3.549396
    MTL1TV 1.08 49.8008 0.9956 2.450209
    Cartesian sampling 34% TV 6.98 33.5339 0.8921 1.655883
    MCTV 5.64 35.3929 0.9243 9.734209
    TTV 5.64 35.3861 0.9374 2.348402
    MTL1TV 5.40 35.7689 0.9478 2.347148
    Brain2 Radial sampling 3% TV 18.88 22.7062 0.5848 1.651263
    MCTV 17.91 23.1698 0.6415 9.983066
    TTV 16.36 23.9466 0.7137 2.173014
    MTL1TV 16.28 23.9891 0.7244 2.149936
    Random sampling 30% TV 3.61 37.0681 0.9468 1.682748
    MCTV 2.99 38.7059 0.9646 9.940363
    TTV 2.58 39.9759 0.9723 2.426169
    MTL1TV 2.37 40.7350 0.9771 2.414485
    Cartesian sampling 34% TV 6.74 31.6495 0.9096 1.643168
    MCTV 6.17 32.4159 0.9240 9.893311
    TTV 6.29 32.2556 0.9208 2.343147
    MTL1TV 6.10 32.5116 0.9241 2.389379

     | Show Table
    DownLoad: CSV

    The visual comparison of reconstruction results and the error images are shown in Figures 57. In Figure 5, the Shepp-Logan phantom is used to demonstrate the performance of the proposed method. A Cartesian sampling at a sampling rate of 34% is employed to compare with the four reconstruction models proposed above. The regularization parameters λ for TV, MCTV, TTV, and MTL1TV are set to 0.0001, 0.005, 0.001, and 0.005, respectively. For TTV and MTL1TV, the parameter a is set to 1 and 0.05, respectively. From the Figure 5, it is observed that compared with other methods, MTL1TV has the best reconstruction capability, and the reconstructed image is most similar to the original image in the visual effects, which is reflected by the comparison of the difference images.

    Figure 5.  Results of Shepp-Logan reconstruction using a Cartesian mask with a sampling rate of 34%. (a) Original image; (f) Cartesian sampling; (b)–(e) reconstructed images using TV, MCTV, TTV, and MTL1TV, respectively; (g)–(i) difference images between (a) and (b)–(e).
    Figure 6.  Results of brain reconstruction using a radial mask with a sampling rate of 3%. (a) Original image; (f) Radial sampling; (b)–(e) reconstructed images using TV, MCTV, TTV, and MTL1TV, respectively; (g)–(i) difference images between (a) and (b)–(e).
    Figure 7.  Results of Brain 2 reconstruction using a random mask with a sampling rate of 30%. (a) Original image; (f) Random sampling; (b)–(e) reconstructed images using TV, MCTV, TTV, and MTL1TV, respectively; (g)–(i) difference images between (a) and (b)–(e).

    Figure 6 displays the reconstruction results of brain data under radial sampling with 10 trajectory lines and a 3% sampling ratio. We set the regularization parameters λ of TV, MCTV, TTV, and MTL1TV to 0.000001, 0.000005, 0.001, and 0.001, respectively. The value of parameter a is set as 0.5 for TTV, while it is set as 0.1 for MTL1TV. Visually, there is no significant difference. The residual images illustrate that the proposed method can visually gain better reconstructions.

    For the Brain 2 image, random sampling with a 30% sampling rate and a 0.1 sampling radius was employed to test the image. The regularization parameters λ were 0.00001, 0.00001, 0.0001, and 0.0005. The value of a in TTV was a=0.5, while in MTL1TV, a=0.1. The reconstructed images and errors are shown in Figure 7. It shows that the reconstruction results of the four methods are similar, and MTL1TV is slightly better than the other three methods.

    Additionally, the convergence of the proposed method is empirically tested. In Figure 8, we illustrate the behavior of the proposed MTL1TV model in the iteration process. It respectively shows the curves of the RE, PSNR, and SSIM values on Shepp-Logan phantom data under a radial mask with a 3% sampling ratio versus iteration numbers. It is shown that as the number of iterations increased, the RE value generated by MTL1TV gradually decreased, while the PSNR and SSIM values increased, which exhibits the stable convergence trend in the subsequent iterations.

    Figure 8.  Metric curves versus iteration number.

    In the proposed model, there are two important parameters: λ and a. The λ is the regularization parameter, which trades the sparsity and data consistency. The parameter a has an important impact on the properties of regularization term and it also affects the quality of reconstructed images. Therefore, it is crucial to choose appropriate values of a and λ, which can help us obtain better reconstruction results. The authors in [22] claimed that a=1 is the best among the tests on signals, and the author in [33] empirically selects a=5, which achieves better results in image reconstruction tasks. To demonstrate how to select parameters a and λ, we present the reconstruction results of the Shepp-Logan phantom (256×256) data using a Cartesian mask with a sampling rate of 34% as an example.

    To begin, we conduct the experiments with the fixed parameter a=1 and a=5, while allowing λ to vary. Subsequently, the parameter λ in TV, MCTV, and TTV can also be selected accordingly. The quantitative comparison results are listed in Table 2. From Table 2, we can see that when λ=0.005, the performance of the proposed model is the "best". Not only that, when λ=0.0001 in TV, λ=0.005 in MCTV, λ1=0.001 in TTV (when a=1), and λ2=0.0005 in TTV (when a=5), and the reconstructed metric PSNR achieves the highest value. Therefore, we set the parameter λ=0.0001 in TV, and λ=0.005 in MCTV.

    Table 2.  Reconstruction results of four models under different λ.
    λ 0.00005 0.0001 0.0005 0.001 0.005 0.01
    TV 55.1473 55.2138 54.9330 54.3397 47.3366 39.9643
    MCTV 56.6046 56.6744 57.0797 57.1284 57.6397 44.8390
    TTV(a=1) 28.6023 29.6150 43.5340 70.5970 55.9932 49.4821
    MTL1TV(a=1) 28.1268 28.6025 33.3470 43.5340 61.9251 55.9932
    TTV(a=5) 28.2477 28.8494 35.0278 47.6527 55.8338 49.9436
    MTL1TV(a=5) 28.1495 28.6462 33.4636 43.1281 57.4406 51.5726

     | Show Table
    DownLoad: CSV

    Next, we conduct the experiment with the fixed parameter λ1 and λ2, respectively, to asses the impact of different values of parameter a on reconstruction performance. From Table 3, it can be clearly seen that when the a=1, TTV exhibits excellent performance. Similarly, when a=0.05, MTL1TV achieves the best results. Furthermore, according to Table 3, it can be observed that when a=0.05>2λ2=0.01, the convexity of the MTL1TV model is clearly indicated. In Figure 9, it can be seen that as the parameter a assumes different values, the differences between the reconstructed MR image and the original image also vary. Notably, when a=0.05, the error images are the smallest.

    Table 3.  Reconstruction results of models TTV and MTL1TV under different a.
    Method(λ) a=0.005 a=0.01 a=0.05 a=0.1 a=0.5 a=1 a=5
    TTV(λ1) 50.5093 53.2998 68.8556 67.2232 69.3398 70.5970 47.6527
    TTV(λ2) 42.9035 42.9824 47.4019 47.6102 54.6126 55.9932 55.8338
    MTL1TV(λ2) 28.3835 29.8307 79.7220 74.5064 64.5339 61.9251 57.4406

     | Show Table
    DownLoad: CSV
    Figure 9.  Results of Shepp-Logan reconstruction using a Cartesian mask with a sampling rate of 34%. (a) Original image; (f) Cartesian sampling; (b)–(e) reconstructed images using a=0.005, a=0.01, a=0.05, and a=0.1, respectively; (g)–(i) difference images between (a) and (b)–(e).

    To demonstrate the performance of the proposed model in noisy condition, we conduct experiments on sampled data contaminated with additive Gaussian noise. In these experiments, Gaussian noise with standard deviation (σ=0.02) was added to the real and imaginary parts of sampled data, respectively. Figure 10 presents that the reconstructed results, along with the corresponding errors, for Shepp-Logan data under radial sampling (10 trajectory lines and a 3% sampling ratio). When compared to the reconstruction errors, it is evident that our proposed method effectively suppresses noise and artifacts. Table 4 presents the quantitative results of the three images (in Figure 4) with additive noise under radial sampling (10 trajectory lines and a 3% sampling ratio). Even though all models were given identical parameter settings in both noisy and noise-free environments, it was observed that the values of PSNR, SSIM, and RE for three images decreased under noisy conditions, as shown in Table 4. Nevertheless, they still produced results that were comparable to their noise-free counterparts. When compared to other methods, our method consistently delivered superior results, highlighting its robustness.

    Figure 10.  Reconstruction of Shepp-Logan with additive noise (σ=0.02) using a radial mask with a sampling rate of 3%. (a) Original image; (f) radial sampling; (b)–(e) reconstructed images using TV, MCTV, TTV, and MTL1TV, respectively; (g)–(i) difference images between (a) and (b)–(e).
    Table 4.  Numerical results for MR image reconstruction with additive noise from Radial mask.
    Test image Method RE (%) PSNR(dB) SSIM
    Shepp-Logan TV 21.48 25.5308 0.6446
    MCTV 17.00 27.5653 0.8167
    TTV 4.74 38.6586 0.7574
    MTL1TV 3.11 42.8507 0.8710
    Brain TV 22.31 23.4433 0.4966
    MCTV 21.46 23.7823 0.6257
    TTV 20.91 24.0075 0.6948
    MTL1TV 19.89 24.4386 0.7808
    Brain2 TV 6.80 31.5799 0.9069
    MCTV 6.27 32.2734 0.9214
    TTV 6.41 32.0904 0.9171
    MTL1TV 6.25 32.3101 0.9209

     | Show Table
    DownLoad: CSV

    In this paper, we introduced a non-convex TV regularization model for undersampled MRI reconstruction. In the new model, based on the MTL1 function, the traditional TV is replaced with the MTL1TV norm, thus improving the fitting performance of the model. We used the ADMM algorithm to compute the minimization problem. The experimental results demonstrated the effectiveness and efficiency of the MTL1TV method. In our future research, we seek to delve into the utilization of the proposed methodology in dynamic MRI, tomographic imaging, and electrical impedance tomography, by amalgamating deep learning frameworks with variational models.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    This work was supported by Natural Science Foundation of Hubei Province (2023AFD013), Philosophy and Social Sciences of Educational Commission of Hubei Province of China (22Y109), and Foundation of Hubei Normal University (2022055).

    The authors declare there is no conflict of interest.

    Proof. 1) Define the new variable η=a+x and substitute it into polynomial (2.6), then it becomes

    η3(a+t)η2+λa2=0, (A.1)

    whose discriminant is:

    Δ=λa2[4(a+t)327λa2].

    Due to t>t1 and Δ>0, the cubic equation has three distinct real roots. We change variables as η=y+a3+t3=x+a. The relation between x and y is: x=y2a3+t3. In terms of t, the cubic polynomial is turned into a depressed cubic as:

    y3(a+t)23y+λa22(a+t)327=0.

    The three roots in trigonometric form are:

    y0=a+t3cos(φ3)y1=2(a+t)3cos(φ3+π3)y2=2(a+t)3cos(π3φ3) (A.2)

    where φ=arccos(127λa22(a+t)3). It is obvious that y0>y1>y2 and y2<0. By the relation x=y2a3+t3, the three roots in variable x are xi=yi2a3+t3. From these formulas, we can prove x0>x1>x2 and x0t. Then, the largest root is x0, i.e., x0=y02a3+t3=gλ(t).

    2) We set: η=ax and y=ηa3+t3. So, x=y+2a3+t3. By a similar analysis as in part (1), there are three distinct roots for the polynomial equation: x0<x1<x2 with the smallest solution

    x0=2(at)3cos(φ3)+x3+2a3,

    where φ=arccos(127λa22(at)3). Therefore, x0=gλ(t), when t<t1.

    Proof. We discuss t=0, t>0, and t<0 in the following:

    1) t=0: In this case, optimization objective function is θλ,t(x)=λϕa(x)+12x2. It is true that λϕa(x) and 12x2 are both increasing for x>0, and decreasing with x<0. Thus, θ(0) is the unique minimizer for function θλ,t(x). So x=0, when t=0.

    2) t>0: Since 12(xt)2 and λϕa(x) are both decreasing with x<0, our optimal solution will only be obtained at nonnegative values. Thus, we just need to consider x0. When x0, we obtain

    θλ,t(x)=λa2(a+x)2+xt,

    and

    θλ,t(x)=12λa2(a+x)3.

    It is clear that θλ,t(x) is increasing, and θλ,t(0)=12λa determines the convexity for the function θλ,t(x). In the following proof, we further discuss the value of x by two conditions: λ<a2 and λ>a2.

    a) λ<a2: So, we have infx>0θλ,t(x)=θλ,t(0+)=12λa0, which means function θλ,t(x) is increasing for x0, with minimum value θλ,t(0)=λt=t2t.

    i) When 0xt2, θλ,t(x) is always positive, thus the optimal value x=0.

    ii) When x>t2, θλ,t(x) is first negative, then positive, and x>t2>t1. The unique positive stationary point x of θλ,t(x) satisfies equation: θλ,t(x)=0, which implies

    x(a+x)2t(a+x)2+λa2=0. (B.1)

    According to Lemma 2, the optimal value is x=x0=gλ(t).

    Above all, the value of x is

    x={0,0tt2;gλ(t),x>t2

    under the condition λa2.

    b) λ>a2: In this case, due to the sign of θλ,t(x), we know that function θλ,t(x) is decreasing at first then switches to be increasing. Its minimum point is ¯x=32λa2a and the least value

    θλ,t(¯x)=λa2(a+¯x)2+¯xt=t1t.

    Then, θλ,t(x)t1t with x0.

    i) When 0tt1, function θλ,t(x) is always increasing. Thus, optimal value is x=0.

    ii) When tt2, θλ,t(0+)0. Thus, the function θλ,t(x) is first decreasing and then increasing. There is only one positive optimal stationary point, which is also the optimal solution. Using the Lemma 2, we know that x=θλ,t(x).

    iii) When t2<t<t1, θλ,t(0+)>0. Thus, the function θλ,t(x) is first increasing, then decreasing and finally increasing, which implies that there are two positive stationary points and the larger one is a local minima. Using Lemma 2, the local minimized point will be x0=θλ,t(x), the largest root of (1.1). Since x0t+λa2(a+x0)2=0, which implies λaa+x0=(tx0)(a+x0)a, we have

    θλ,t(x0)θλ,t(0)=12x20x0t+λax0a+x0=x0(12x0t+)λaa+x0=x0(12x0t+(tx0)(a+x0)a)=x20(tx0a12).

    Define a function h(t)=tgλ(t)a2.

    First, we prove that t=t3 is a solution to h(t)=0. Since λ>a2, t3=2λaa2>0. Thus:

    cos(φ(t3))=127λa22(a+t3)3=127λa22(a2+2λa)3.

    Further, by using the relation cos(φ)=4cos3(φ3)3cos(φ3) and 0φ3π3, we have

    cos(φ(t3)3)=2λaa4a2+2λa.

    Plugging this formula into gλ(t3) shows that gλ(t3)=2λaa=t3a2. Therefore, t3 is the root for h(t) and t3(t1,t2).

    Second, we prove that the function h(t) changes sign at point t=t3. We prefer to discuss it in two cases.

    Case 1: x(t3,t2). According to Lemma 2, we know that gλ(t) is the largest root of the cubic polynomial H(y)=y(a+y)2t(a+y)2+λa2 under the condition of t>t1.

    For function H(y), we have H(t)=λa2 and

    H(ta2)=λa2a2(a2+t)2.

    Due to t>t3=2λaa2 and H(ta2)<0, there is a root y=gλ(t), such that y=gλ(t)(ta2,t) for the equation H(y)=0. That is, tgλ(t)<a2, and, thus, h(t)<0. Case 2: x(t1,t3). We have H(ta2)>0 and H(t)>0. Due to the proof in Lemma 2, one possible situation is that there are two roots x0 and x1 within interval (ta2,t). However, we can exclude this case. This is because

    x0x1=2(a+t)3{cos(φ3)cos(φ3+π3)}=2(a+t)3{2sin(φ3)+π6)sin(π6)}=2(a+t)3sin(φ3+π6).

    Furthermore, x0x1>a2 holds for t>t1>a2 and λ>a2. This is a contradiction of the assumption that x0 and x1 are in (ta2,t). Thus, H(y)=0 has no root in (ta2,t).

    Therefore, x0=gλ(t)<ta2. That is to say, h(t)>0.

    From the discussion earlier, it is true that the optimal solution x=0, if 0<tt3, and x=x0=gλ(t), if t>t3.

    To sum up, we know that under the condition λ>a2,

    x={0,0tt3;gλ(t),x>t3

    3) t<0: Because

    infxθλ,t(x)=infxθλ,t(x)=infx12(x|t|)2+ϕa(t),

    x(t)=x(t), which implies that the formula obtained when t>0 above can extend to the case: t<0 by odd symmetry.

    Summarizing results from all cases, the proof is completed.

    Proof. It is clear to see that the function

    θλ,t(x)=λϕa(x)+12(xt)2

    is differentiable on R{0}. For x0, the derivative of θλ,t(x) is given by

    θλ,t(x)=λa2(a+|x|)2sgn(x)+(xt),x0.

    Let us find the range of a for which θλ,t(x) is convex. Notice that

    ϕa(0+)=1>1=ϕa(0)

    and

    θλ,t(x)=12λa2(a+x)3,x>0.

    When we set a2λ, the function θλ,t is increasing, then the function θλ,t(x) is strictly convex.



    [1] I. Pykett, J. Newhouse, F. Buonanno, T. Brady, M. Goldman, J. Kistler, et al., Principles of nuclear magnetic resonance imaging, Radiology, 143 (1982), 157–168. https://doi.org/10.1148/radiology.143.1.7038763 doi: 10.1148/radiology.143.1.7038763
    [2] X. Gu, W. Xue, Y. Sun, X. Qi, X. Luo, Y. He, Magnetic resonance image restoration via least absolute deviations measure with isotropic total variation constraint, Math. Biosci. Eng., 20 (2023), 10590–10609. https://doi.org/10.3934/mbe.2023468 doi: 10.3934/mbe.2023468
    [3] Y. Beauferris, J. Teuwen, D. Karkalousos, N. Moriakov, M. Caan, G. Yiasemis, et al., Multi-coil MRI reconstruction challenge assessing brain MRI reconstruction models and their generalizability to varying coil configurations, Front. Neurosci., 16 (2022), 1–16. https://doi.org/10.3389/fnins.2022.919186 doi: 10.3389/fnins.2022.919186
    [4] M. Lustig, D. Donoho, J. M. Pauly, Sparse MRI: The application of compressed sensing for rapid MR imaging, Magn. Reson. Med., 58 (2007), 1182–1195. https://doi.org/10.1002/mrm.21391 doi: 10.1002/mrm.21391
    [5] J. He, Q. Liu, A. Christodoulou, C. Ma, F. Lam, Z. Liang, Accelerated high-dimensional MR imaging with sparse sampling using low-rank tensors, IEEE Trans. Med. Imaging, 35 (2016), 2119–2129. https://doi.org/10.1109/TMI.2016.2550204 doi: 10.1109/TMI.2016.2550204
    [6] A. Tran, T. Nguyen, P. Doan, D. Tran, D. Tran, Parallel magnetic resonance imaging acceleration with a hybrid sensing approach, Math. Biosci. Eng., 18 (2021), 2288–2302. https://doi.org/10.3934/mbe.2021116 doi: 10.3934/mbe.2021116
    [7] F. Knoll, K. Hammernik, C. Zhang, S. Moeller, T. Sodickson, D. Sodickson, et al., Deep-Learning Methods for Parallel Magnetic Resonance Imaging Reconstruction: A Survey of the Current Approaches, Trends, and Issues, IEEE Signal Process. Mag., 37 (2020), 128–140. https://doi.org/10.1109/MSP.2019.2950640 doi: 10.1109/MSP.2019.2950640
    [8] D. Donoho, Compressed sensing, IEEE Trans. Inf. Theory, 52 (2006), 1289–1306. https://doi.org/10.1109/TIT.2006.871582 doi: 10.1109/TIT.2006.871582
    [9] J. Ye, Compressed sensing MRI: a review from signal processing perspective, BMC Biomed. Eng., 1 (2019), 1–17. https://doi.org/10.1186/s42490-019-0006-z doi: 10.1186/s42490-019-0006-z
    [10] X. Li, R. Feng, F. Xiao, Y. Yin, D. Cao, X. Wu, et al., Sparse reconstruction of magnetic resonance image combined with two-step iteration and adaptive shrinkage factor, Math. Biosci. Eng., 19 (2022), 13214–13226. https://doi.org/10.3934/mbe.2022618 doi: 10.3934/mbe.2022618
    [11] L. Rudin, S. Osher, E. Fatemi, Nonlinear total variation based noise removal algorithms, Phys. D, 60 (1992), 259–268. https://doi.org/10.1016/0167-2789(92)90242-F doi: 10.1016/0167-2789(92)90242-F
    [12] I. Selesnick, Sparse regularization via convex analysis, IEEE Trans. Signal Process., 65 (2017), 4481–4494. https://doi.org/10.1109/TSP.2017.2711501 doi: 10.1109/TSP.2017.2711501
    [13] D. Peleg, R. Meir, A bilinear formulation for vector sparsity optimization, Signal Process., 88 (2008), 375–389. https://doi.org/10.1016/j.sigpro.2007.08.015 doi: 10.1016/j.sigpro.2007.08.015
    [14] J. Fan, R. Li, Variable selection via nonconcave penalized likelihood and its oracle properties, J. Am. Stat. Assoc., 96 (2001), 1348–1360. https://doi.org/10.1198/016214501753382273 doi: 10.1198/016214501753382273
    [15] C. Zhang, Nearly unbiased variable selection under minimax concave penalty, Ann. Stat., 38 (2010), 894–942. https://doi.org/10.1214/09-AOS729 doi: 10.1214/09-AOS729
    [16] F. Zhang, H. Wang, W. Qin, X. Zhao, J. Wang, Generalized nonconvex regularization for tensor RPCA and its applications in visual inpainting, Appl. Intell., 53 (2023), 23124–23146. https://doi.org/10.1007/s10489-023-04744-9 doi: 10.1007/s10489-023-04744-9
    [17] J. Zou, M. Shen, Y. Zhang, H. Li, G. Liu, S. Ding, Total variation denoising with non-convex regularizers, IEEE Access, 7 (2018), 4422–4431. https://doi.org/10.1109/ACCESS.2018.2888944 doi: 10.1109/ACCESS.2018.2888944
    [18] M. Shen, J. Li, T. Zhang, J. Zou, Magnetic resonance imaging reconstruction via non-convex total variation regularization, Int. J. Imaging Syst. Technol., 31 (2021), 412–424. https://doi.org/10.1002/ima.22463 doi: 10.1002/ima.22463
    [19] I. Selesnick, I. Bayram, Sparse signal estimation by maximally sparse convex optimization, IEEE Trans. Signal Process., 62 (2014), 1078–1092. https://doi.org/10.1109/TSP.2014.2298839 doi: 10.1109/TSP.2014.2298839
    [20] S. Zhang, J. Xin, Minimization of transformed L1 penalty: Closed form representation and iterative thresholding algorithms, Commun. Math. Sci., 15 (2017), 511–537. https://doi.org/10.4310/CMS.2017.v15.n2.a9 doi: 10.4310/CMS.2017.v15.n2.a9
    [21] S. Zhang, J. Xin, Minimization of transformed L1 penalty: theory, difference of convex function algorithm, and robust application in compressed sensing, Math. Program., 169 (2018), 307–336. https://doi.org/10.1007/s10107-018-1236-x doi: 10.1007/s10107-018-1236-x
    [22] H. Li, Q. Zhang, A. Cui, J. Peng, Minimization of fraction function penalty in compressed sensing, IEEE Trans. Neural Networks Learn. Syst., 31 (2020), 1626–1637. https://doi.org/10.1109/TNNLS.2019.2921404 doi: 10.1109/TNNLS.2019.2921404
    [23] J. Li, Z. Xie, G. Liu, L. Yang, J. Zou, Diffusion optical tomography reconstruction based on convex-nonconvex graph total variation regularization, Math. Meth. Appl. Sci., 23 (2023), 4534–4545. https://orcid.org/0000-0001-7897-7151
    [24] Y. Liu, H. Du, Z. Wang, W. Mei, Convex MR brain image reconstruction via no-convex total variation minimization, Int. J. Imaging Syst. Technol., 28 (2018), 246–253. https://doi.org/10.1002/ima.22275 doi: 10.1002/ima.22275
    [25] Z. Luo, Z. Zhu, B. Zhang, An AtanTV nonconvex regularization model for MRI reconstruction, J. Sens., 2022 (2022), 1–15. https://doi.org/10.1155/2022/1758996 doi: 10.1155/2022/1758996
    [26] Z. Luo, Z. Zhu, B. Zhang, An SCADTV nonconvex regularization approach for magnetic resonance imaging, IAENG Int. J. Comput. Sci., 48 (2021), 1005–1012. https://api.semanticscholar.org/CorpusID: 248818745
    [27] Y. Lu, B. Zhang, Z. Zhu, Y. Liu, A CauchyTV non-convex regularization model for MRI reconstruction, Signal, Image Video Process., 17 (2023), 3275–3282. https://doi.org/10.1007/s11760-023-02542-x doi: 10.1007/s11760-023-02542-x
    [28] S. Boyd, N. Parikh, E. Chu, B. Peleato, J. Eckstein, Distributed optimization and statistical learning via the alternating direction method of multipliers, Found. Trends Mach. Learn., 3 (2011), 1–122. https://doi.org/10.1561/2200000016 doi: 10.1561/2200000016
    [29] D. G. Luenberger, Y. Ye, Linear and Nonlinear Programming, 3nd edition, Springer, USA, 2008. https://doi.org/10.1007/978-3-319-18842-3
    [30] J. Yang, Y. Zhang, W. Yin, A fast alternating direction method for TVL1-L2 signal reconstruction from partial fourier data, IEEE J. Sel. Top. Signal Process., 4 (2010), 288–297. https://doi.org/10.1109/JSTSP.2010.2042333 doi: 10.1109/JSTSP.2010.2042333
    [31] B. Zhang, G. Zhu, Z. Zhu, S. Kwong, Alternating direction method of multipliers for nonconvex log total variation image restoration, Appl. Math. Modell., 114 (2023), 338–359. https://doi.org/10.1016/j.apm.2022.09.018 doi: 10.1016/j.apm.2022.09.018
    [32] J. You, Y. Jiao, X. Lu, T. Zeng, A nonconvex model with minimax concave penalty for image restoration, J. Sci. Comput., 78 (2019), 1063–1086. https://doi.org/10.1007/s10915-018-0801-z doi: 10.1007/s10915-018-0801-z
    [33] L. Huo, W. Chen, H. Ge, M. K. Ng, Stable image reconstruction using transformed total variation minimization, SIAM J. Imaging Sci., 15 (2022), 1104–1139. https://doi.org/10.1137/21M1438566 doi: 10.1137/21M1438566
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(985) PDF downloads(31) Cited by(0)

Figures and Tables

Figures(10)  /  Tables(4)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog