Loading [MathJax]/jax/element/mml/optable/BasicLatin.js
Research article

MRI image enhancement based on feature clustering in the NSCT domain

  • Received: 29 December 2021 Revised: 24 May 2022 Accepted: 14 June 2022 Published: 23 June 2022
  • MSC : 91C20, 68U10

  • The noise and low clarity in magnetic resonance imaging (MRI) images impede the doctor's diagnosis. An MRI image enhancement method is proposed in the non-subsampled contourlet transform (NSCT) domain. The coefficients of the NSCT are classified as noise component, weak edges component and strong edges component by feature clustering. We modified the transform coefficients to enhance the MRI images. The coefficients corresponding to noise are set to zero, the coefficients corresponding to strong edges are essentially unchanged, and the coefficients corresponding to weak edges are enhanced by a simplified nonlinear gain function. It is shown that the proposed MRI image enhancement method has advantages in visual quality and objective evaluation indexes compared to the state-of-the-art methods.

    Citation: Xia Chang, Haixia Zhao, Zhenxia Xue. MRI image enhancement based on feature clustering in the NSCT domain[J]. AIMS Mathematics, 2022, 7(8): 15633-15658. doi: 10.3934/math.2022856

    Related Papers:

    [1] Yuxin Luo, Yu Fang, Guofei Zeng, Yibin Lu, Li Du, Lisha Nie, Pu-Yeh Wu, Dechuan Zhang, Longling Fan . DAFNet: A dual attention-guided fuzzy network for cardiac MRI segmentation. AIMS Mathematics, 2024, 9(4): 8814-8833. doi: 10.3934/math.2024429
    [2] S. Neelakandan, Sathishkumar Veerappampalayam Easwaramoorthy, A. Chinnasamy, Jaehyuk Cho . Fuzzy adaptive learning control network (FALCN) for image clustering and content-based image retrieval on noisy dataset. AIMS Mathematics, 2023, 8(8): 18314-18338. doi: 10.3934/math.2023931
    [3] Xiang Gao, Linzhang Lu, Qilong Liu . Non-negative Tucker decomposition with double constraints for multiway dimensionality reduction. AIMS Mathematics, 2024, 9(8): 21755-21785. doi: 10.3934/math.20241058
    [4] Thavavel Vaiyapuri, Prasanalakshmi Balaji, S. Shridevi, Santhi Muttipoll Dharmarajlu, Nourah Ali AlAseem . An attention-based bidirectional long short-term memory based optimal deep learning technique for bone cancer detection and classifications. AIMS Mathematics, 2024, 9(6): 16704-16720. doi: 10.3934/math.2024810
    [5] Iliyas Karim khan, Hanita Binti Daud, Nooraini binti Zainuddin, Rajalingam Sokkalingam, Abdussamad, Abdul Museeb, Agha Inayat . Addressing limitations of the K-means clustering algorithm: outliers, non-spherical data, and optimal cluster selection. AIMS Mathematics, 2024, 9(9): 25070-25097. doi: 10.3934/math.20241222
    [6] Wenxue Sun, Huiyuan Zhao, Xiao Zhang, Yuchao Sun, Xiaoxin Liu, Xueling Lv, Di Fan . Zero-watermarking Algorithm for Audio and Video Matching Verification. AIMS Mathematics, 2022, 7(5): 8390-8407. doi: 10.3934/math.2022468
    [7] Caicai Feng, Saratha Sathasivam, Nurshazneem Roslan, Muraly Velavan . 2-SAT discrete Hopfield neural networks optimization via Crow search and fuzzy dynamical clustering approach. AIMS Mathematics, 2024, 9(4): 9232-9266. doi: 10.3934/math.2024450
    [8] Sergio Iglesias-Perez, Alberto Partida, Regino Criado . The advantages of k-visibility: A comparative analysis of several time series clustering algorithms. AIMS Mathematics, 2024, 9(12): 35551-35569. doi: 10.3934/math.20241687
    [9] Yan Ma, Defeng Kong . Super-resolution reconstruction algorithm for dim and blurred traffic sign images in complex environments. AIMS Mathematics, 2024, 9(6): 14525-14548. doi: 10.3934/math.2024706
    [10] Yang Li, Tianshi Wang, Li Liu . Random style transfer for person re-identification with one example. AIMS Mathematics, 2021, 6(5): 4715-4733. doi: 10.3934/math.2021277
  • The noise and low clarity in magnetic resonance imaging (MRI) images impede the doctor's diagnosis. An MRI image enhancement method is proposed in the non-subsampled contourlet transform (NSCT) domain. The coefficients of the NSCT are classified as noise component, weak edges component and strong edges component by feature clustering. We modified the transform coefficients to enhance the MRI images. The coefficients corresponding to noise are set to zero, the coefficients corresponding to strong edges are essentially unchanged, and the coefficients corresponding to weak edges are enhanced by a simplified nonlinear gain function. It is shown that the proposed MRI image enhancement method has advantages in visual quality and objective evaluation indexes compared to the state-of-the-art methods.



    Magnetic resonance imaging (MRI), a kind of non-invasive medical imaging, can reconstruct the internal structure of the human body. MRI has been widely used in medical diagnosis [1,2,3]. The MRI images with abundant details and high clarity are effective in helping doctors to make observations and get a diagnosis for the interior tissue of the human body.

    There exists noise and the contrast of light and shade is high in MRI images, that impede the doctor's diagnosis. Recently we can obtain the MRI images nearly undisturbed by noise with the development of imaging technology and the improvement of the hardware devices. MRI images are used as a diagnostic evidence, so the interior tissue of the human body in the MRI images are expected not only to be visible but also be seen clearly. However the weak details of the MRI images are easily covered up. Highlighting the details to benefit diagnosis is essential in MRI images enhancement.

    The enhancement methods [4,5,6] available include spatial domain based methods, transform domain based methods and deep learning based methods. The spatial domain based methods contain histogram equalization (HE) [7,8], contrast limited adaptive histogram equalization (CLAHE) [9,10], unsharp-mask filtering (UM) [11,12] and et al. [13,14]. Spatial domain based methods have some deficiencies confronting with the MRI images enhancement. The HE concerns with the distribution of the whole gray-level in images. When the contrast of light and shade is high in an image, the HE will introduce obvious whitewashed effect to the enhanced result. That affects the image interpretation [15]. The CLAHE considers the distribution of the local gray-level in images and can enhance the contrast of the local region. But this method can also enhance the noise appearing in the scene. The UM enhances the details in the MRI images effectively, while it is susceptible to noise. The noise existing in the images, even a tiny amounts of noise, will be enhanced by the UM [16]. The enhancement methods based on transform have advantages in MRI images enhancement [17,18,19,20,21]. The transform based enhancement methods distinguish the coefficients corresponding to noise, weak edges, and strong edges by appropriate thresholds. The enhancement procedure is accomplished by setting the coefficients corresponding to noise as 0, enhancing the coefficients corresponding to weak edges, and retaining the coefficients corresponding to strong edges. However, the appropriate thresholds are difficult to choose if there is lack of prior knowledge of MRI images. Deep learning based methods are a new kind of methods for image enhancement, which have achieved favorable results. It can't be ignored that the set of parameters in the deep network is complicated, and the design of parameters will directly affect the efficiency of the algorithm.

    Wavelet transform possesses favorable properties in time-frequency analysis, which has been used widely in MRI images enhancement. It has been found that wavelet transform can't capture the intrinsic geometrical information in images with further research [22]. The NSCT, a multiscale geometrical analysis method, makes up for the deficiency of the wavelet transform [23]. The NSCT can capture the edges, contours and texture information in images effectively [24,25]. The NSCT is suitable for expressing the MRI images bearing abundant detail information.

    To overcome the drawbacks of the image enhancement methods available and take advantage of the good descriptive ability of the NSCT, we propose an MRI image enhancement method based on feature clustering. According to the different exhibitions of the noise, weak edge and strong edge in the high-frequency directional subbands, the three-dimensional feature is extracted. Without adjusting complicated threshold parameters, the Fuzzy C-means (FCM) clustering method is introduced in this paper to classify the noise component, weak edge component and strong edge component of the high-frequency directional subbands coefficients. The gain rule is designed to modify the high-frequency coefficients. Experiments show its effectiveness.

    The NSCT transform is a shift-invariant version of the contourlet transform. The NSCT inherits the favorable characteristics of the contourlet transform, as multi-scale, locality, multi-directionality and anisotropy. The NSCT allows redundancy, therefore, it can possess abundant basis functions. That is very favourable to express the MRI images with plenty of detail information. High-frequency directional subbands in different scales and different directions will be obtained after the NSCT decomposition of an image. The High-frequency directional subbands and the original image have the same size. The wavelet transform has high-frequency subbands with limited directions. Unlike the wavelet transform, the number of the high-frequency subbands of the NSCT is the arbitrary powers of two, which can be set up artificially. The detail information of images in fine scales is more plentiful than that in coarse scales. So the number of the NSCT high-frequency subbands in fine scales is set up with a large value in order to capture more detail information.

    Figure 1 is the NSCT decomposition of a real MRI image on the abdomen. We decompose the MRI image into 4 scales. The numbers of high-frequency directional subbands from fine scale to coarse scale are 8, 8, 4 and 4 respectively. The equivalent filter for the k-layer scale decomposition of the signal is:

    Heqn(Z)={H1(Z2n1)n2j=0H0(Z2j)1n<2k,n1j=0H0(Z2j)n=2k. (1)

    A directional decomposition can be obtained by iterating the non-downsampling direction filter bank U(Z) :

    Ueqk(Z)=Ui(Z)Uj(zQ). (2)

    Q is a matrix.

    Q=(1111). (3)

    Figure 1(a) is the original MRI image. Figure 1(b) is the partition graph of frequency in 8 directions. It can be seen that the frequency domain is divided into 8 wedge-shaped subbands. Figure 1(c1)-(c8) are the basis functions carrying out the NSCT decomposition in 8 directions. Figure 1(d1)-(d8) are the high-frequency directional subbands obtained corresponding to the basic functions in Figure 1(c1)-(c8). The basic functions of the NSCT have abundant directions and shapes, so the NSCT is able to capture the geometrical structure effectively.

    Take an observation about the high-frequency directional subbands in the same scale, as shown in Figure 1(d1)-(d8). As in Ref. [20], it can be found that:

    1) Strong edges have significant directional characteristic. At the same scale, the coefficients corresponding to strong edges exhibit large amplitude in each high-frequency directional subband.

    2) Weak edges also have geometrical structure. At the same scale, the coefficients corresponding to weak edges exhibit large amplitude in some high-frequency directional subbands while exhibit small amplitude in others.

    3) Noise does not have geometrical structure. At the same scale, the coefficients corresponding to noise exhibit rather small amplitude in each high-frequency directional subband.

    These characteristics present the advantages of the NSCT in multi-directional expression of images. We also acquire the basis to divide coefficients as strong edges component, weak edges component and noise component.

    Figure 1.  NSCT decomposition of a real abdominal MRI image.

    The image enhancement methods based on transform distinguish the coefficients into strong edges component, weak edges component and noise component by appropriate thresholds. The threshold can be seem as the 1-dimensional feature. According to the observation of NSCT coefficients in Section 2, we don't cling to 1-dimensional feature and provide the strategy of feature selection. We expect the features selected are able to classify the NSCT coefficients as strong edges component, weak edges component and noise component effectively.

    I(m,n) is the input image with the size M×N. After employing a L-level decomposition of the NSCT, we will get high-frequency directional subbands {Dl,i(m,n)} in different scales. 0l<L1, 1ikl, 1mM, 1nN. kl denotes the number of high-frequency directional subbands at the scale 2l. The high-frequency directional subbands at the same scale are represented as a matrix Dl with the size M×N×kl. The elements of Dl at the location (m,n) constitute a kl -dimensional vector. That is Dl(m,n)=[Dl,1(m,n),Dl,2(m,n),...,Dl,kl(m,n)]T. We extract 3-dimensional feature from Dl, viz. mean value, variance and maximum value. The strategy of feature selection is detailed as follows:

    1) The mean value feature

    The coefficients corresponding to noise have small modulus in each high-frequency directional subband at the same scale. So, if Dl(m,n) corresponds to noise, the mean value of the elements modulus in Dl(m,n) is small. The coefficients corresponding to weak edges have large modulus in some high-frequency directional subbands at the same scale. So, if Dl(m,n) corresponds to weak edges, the mean value of the elements modulus in Dl(m,n) is at medium level. The coefficients corresponding to strong edges have large modulus in each high-frequency directional subband at the same scale. So, if Dl(m,n) corresponds to strong edges, the mean value of the elements modulus in Dl(m,n) is large. We adopt the mean value of the elements modulus in Dl(m,n) as a feature to distinguish the three different components. The feature is represented as Zlmean(m,n). We have

    Zlmean(m,n)=1klkli=1|Dl,i(m,n)|. (4)

    2) The standard deviation feature

    The standard deviation is generally used to measure the dispersion of data apart from the mean value. At the same scale, the coefficients corresponding to weak edges in some high-frequency directional subbands have large modulus, but the coefficients corresponding to weak edges have small modulus in other high-frequency directional subbands. So, if Dl(m,n) corresponds to weak edges, the dispersion of elements modulus in Dl(m,n) is large. That means the standard deviation of elements modulus in Dl(m,n) is large. The modulus of the coefficients corresponding to noise or strong edges have small dispersion in high-frequency directional subbands at the same scale. If Dl(m,n) corresponds to noise or strong edges, the standard deviation of elements modulus in Dl(m,n) is small. We adopt the standard deviation of the elements modulus in Dl(m,n) as another feature to distinguish the three different components. The feature is represented as Zlstd(m,n). We have

    Zlstd(m,n)=1kl1kli=1(|Dl,i(m,n)|1klkli=1|Dl,i(m,n)|)2. (5)

    3) The maximal modulus feature

    At the same scale, the coefficients corresponding to noise in each high-frequency directional subbands have small modulus. If Dl(m,n) corresponds to noise, the maximal modulus of elements in Dl(m,n) is still small. The coefficients corresponding to weak edges in some high-frequency directional subbands have large modulus at the same scale. If Dl(m,n) corresponds to weak edges, the maximal modulus of elements in Dl(m,n) is large. The coefficients corresponding to strong edges in each high-frequency directional subbands have large modulus. If Dl(m,n) corresponds to strong edges, the maximal modulus of elements in Dl(m,n) is larger than that corresponding to weak edges or noise. We adopt the maximal modulus of elements in Dl(m,n) as the third feature to distinguish the three different components. The feature is represented as Zmax. We have

    Z_{\max }^{l}(m, n)=\max \left\{\left|D_{l, i}(m, n)\right|\right\}_{1 \leq i \leq k_{l}}. (6)

    In conclusion, we extract 3-D dimensional feature vector from high-frequency directional subbands at the same scale. That is

    {Z^l}(m, n) = {[Z_{mean}^l(m, n), Z_{std}^l(m, n), Z_{\max }^l(m, n)]^T} . (7)

    Without estimating the threshold, we introduce the FCM [26,27] to distinguish the three components in NSCT coefficients by using the 3-D dimensional feature vector constructed. The FCM is an unsupervised classification methods, which acquires labels of the data by minimizing the global cost function. The features of coefficients at different scale are clustered respectively. The data to be clustered is {\boldsymbol{Z}^l}(m, n) , 0 \leqslant l < L - 1 , 1 \leqslant m \leqslant M , 1 \leqslant n \leqslant N . At the scale {2^{ - l}} , if {\boldsymbol{Z}^l}(m, n) is divided into a certain class (noise, weak edges or strong edges), the coefficients at the scale {2^{ - l}} in each high-frequency directional subband at the location (m, n) are deem to belong to this class.

    The feature clustering method based on FCM needs to minimize the objective function. That is

    \min J\left(\boldsymbol{U}, \boldsymbol{V} ; \boldsymbol{Z}^{l}\right)=\sum\limits_{p=1}^{3} \sum\limits_{q=1}^{M \times N} U_{p, q}^{r}\left\|\boldsymbol{Z}_{q}^{l}-\boldsymbol{V}_{p}\right\|^{2}, (8)

    where the objective function J(\boldsymbol{U}, \boldsymbol{V};{\boldsymbol{Z}^l}) is the weighted square of error between data \boldsymbol{Z}_q^l and cluster center {\boldsymbol{V}_p} . {\left\{ {\boldsymbol{Z}_q^l} \right\}_{1 \leqslant q \leqslant MN}} = {\left\{ {{Z^l}(m, n)} \right\}_{1 \leqslant m \leqslant M, 1 \leqslant n \leqslant N}} . That means the 2-dimensional indexes of the data \left\{ {{\boldsymbol{Z}^l}(m, n)} \right\} are changed to the 1-dimensional form. r is the weighted index in the FCM based clustering. The r is set as 2 in our algorithm. {\boldsymbol{V}_p} is computed by

    {\boldsymbol{V}_p} = \frac{{\sum\limits_{q = 1}^{M \times N} {{{\left( {{U_{p, q}}} \right)}^2}} Z_q^l}}{{\sum\limits_{q = 1}^{M \times N} {{{\left( {{U_{p, q}}} \right)}^2}} }} , 1 \leqslant p \leqslant 3 . (9)

    {d_{p, q}} is the distance between data \boldsymbol{Z}_q^l and cluster center {\boldsymbol{V}_p} , which is defined as

    {d_{p, q}} = {\left\| {\boldsymbol{Z}_q^l - {\boldsymbol{V}_p}} \right\|^2} \text{, } 1 \leqslant p \leqslant 3 \text{, }. 1 \leqslant q \leqslant M \times N . (10)

    {U_{p, q}} is the degree of membership from data \boldsymbol{Z}_q^l to cluster center {\boldsymbol{V}_p} .

    {U_{p, q}} = \frac{{1/{d_{p, q}}}}{{\sum\limits_{{p^{'}} = 1}^3 {\left( {1/{d_{{p^{'}}, q}}} \right)} }} . (11)
    \sum\limits_{p = 1}^3 {{U_{p, q}}} = 1 \text{, } 1 \leqslant q \leqslant M \times N .

    The objective function J(\boldsymbol{U}, \boldsymbol{V};{\boldsymbol{Z}^l}) is minimized by iterating formulas (6)-(8) until the optimal membership matrix \boldsymbol{U}_q^ * = {[U_{1, q}^ *, U_{2, q}^ *, U_{3, q}^ *]^T} of data {\left\{ {\boldsymbol{Z}_q^l} \right\}_{1 \leqslant q \leqslant MN}} is obtained. For the reason of timeliness, the iteration is stopped in the case that the number of iterative time is more than 50 or the change is smaller than 0.001 in {\boldsymbol{V}_p} and {U_{p, q}} . U_{{p^{'}}, q}^ * = \max \left\{ {U_{1, q}^ *, U_{2, q}^ *, U_{3, q}^ * } \right\} , {p^{'}} \in \left\{ {1, 2, 3} \right\} . If the degree of membership between \boldsymbol{Z}_q^l and class {p^{'}} has the largest value, \boldsymbol{Z}_q^l is classified into class {p^{'}} . The feature clustering method based on FCM in NSCT domain for dividing the noise component, weak edges component and strong edges component is summarized in Table 1.

    Table 1.  Feature clustering method based on FCM in NSCT domain.
    Procedure:
    Input: \left\{ {{D_{l, i}}(m, n)} \right\} , the coefficients of high-frequency directional subband at scale l , 1 \leqslant i \leqslant {k_l}
    Step 1. Compute the feature vector {\boldsymbol{Z}^l}(m, n) = {[Z_{mean}^l(m, n), Z_{std}^l(m, n), Z_{\max }^l(m, n)]^T} of the high-frequency directional subband. Change the \left\{ {{\boldsymbol{Z}^l}(m, n)} \right\} as \left\{ {\boldsymbol{Z}_q^l} \right\} .
    Step 2. Initialization: {\boldsymbol{V}_p} , {U_{p, q}} , 1 \leqslant p \leqslant 3 , 1 \leqslant q \leqslant M \times N , 且 \sum\limits_{p = 1}^3 {{U_{p, q}}(0)} = 1 , tem = 0 .
    Step 3. tem = tem + 1 .
    Step 4. Update {\boldsymbol{V}_p} by formula (9).
    Step 5. Update {U_{p, q}} by formula (11).
    Step 6. If tem < 50 and change is larger than 0.001 in {\boldsymbol{V}_p} and {U_{p, q}} , turn to Step 3. Otherwise stop iterate and obtain the optimal membership matrix \boldsymbol{U}_q^ * = {[U_{1, q}^ *, U_{2, q}^ *, U_{3, q}^ *]^T} .
    Step 7. Class \left\{ {{D_{l, i}}(m, n)} \right\} as noise component, weak edges component and strong edges component according to U_q^ * .
    Output: The class attributes of \left\{ {{D_{l, i}}(m, n)} \right\} are obtained.

     | Show Table
    DownLoad: CSV

    We apply NSCT to the abdominal MRI image (Figure 2(a)) and cluster the coefficients of the high-frequency directional subbands by the method described in Table 1. After acquiring the category attributes of the high-frequency coefficients, we set low-frequency coefficients to zero and make NSCT inverse transform only to the coefficients corresponding to noise, weak edges or strong edges respectively. The clustering results of the three high-frequency components will be obtained. Figure 2(b)-(d) are the noise, weak edges or strong edges respectively. Comparing the figures of the three high-frequency components with the original abdominal MRI image, it can be seen that the proposed method classifies the three components in images effectively. Figure 3 gives the modular value distribution of the coefficients after clustering on scale 1. It is observed that high-frequency coefficients at the scale 1 is divided into three different interval segments after FCM feature clustering.

    Figure 2.  Clustering results of the high-frequency coefficients.
    Figure 3.  Histogram distribution of the coefficient modules.

    The high-frequency coefficients {D_{l, i}}(m, n) are classified as noise component, weak edges component and strong edges component by feature clustering in NSCT domain. 0 \leqslant l < L - 1 , 1 \leqslant i \leqslant {k_l} , 1 \leqslant m \leqslant M , We adopt different gain factors to coefficients of different classifications. D_{l, i}^I(m, n) and D_{l, i}^{I'}(m, n) are denote the high-frequency coefficients of the input image I and enhanced image I' respectively. The gain rules are defined as:

    D_{l, i}^{I'}(m, n)' = \left\{ \begin{gathered} {a_0} \bullet {\text{ }}D_{l, i}^I(m, n)&&{\text{ noise component}} \hfill \\ {a_1}\left( {D_{l, i}^I(m, n)/{A_l}} \right) \bullet {A_l}&&{\text{ weak edges component }} \hfill \\ {a_2} \bullet D_{l, i}^I(m, n)&&{\text{ strong edges component}} \hfill \\ \end{gathered} \right. , (12)

    where {A_l} is the maximum modulus of high-frequency coefficients at the scale {2^{ - l}} . Set {a_0} = 0 and the coefficients corresponding to noise are set to 0. The coefficients corresponding to strong edges will be enhanced moderately by adjusting parameter {a_2} . Set {a_2} > 1 . {a_1}(\bullet) is the nonlinear gain function.

    The nonlinear gain function should satisfy monotonicity and antisymmetry. Meanwhile the component with low contrast should be highly enhanced, the sharp edges should be kept and avoid over enhancement. Due to the selection of thresholds included, the classical nonlinear gain function proposed by A. F. Laine needs adjusting several parameters (the generalized adaptive gain method) [28,29]. We simplify the gain function and adopt only one parameter t to control the degree of enhancement. Set t > 1 . The lager the t is set, the higher the degree of enhancement will be. The value of t needs to be chosen properly in order to avoid over enhancement. Figure 4 shows the curves of nonlinear gain function {a_1}(\bullet) . The four curves from the bottom to the top correspond to the setting of t is 5, 6, 7 and 8 respectively. The expression of {a_1}(x) is

    {a_1}(x) = \frac{1}{{1 + {e^{ - t \bullet x}}}} - \frac{1}{{1 + {e^{t \bullet x}}}}, (13)

    where x is the input data and - 1 \leqslant x \leqslant 1 .

    Figure 4.  The curves of nonlinear gain function {a_1}(\bullet) .

    And in our algorithm

    {a_1}(D_{l, i}^I(m, n)/{A_l}) = \frac{1}{{1 + {e^{ - t \bullet (D_{l, i}^I(m, n)/{A_l})}}}} - \frac{1}{{1 + {e^{t \bullet (D_{l, i}^I(m, n)/{A_l})}}}}. (14)

    The low-frequency subband and high-frequency directional subbands are obtained by applying NSCT decomposition to the MRI image. In our enhancement algorithm, we keep the coefficients of low-frequency subband which describe the profile information of images. The coefficients of high-frequency subbands describe the detail information, such as edges, contours and textures, etc. Taking advantages of the different expression of noise, weak edges and strong edges in high-frequency subbands, we classify the coefficients of high-frequency subband by feature clustering. The new high-frequency subbands with modified coefficients will be obtained after setting the noise component as zero, enhancing the weak edges component and adjusting the strong component moderately. Applying NSCT inverse transform to the low-frequency subband of the original image and the newly acquired high-frequency subbands, the final enhanced MRI images will be obtained. The procedure of MRI images enhancement based on feature clustering in NSCT domain is shown in Figure 5. The enhancement algorithm is summarized in Table 2.

    Table 2.  MRI images enhancement based on feature clustering in NSCT domain.
    Procedure:
    Input: An MRI image I(m, n) , 1 \leqslant m \leqslant M , 1 \leqslant n \leqslant N .
    Step 1. Apply L -level NSCT decomposition to image I(m, n) . The low-frequency subband coefficients and high-frequency subbands coefficients {D_{l, i}}(m, n) are obtained. 0 \leqslant l < L - 1 , 1 \leqslant i \leqslant {k_l} .
    Step 2. Use the algorithm expounded in Table 1 to classify the high-frequency subbands coefficients {D_{l, i}}(m, n) into noise component, weak edges component and strong edges component.
    Step 3. Use the gain rules as formula (12) to modify the high-frequency subbands coefficients {D_{l, i}}(m, n) , which have owned the category attributes.
    Step 4. Apply NSCT inverse transform to the low-frequency subband obtained in step 1 and the modified high-frequency subbands coefficients in step 3. The enhanced image I'(m, n) is obtained.
    Output: The enhanced MRI image I'(m, n) .

     | Show Table
    DownLoad: CSV
    Figure 5.  The procedure of the proposed method.

    The test images used in experiments are real MRI images with little noise. The abdominal MRI images are from the CHAOS datasets downloaded at https://chaos.grand-challenge.org/Download/. The brain MRI images are from the Whole Brain Atlas datasets downloaded from http://www.med.harvard.edu/AANLIB/home.html. The size of the test images is 256×256. The methods based on spatial domain, as HE, CLAHE and UM, the GAG parameter adjusting methods [28,29] in wavelet (WT), contourlet (CT), stationary wavelet (SWT) and NSCT domain (NSCT) respectively, and the SMIPC [30] are compared with the proposed enhancement method. The algorithm performs contrast experiments at the PC end of the MATLAB R2018b. The HE, CLAHE UM, SWT adopt the MATLAB built-in function. The HE and UM use the default parameters. The CLAHE uses the common parameters, as 'ClipLimit' = 0.02 and 'Distribution' = 'rayleigh'. The parameters [28] adopted in wavelet (WT), contourlet (CT), stationary wavelet (SWT) and NSCT (NSCT) based enhancement methods are {T_1} = 0.0025 , {T_2} = 0.05 , {T_3} = 0.9 , c = 10 and b = 0.2.

    We apply 4-level decomposition to test images. The wavelet transform and stationary wavelet transform adopt '9-7' wavelet basis function. Contourlet transform and NSCT adopt the '9-7' pyramid decomposition and the McClellan transformed directional filters [30]. From fine scale to coarse scale, the number of high-frequency directional subbands is 8, 8, 4, and 4 respectively. We set {a_2} = {\text{1}}.2{\text{5}} in the gain rule, viz. formula (12). We choose parameter t = 7 in the nonlinear gain function {a_1}(\bullet) , viz. formula (14).

    The number of highfrequency directional subbands subband was chosen based on the following considerations: 1) The number of directional subbands on the fine scale should be higher than that on the coarse scale in order to capture the more details. 2) The number of directional subbands on one scale should not be too much to prevent data redundancy and time loss in data mining. The parameters we choosed in the gain rule according to experiment experience. The two parameters, {a_2} = {\text{1}}.2{\text{5}} , t = 7 , were not adjusted for different MRI images after setting.

    Figures 6, 8, 10, 12 and 14 are the enhancement results of MRI images. HE based method produces obvious whitewashed effect, that reduces the readability of the image. CLAHE based and UM based methods enhance the details of the images, meanwhile enhance the noise in the images. The methods based on WT, CT, SWT and NSCT acquire enhancement results with clear details and smooth uniform regions. The enhancement capacity of the SMIPC for detail is limited. Compared to other methods, our method obtains the most abundant details and highlights the weak edges which hard to observe in the original images (as the middle and lower part of the abdominal cavity and the the right part of the abdominal cavity in Figures 6, 8 and 10, as the middle part of the brain in Figures 12 and 14). The enhancement results we acquired have smooth uniform regions and high clarity.

    Figures 7, 9, 11, 13 and 15 are the histogram distribution of the enhancement results. The MRI images with large difference in intensity distribution. The number of the pixels is zero at the left part of the gray value internal in the histogram distribution of HE results, which makes the enhancement image demonstrating whitewashed effect. CLAHE method extends the grayscale range of the original images. But the outline of histogram distribution of the CLAHE results have large difference compared to that of original image. The outlines of histogram distributions of UM, WT, CT, SWT, NSCT, SMIPC and our results are similar, which indicates that the enhancement results maintain the whole structure of the original images.

    Tables 3-7 give the values of objective evaluation indexes on information entropy (IE), gray mean of the pixels (MV), standard variance (STD), average gradient (AG), grayscale contrast (GC), structural similarity index (SSIM). Information entropy reflects the amount of information contained in the enhanced images. The higher value of the information entropy is, the more information entropy of the enhanced images is. The gray mean of the pixels reflects the average brightness of the images. The gray mean of the pixels will increase after image enhancement. The standard variance reflects the degree of the dispersion that the gray value of pixels deviate from the gray mean. The detail information in the images is more abundant, the standard variance of the image is larger. The average gradient reflects the richness of the texture details and the clarity of the image. The AG is defined as:

    AG = \frac{1}{{MN}}\sum\limits_{i = 1}^M {\sum\limits_{j = 1}^N {\sqrt {\frac{{\Delta I_x^2 + \Delta I_y^2}}{2}} } } , (15)

    where \Delta {I_x} = f(x, y) - f(x - 1, y) f and \Delta {I_y} = f(x, y) - f(x, y - 1) . The larger AG value indicates that the image possesses more abundant detail information and higher definition. The GC reflects the hierarchical change of the image from black to white. The higher the GC value, the stronger the gray expression ability. The GC is defined as:

    GC = \sum\limits_\delta {\delta {{(i, j)}^2}} {P_\delta }(i, j) , (16)

    where \delta (i, j) = \left| {i - j} \right| is a gray value difference of adjacent pixels, {P_\delta }(i, j) is the distribution probability of the pixels with the gray difference of the neighboring pixels at \delta . The SSIM is an index to measure the similarity of two images. When the two images are identical, the value of the SSIM is equal to 1.

    The whitewashed effect and the enhanced noise in the results cause an obviously increase in the values of evaluation indexes of HE, CLAHE and UM results. Compared to the WT, CT, SWT, NSCT and SMIPC methods, our method has a superiority in the indexes of average gradient, grayscale contrast, and is at a middle level on the other indexes. The proposed method bears an advantage is that it can effectively enhance the weak edge components in the image. The prominences of the components make the details of the enhancement results increase and the changes in gray scale increase. So the values obtained by our algorithm can be observed in the AG and GC indicators are relatively high. The UM method does not take into account the noise in the image. In the enhancement process, the noise is also enhanced as a detail component (especially the smoothing region of body in Figure 6), which makes the value of the evaluation indexes high. Our resulting images with the detail components highlighted are different from the original image. While the more similar the experimental results are to the original image, the larger the SSIM value. So our results do not have advantages on SSIM.

    Figure 6.  Enhancement results of the abdomen 1.
    Figure 7.  Histogram distribution of the abdomen 1.
    Figure 8.  Enhancement results of the abdomen 2.
    Figure 9.  Histogram distribution of the abdomen 2.
    Figure 10.  Enhancement results of the abdomen 3.
    Figure 11.  Histogram distribution of the abdomen 3.
    Figure 12.  Enhancement results of the brain 1.
    Figure 13.  Histogram distribution of the brain 1.
    Figure 14.  Enhancement results of the brain 2.
    Figure 15.  Histogram distribution of the brain 2.
    Table 3.  Comparison of the objective evaluation indexes of the abdomen 1.
    Abdomen 1
    IE MV STD AG GC SSIM
    Original Image 6.3739 48.8064 57.1799 5.4466 2.2965 1.0000
    HE 5.0905 133.3824 66.7910 6.8195 1.2442 0.3707
    CLAHE 6.7161 86.4245 57.0363 9.4524 2.6722 0.4272
    UM 6.0738 50.0281 62.6574 11.3711 8.9909 0.8272
    WT 6.1698 50.2151 63.6534 6.4244 3.3694 0.6249
    CT 6.1731 50.3587 63.5974 8.5496 5.4104 0.8295
    SWT 6.1684 50.5205 64.9781 6.2092 3.2312 0.6494
    NSCT 6.1591 50.7755 65.5958 6.1304 3.1810 0.6489
    SMIPC 6.3110 68.6816 69.1470 6.5917 2.6149 0.7737
    NSCT_FC 6.1931 51.1351 66.1048 9.2908 6.5626 0.5133

     | Show Table
    DownLoad: CSV
    Table 4.  Comparison of the objective evaluation indexes of the abdomen 2.
    Abdomen 2
    IE MV STD AG GC SSIM
    Original Image 3.9101 43.4136 64.7740 5.5654 2.7607 1.0000
    HE 2.6748 176.9266 29.8868 2.8715 0.2721 0.2568
    CLAHE 4.0196 60.0482 57.8522 6.3480 2.0043 0.3708
    UM 3.7338 44.0154 70.1693 10.2870 7.0370 0.9059
    WT 5.8572 37.2316 52.2740 6.6213 3.1618 0.3895
    CT 5.9076 35.6707 49.7857 6.4851 3.1220 0.3400
    SWT 5.6056 33.1982 48.2115 5.8744 2.8465 0.7501
    NSCT 5.8461 37.3192 53.4463 6.4139 3.0724 0.7147
    SMIPC 3.8217 56.8934 79.2096 5.7250 3.0118 0.9375
    NSCT_FC 4.8865 47.3493 75.7121 9.2782 6.0097 0.5386

     | Show Table
    DownLoad: CSV
    Table 5.  Comparison of the objective evaluation indexes of the abdomen 3.
    Abdomen 3
    IE MV STD AG GC SSIM
    Original Image 4.1618 47.4574 62.6610 5.4680 2.1883 1.0000
    HE 2.9851 169.7861 34.5959 3.8235 0.4500 0.3236
    CLAHE 4.4257 64.9121 60.0625 7.3549 2.3580 0.3871
    UM 4.1056 47.9313 68.7032 11.0887 6.9016 0.8889
    WT 5.8235 39.2911 52.5718 7.0580 3.4523 0.4230
    CT 5.7223 34.4960 45.7392 6.2431 2.9799 0.3755
    SWT 5.7487 40.2207 54.9499 7.1051 3.4752 0.7850
    NSCT 5.7600 38.3278 52.3462 6.7303 3.2830 0.7717
    SMIPC 4.0827 62.4181 77.9954 5.4587 2.1270 0.9587
    NSCT_FC 5.1642 49.9835 73.6336 10.0034 6.1659 0.5448

     | Show Table
    DownLoad: CSV
    Table 6.  Comparison of the objective evaluation indexes of the brain 1.
    Brain 1
    IE MV STD AG GC SSIM
    Original Image 3.0561 40.0297 62.4801 5.7936 3.3953 1.0000
    HE 2.6469 176.8861 29.9093 3.0011 0.2647 0.2194
    CLAHE 3.8517 55.3079 51.7983 5.8899 2.0572 0.3905
    UM 3.5959 40.8784 68.4163 12.4311 10.8083 0.8632
    WT 4.9340 28.6394 45.0370 6.3359 3.4842 0.4686
    CT 4.9679 29.6385 46.6556 6.5816 3.6033 0.4322
    SWT 4.7813 29.4172 46.5978 6.3766 3.4717 0.8891
    NSCT 4.6050 29.1858 46.2259 6.1893 3.3588 0.8922
    SMIPC 3.0561 53.7925 77.2596 6.2090 4.2998 0.8040
    NSCT_FC 4.1499 42.1759 70.8711 10.5256 7.7725 0.5365

     | Show Table
    DownLoad: CSV
    Table 7.  Comparison of the objective evaluation indexes of the brain 2.
    Brain 2
    IE MV STD AG GC SSIM
    Original Image 3.7625 50.7798 62.9891 7.1894 4.5357 1.0000
    HE 3.2745 163.0556 39.1893 4.9636 0.7210 0.3552
    CLAHE 4.7414 66.6889 55.5672 7.5747 2.8895 0.5019
    UM 4.4331 53.0085 71.1020 14.7573 12.9256 0.8478
    WT 5.8859 35.4493 43.1308 7.0324 4.0154 0.4878
    CT 5.9644 37.8112 46.1570 7.5781 4.3002 0.4552
    SWT 5.8359 36.2405 44.4183 6.9981 3.9809 0.8143
    NSCT 5.8172 35.4688 43.2735 6.7194 3.8077 0.8053
    SMIPC 3.7625 66.4997 77.1337 7.9408 5.3020 0.9413
    NSCT_FC 5.3062 54.3998 72.7941 12.2566 9.6460 0.6038

     | Show Table
    DownLoad: CSV

    MRI images contain abundant edges, contours and texture information. The MRI images with clear object region and high contrast usually can't be obtained due to the limitation of imaging condition, that affects the diagnosis. The visual quality of MRI images should be improved. An MRI images enhancement method based on feature clustering in NSCT domain is proposed. Benefiting from the favorable description ability of the NSCT, according to the different expression forms of noise, weak edges and strong edges in high-frequency subbands, we extract the 3-dimensional feature of high-frequency coefficients in the same scale. The FCM clustering algorithm is introduced to classify the high-frequency coefficients into noise component, weak edges component and strong edges component. The high-frequency coefficients are modified by the gain rules and gain functions designed without adjusting tedious threshold parameters. Compared with the state of art methods based on spatial domain and transform domain, our method has a certain superiority. The enhanced results we obtained possess high clarity, distinct details and smooth uniform area, that will help doctors make diagnosis. The method proposed is an effective and feasible method for MRI image enhancement.

    This work was supported in part by the Ningxia Natural Science Foundation (No. 2022AAC03236), by the National Natural Science Foundation of China (No. 11961001 and No. 61907012), by the First Class Disciplines Foundation of Ningxia (No. NXYLXK2017B09), by the Major Project of North Minzu University (No. ZDZX201801), by the Special Project of North Minzu University (No. FWNX01), by the Master Degree Candidate Innovation Program (No. YCX22106) and by the Hongyang soft project.

    The authors declare that there is no conflict of interests.



    [1] W. S. Hinshaw, P. A. Bottomley, G. N. Holland, Radiographic thin-section image of the human wrist by nuclear magnetic resonance, Nature, 270 (1977), 722-723. https://doi.org/10.1038/270722a0 doi: 10.1038/270722a0
    [2] K. H. Wan, S. H. Shi, C. Z. Shao, Y. B. Liu, M. Li, T. S. Hou, et al., An MRI study of psoas major and abdominal large vessels with respect to the X/DLIF approach, Eur. Spine J., 20 (2011), 557-562. https://doi.org/10.7507/1002-1892.20160291 doi: 10.7507/1002-1892.20160291
    [3] R. H. Wang, J. H. Lv, S. L. Ma, A MRI image segmentation method based on medical semaphore calculating in medical multimedia big data environment, Multimed. Tools Appl., 77 (2018), 9995-10015. https://doi.org/10.1007/s11042-017-4591-3 doi: 10.1007/s11042-017-4591-3
    [4] S. S. Bedi, K. Rati, Various image enhancement techniques-a critical review, Int. J. Adv. Res. Comput. Commun. Eng., 2 (2013), 1605-1609.
    [5] R. Kaur, K. Kaur, Study of Image enhancement techniques in image processing: A review, Int. J. Eng. Manuf., 6 (2016), 38-50. https://doi.org/10.5815/ijem.2016.06.04 doi: 10.5815/ijem.2016.06.04
    [6] K. Singh, A. Seth, H. S. Sandhu, A comprehensive review of convolutional neural network based image enhancement techniques, IEEE Int. Conf. Syst. Comput. Autom. Netw., 2019, 1-6. https://doi.org/10.1109/ICSCAN.2019.8878706 doi: 10.1109/ICSCAN.2019.8878706
    [7] R. Hummel, Image enhancement by histogram transformation, Comput. Graphics Image Process., 6 (1997), 184-195. https://doi.org/10.1016/S0146-664X(77)80011-7 doi: 10.1016/S0146-664X(77)80011-7
    [8] Y. Liu, J. Guo, J. Yu, Contrast enhancement using stratifified parametric-oriented histogram equalization, IEEE T. Circ. Syst. Vid., 27 (2017), 1171-1181. https://doi.org/10.1109/TCSVT.2016.2527338 doi: 10.1109/TCSVT.2016.2527338
    [9] K. Zuiderveld, Contrast limited adaptive histogram equalization, Academic Press Professional, Inc., 1994. https://doi.org/10.1016/B978-0-12-336156-1.50061-6
    [10] J. Sheeba, S. Parasuraman, K. Amudha, Contrast enhancement and brightness preserving of digital mammograms using fuzzy clipped contrast-limited adaptive histogram equalization algorithm, Appl. Soft Comput., 42 (2016), 167-177. https://doi.org/10.1016/j.asoc.2016.01.039 doi: 10.1016/j.asoc.2016.01.039
    [11] B. Vikrant, M. Mukul, U. Shabana, Human visual system based unsharp masking for enhancement of mammographic images, J. Comput. Sci., 21 (2017), 387-393. https://doi.org/10.1016/j.jocs.2016.07.015 doi: 10.1016/j.jocs.2016.07.015
    [12] W. Ye, K. Ma, Blurriness-guided unsharp masking, IEEE T. Image Proces., 27 (2018), 4465-4477. https://doi.org/10.1109/TIP.2018.2838660 doi: 10.1109/TIP.2018.2838660
    [13] S. K. Chandra, M. K. Bajpai, Fractional mesh-free linear diffusion method for image enhancement and segmentation for automatic tumor classification, Biomed. Signal Proces., 58 (2020), 101841. https://doi.org/10.1016/j.bspc.2019.101841 doi: 10.1016/j.bspc.2019.101841
    [14] A. R. Al-Shamasneh, H. A. Jalab, S. Palaiahnakote, U. H. Obaidellah, R. W. Ibrahim, M. T. El-Melegy, A new local fractional entropy-based model for kidney MRI image enhancement, Entropy, 20 (2018), 344. https://doi.org/10.3390/e20050344 doi: 10.3390/e20050344
    [15] Z. Y. Chen, B. R. Abidi, D. L. Page, M. A. Abidi, Gray-level grouping: An automatic method for optimized image contrast enhancement part 1: The basic method, IEEE T. Image Proces., 15 (2006), 2290-2302. https://doi.org/10.1109/TIP.2006.875204 doi: 10.1109/TIP.2006.875204
    [16] A. Polesel, G. Ramponi, V. J. Mathews, Image enhancement via adaptive unsharp masking, IEEE T. Image Proces., 9 (2000), 505-510. https://doi.org/10.1109/83.826787 doi: 10.1109/83.826787
    [17] A. M. Chikhalikar, N. V. Dharwadkar, Model for enhancement and segmentation of magnetic resonance images for brain tumor classification, Pattern Recog. Image, 31 (2021), 49-59. https://doi.org/10.1134/S1054661821010065 doi: 10.1134/S1054661821010065
    [18] L. Cao, H. Q Li, Y. J. Zhang, Retinal image enhancement using low-pass filtering and α-rooting, Signal Proces., 170 (2020), 107445. https://doi.org/10.1016/j.sigpro.2019.107445 doi: 10.1016/j.sigpro.2019.107445
    [19] D. Wei, Z. Wang. Channel rearrangement multi-branch network for image super-resolution, Digit. Signal Proces., 120 (2022), 103254. https://doi.org/10.1016/j.dsp.2021.103254 doi: 10.1016/j.dsp.2021.103254
    [20] D. Wei, Y. M. Li, Convolution and multichannel sampling for the offset linear canonical transform and their applications, IEEE T. Signal Proces., 67 (2019), 6009-6024. https://doi.org/10.1109/TSP.2019.2951191 doi: 10.1109/TSP.2019.2951191
    [21] D. Wei, Linear canonical stockwell transform: Theory and applications, IEEE T. Signal Proces., 70 (2022), 1333-1347. https://doi.org/10.1109/TSP.2022.3152402 doi: 10.1109/TSP.2022.3152402
    [22] D. L. Donoho, A. G. Flesia, Can recent innovations in harmonic analysis 'explain' key findings in natural image statistics? Network-Comp. Neural, 12 (2001), 371-393. https://doi.org/10.1080/net.12.3.371.393 doi: 10.1080/net.12.3.371.393
    [23] A. L. Cunha, J. Zhou, M. N. Do, The nonsubsampled contourlet transform: Theory, design, and applications, IEEE T. Image Proces., 15 (2006), 3089-3101. https://doi.org/10.1109/TIP.2006.877507 doi: 10.1109/TIP.2006.877507
    [24] X. Chang, L. C. Jiao, F. Liu, Y. H Sha, SAR Image despeckling using scale mixtures of gaussians in the nonsubsampled contourlet domain, Chinese J. Electron., 24 (2015), 205-211. https://doi.org/10.1049/cje.2015.01.034 doi: 10.1049/cje.2015.01.034
    [25] M. M. Hasan, M. M. Hossain, S. Mia, A. M. Shmed, R. M. Mahman, A combined approach of non-subsampled contourlet transform and convolutional neural network to detect gastrointestinal polyp, Multimed. Tools Appl., 2022, 9949-9968 https://doi.org/10.1007/s11042-022-12250-2 doi: 10.1007/s11042-022-12250-2
    [26] I. Gath, A. B. Geva, Unsupervised optimal fuzzy clustering, IEEE T. Pattern Anal., 11 (1989), 773-780. https://doi.org/10.1109/34.192473 doi: 10.1109/34.192473
    [27] T. B. Ren, H. H. Wang, H. L. Feng, C. S. Xu, G. S. Liu, P. Ding, Study on the improved fuzzy clustering algorithm and its application in brain image segmentation, Appl. Soft Comput., 81 (2019), 105503. https://doi.org/10.1016/j.asoc.2019.105503 doi: 10.1016/j.asoc.2019.105503
    [28] A. F. Laine, X. Zong, A multiscale sub-octave wavelet transform for de-noising and enhancement, Wavelet Appl. Signal Image Proces. IV, 2825 (1996), 238-249. https://doi.org/10.1117/12.255235 doi: 10.1117/12.255235
    [29] R. E. Glenn, L. Demetrio, M. P. Vishal, Directional multiscale processing of images using wavelets with composite dilations, J. Math. Imaging Vis., 48 (2014), 13-34. https://doi.org/10.1007/s10851-012-0385-4 doi: 10.1007/s10851-012-0385-4
    [30] Z. Al-Ameen, Contrast enhancement of medical images using statistical methods with image processing concepts, IEEE Publisher, Erbil, Iraq, 2020,169-173. https://doi.org/10.1109/IEC49899.2020.9122925
    [31] A. Cohen, I. Daubechies, Non-separable bidimensional wavelet bases, Rev. Mat. Iberoam., 9 (1993), 51-137. https://doi.org/10.4171/RMI/133 doi: 10.4171/RMI/133
  • Reader Comments
  • © 2022 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1664) PDF downloads(52) Cited by(0)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog