Research article Special Issues

Iterator-Net: sinogram-based CT image reconstruction


  • Image reconstruction is extremely important for computed tomography (CT) imaging, so it is significant to be continuously improved. The unfolding dynamics method combines a deep learning model with a traditional iterative algorithm. It is interpretable and has a fast reconstruction speed, but the essence of the algorithm is to replace the approximation operator in the optimization objective with a learning operator in the form of a convolutional neural network. In this paper, we firstly design a new iterator network (iNet), which is based on the universal approximation theorem and tries to simulate the functional relationship between the former and the latter in the maximum-likelihood expectation maximization (MLEM) algorithm. To evaluate the effectiveness of the method, we conduct experiments on a CT dataset, and the results show that our iNet method improves the quality of reconstructed images.

    Citation: Limin Ma, Yudong Yao, Yueyang Teng. Iterator-Net: sinogram-based CT image reconstruction[J]. Mathematical Biosciences and Engineering, 2022, 19(12): 13050-13061. doi: 10.3934/mbe.2022609

    Related Papers:

    [1] Huiying Li, Yizhuang Song . Sparse-view X-ray CT based on a box-constrained nonlinear weighted anisotropic TV regularization. Mathematical Biosciences and Engineering, 2024, 21(4): 5047-5067. doi: 10.3934/mbe.2024223
    [2] Si Li, Limei Peng, Fenghuan Li, Zengguo Liang . Low-dose sinogram restoration enabled by conditional GAN with cross-domain regularization in SPECT imaging. Mathematical Biosciences and Engineering, 2023, 20(6): 9728-9758. doi: 10.3934/mbe.2023427
    [3] Si Li, Wenquan Ye, Fenghuan Li . LU-Net: combining LSTM and U-Net for sinogram synthesis in sparse-view SPECT reconstruction. Mathematical Biosciences and Engineering, 2022, 19(4): 4320-4340. doi: 10.3934/mbe.2022200
    [4] Xuefei Deng, Yu Liu, Hao Chen . Three-dimensional image reconstruction based on improved U-net network for anatomy of pulmonary segmentectomy. Mathematical Biosciences and Engineering, 2021, 18(4): 3313-3322. doi: 10.3934/mbe.2021165
    [5] Anam Mehmood, Ishtiaq Rasool Khan, Hassan Dawood, Hussain Dawood . A non-uniform quantization scheme for visualization of CT images. Mathematical Biosciences and Engineering, 2021, 18(4): 4311-4326. doi: 10.3934/mbe.2021216
    [6] Tianran Yuan, Hongsheng Zhang, Hao Liu, Juan Du, Huiming Yu, Yimin Wang, Yabin Xu . Watertight 2-manifold 3D bone surface model reconstruction from CT images based on visual hyper-spherical mapping. Mathematical Biosciences and Engineering, 2021, 18(2): 1280-1313. doi: 10.3934/mbe.2021068
    [7] Jun Liu, Zhenhua Yan, Chaochao Zhou, Liren Shao, Yuanyuan Han, Yusheng Song . mfeeU-Net: A multi-scale feature extraction and enhancement U-Net for automatic liver segmentation from CT Images. Mathematical Biosciences and Engineering, 2023, 20(5): 7784-7801. doi: 10.3934/mbe.2023336
    [8] Yingjian Yang, Qiang Li, Yingwei Guo, Yang Liu, Xian Li, Jiaqi Guo, Wei Li, Lei Cheng, Huai Chen, Yan Kang . Lung parenchyma parameters measure of rats from pulmonary window computed tomography images based on ResU-Net model for medical respiratory researches. Mathematical Biosciences and Engineering, 2021, 18(4): 4193-4211. doi: 10.3934/mbe.2021210
    [9] Dayu Xiao, Jianhua Li, Ruotong Zhao, Shouliang Qi, Yan Kang . Iterative CT reconstruction based on ADMM using shearlet sparse regularization. Mathematical Biosciences and Engineering, 2022, 19(12): 11840-11853. doi: 10.3934/mbe.2022552
    [10] Shuaiyu Bu, Yuanyuan Li, Guoqiang Liu, Yifan Li . MAET-SAM: Magneto-Acousto-Electrical Tomography segmentation network based on the segment anything model. Mathematical Biosciences and Engineering, 2025, 22(3): 585-603. doi: 10.3934/mbe.2025022
  • Image reconstruction is extremely important for computed tomography (CT) imaging, so it is significant to be continuously improved. The unfolding dynamics method combines a deep learning model with a traditional iterative algorithm. It is interpretable and has a fast reconstruction speed, but the essence of the algorithm is to replace the approximation operator in the optimization objective with a learning operator in the form of a convolutional neural network. In this paper, we firstly design a new iterator network (iNet), which is based on the universal approximation theorem and tries to simulate the functional relationship between the former and the latter in the maximum-likelihood expectation maximization (MLEM) algorithm. To evaluate the effectiveness of the method, we conduct experiments on a CT dataset, and the results show that our iNet method improves the quality of reconstructed images.



    Computed tomography (CT) is one of the major imaging techniques that is widely used in clinical practice. Different types of imaging modalities have both similarities and characteristics in their tomographic image reconstruction theories and techniques. Recently, molecular imaging, small animal imaging and the new CT devices have become the focus of international peers. The success of new imaging devices is inseparable from the development of tomographic image reconstruction technology. In the past 40 years, a variety of tomographic reconstruction algorithms have been developed. These algorithms are divided into three categories: mathematical reconstruction algorithms, deep learning reconstruction algorithms and unrolled dynamics (UD) reconstruction method [1].

    The mathematical reconstruction algorithm is modeled according to imaging principles and has a reliable interpretation, but it does not make use of the prior knowledge of the data. Deep learning makes full use of the prior knowledge of the data, but it is not interpretable. The UD reconstruction algorithm combines the advantages of deep learning techniques and traditional reconstruction algorithms. It has been widely used in medical image processing [2,3,4,5,6,7,8]. ISTA-net [9] is a structured deep network that learns a nonlinear transform in an end-to-end manner. This transform replaces the complex transform associated with the proximal mapping. ISTA-NET++, another end-to-end ISTA-unfolding deep network, handles a multi-ratio task using a single model. This kind of method improves the interpretability of the deep learning model and opens up a new way for medical image reconstruction.

    The application of UD methods in CT image reconstruction is also gradually increasing [1,11,12]. Dong et al. proposed the meta-inversion network (MetaInv-Net) [1] which is a representative UD method. Its backbone network architecture is composed of many submodules related to the conjugate gradient (CG) algorithm. Instead of imitating the iterative method, Jin et al. applied a convolutional neural network (CNN) to an inverse problem and proposed FBPConvNet [10]. It is another UD method for CT reconstruction, which is used to explore an advanced CNN architecture. Hu et al. [11] designed a UD method, called learned experts' assessment-based reconstruction network (LEARN). They unfolded an iterative reconstruction scheme into many submodules of a neural network. However, these UD methods all have a common disadvantage. They all simulate regularization operators in optimization problems without the iterative algorithm itself. Therefore, this paper attempts to explore a deep neural network to learn the map of the iterative algorithm itself.

    In this paper, we propose a novel UD method for CT reconstruction, which is fundamentally different from the methods mentioned above. Instead of using neural networks to simulate the approximation operator in iterative algorithms, we design a model to simulate the map between the former and the latter in maximum-likelihood expectation maximization (MLEM) and we call it iNet. We train the network to approximate the real map as closely as possible. This model improves the quality of reconstructed CT images and does not require explicit iterative expressions. This method also has strong generalization performance. Namely, we do not have to study numerical solutions for optimization problems and can use neural networks to learn an iterative algorithm to solve the corresponding optimization problems. This work can provide new inspiration for UD methods.

    The image reconstruction problem is modeled as follows:

    AX=P (1)

    where ARmn denotes the system matrix of the inverse problem, aij represents the element in row i and column j of A, X=[x1,,xn]T denotes the unknown image, and P=[p1,,pm]T denotes the known measured data. Usually, we give an estimate of X, and then the estimate of the corresponding measured value is expressed as Y=[y1,,ym]T, where yi=nj=1aijxj. Supposing P obeys Gaussian distribution with an expectation μi=yi and the same variance σ, we can get the probability density of P distribution:

    f(pi|μi,σ2)=(1σ2π)exp[12σ2(piμi)2] (2)

    The likelihood function is established as follows:

    Prob=mi=1f(piμi,σ2)=(1σ2π)mexp[12σ2mi=1(piμi)2] (3)

    Take the logarithm of this likelihood function:

    ln(Prob)=12σ2mi=1(piμi)2mln(σ2π) (4)

    According to the previous assumptions, we know that the above equation is equivalent to the following equation:

    ln(Prob)=12σ2mi=1(piyi)2mln(σ2π) (5)

    To find the extreme value of ln(Prob), we find the partial derivative of the above Eq (5) with respect to xl,

    ln(Prob)xl=1σ2mi=1[(piyi)yixl] (6)

    Bring yixl=xl[mi=1aijxj]=ail into Eq (6), we get the following formula:

    ln(Prob)xl=1σ2mi=1[(pinj=1aijxj)ail] (7)

    By the corresponding Kuhn-Tucker (KT) conditions, we know that

    xlln(Prob)xl=1σ2xlmi=1[(pinj=1aijxj)ail]=0 (8)

    By the fixed point theory, we achieve the update rule:

    x(k+1)l=x(k)lmi=1piailmi=1ail(nj=1aijx(k)j) (9)

    Then, the final MLEM algorithm is obtained as follows:

    Xk+1=Xk.ATYATAXk (10)

    where Xk and Xk+1 represent the results of the k -th and (k+1) -th iteration of X. To avoid the case where the denominator in Eq (10) is zero, we usually introduce an arbitrary small positive number ε, and obtain the general form of Eq (10) as follows:

    Xk+1=Xk.ATYATAXk+ε (11)

    Deep learning models are used to learn mapping relations in the data space, including linear and nonlinear mapping relations. Initially, it was thought that learning nonlinear mappings required specialized nonlinear models. Fortunately, the universal approximation theorem shows that a feedforward neural network with a linear output layer and at least one layer containing any kind of "squeezed" activation function can approximate a Borel measurable function from one finite dimensional space to another with arbitrary accuracy. Specifically, the neural network built by the hidden layers with arbitrary width and depth approximates continuous functions with arbitrary accuracy [12,13]. Moreover, neural networks with a single hidden layer, as shown in these works [14,16], can accurately approximate any nonlinear continuous function.

    We propose a kind of iterator network to simulate the solution form of the MLEM algorithm and call it iNet. This model improves the quality of reconstructed CT images and does not need explicit iterative expressions. To understand the iNet, we present the conceptual explanation and the architecture of iNet in this section.

    We know that the iterative form of the MLEM algorithm is as follows:

    Xk+1=Xk.ATYATAXk+ε (12)

    Observing the above expression, we obviously find the following conclusions:

    X1=f(X0,ATY)
    X2=f(X1,ATY)
    X3=f(X2,ATY)
    Xk+1=f(Xk,ATY)

    Where f represents the map of the former and the latter. We train the iNet to learn the map f instead of manually computing the explicit expression of the MLEM algorithm.

    Figure 1 shows the detailed network structure of the proposed iNet. It is easy to find that the iNet consists of a convolution part, deconvolution part and skip connections. The convolution part consists of five convolution layers followed by Rectified Linear Unit (ReLU) activation function. The number of convolution kernels is 96, the size of all convolution kernels used is 3 × 3 and the fill of the convolution is "valid". Therefore, the number and the dimension of feature maps after each convolution are 96 and reduced by 4, respectively. The deconvolution part consists of a deconvolution layer, three ReLU activation functions followed by a deconvolution layer, and a deconvolution layer followed by ReLU. We use two skip connections between the convolution and deconvolution, which can avoid the loss of image detail. On the other hand, the skip connection also solves the problems caused by overly deep convolutional layers. It's worth noting that the number of convolution kernels of the final layer is 1. To make the network structure clearer, we explained these symbols of "Adding of pixel level" by adding the Figure 2.

    Figure 1.  Schematic flow chart of the proposed iNet.
    Figure 2.  The explanation of "adding of pixel level".

    To verify the feasibility of our iNet, we conduct experiments on a CT dataset and report qualitative and quantitative results. Firstly, we introduce the dataset in detail. Then, we selected some reconstructed images for display. Finally, we calculate the peak signal-to-noise ratio (PSNR) and structure similarity index measure (SSIM) of the reconstructed image.

    The dataset is stored in the cancer imaging archive (TCIA). It contains clinical CT images of 149 patients scanned using the Lightspeed VCT scanner (GE Healthcare, Waukesha, WI). The data of each patient contains normal-dose and simulated low-dose projection images, normal-dose CT images, and the corresponding clinical information. We used data from 5 patients as the training set and data from 1 patient as the testing set. A scanned image of each patient includes about 300 2D slices, which can be found at the official website of TCIA. These 2D slices were then projected using a systematic matrix forward projection to generate corresponding sinogram data. Then the sinogram data was used to generate corresponding CT images with the MLEM algorithm. Figure 3 shows the schematic diagram of the experimental data generated using a slice. Figure 4 shows the dataset used to train and test the proposed method.

    Figure 3.  Schematic diagram of the experimental data generated using a slice.
    Figure 4.  Experimental data for network training and testing.

    Taking one of the slices as an example, we show the data reconstruction in Figure 3:

    1). For each CT slice of 512 × 512 size, interpolation is performed using resize function to obtain a CT slice of 256 × 256.

    2). The above slice was then projected using a systematic matrix forward projection to generate corresponding sinogram data.

    3). Finally, the CT image was generated using the above sinogram with the MLEM algorithm.

    To evaluate the effect of iNet, we compared the reconstructed image generated by iNet with by the three representative algorithms. They are the alternating direction method of multipliers (ADMM), filtered back projection (FBP) and MLEM.

    To quantitatively compare different CT reconstruction methods, PSNR and SSIM are calculated of the reconstructed CT PSNR is defined as:

    PSNR=10log10(u2maxMSE) (13)

    where umax is the maximum pixel value of the image, and MSE indicates the mean square error between the reconstructed result and the target image. SSIM is defined as:

    SSIM(x,y)=(ux+uy+c1)(2σxy+c2)(u2x+u2y+c1)(σ2x+σ2y+c2) (14)

    where x and y represent the reconstructed image and the target image, ux and uy are the mean values of x, y, σx and σy are the standard deviations of x, y, respectively. c1, c2 are two very small constants to avoid the denominator to be zero. SSIM measures the similarity between two images, and the closer the value is to 1, the more similar the two images are to each other.

    We report and discuss the performance of our iNet according to experimental results. Figure 5 is the visualization of the three reconstructed images. The first column is the real image. The second, third, fourth, and last columns are the results generated by ADMM, FBP, MLEM, and the proposed iNet. The parts indicated by red and blue rectangle boxes represent regions of interest (ROIs) and the corresponding zoomed, respectively. The reconstructed image in Figure 5(b) generated by the ADMM algorithm has obvious blurring. The reconstructed images in Figure 5(c)-(e) have better performance. Compared with other methods mentioned in this work, the proposed method generated more reliable reconstructed images, and we also proved this conclusion with quantitative values.

    Figure 5.  Three CT slices reconstructed by ADMM, FBP, MLEM, and our proposed method, respectively.

    We calculated the PSNR and SSIM of reconstructed images shown in Figure 5 and listed the results in Table 1. It is easy to find that the PSNR and SSIM values of the image reconstructed by iNet are the highest. Additionally, Figure 6-8 show the profiles of the reconstructed images shown in Figure 5. The blue curves correspond to the label (reference) and the orange curves correspond to the different reconstruction methods. It can be seen that the reconstructed images of our iNet are closer to the reference images.

    Table 1.  Quantitative evaluation of different methods at test dataset.
    Frist CT image Second CT image Third CT image
    PSNR SSIM PSNR SSIM PSNR SSIM
    ADMM 25.0550 0.6073 24.2089 0.6974 24.3659 0.7079
    FBP 25.4607 0.6673 23.3797 0.6741 23.2561 0.6380
    MLEM 35.1184 0.9500 31.1246 0.9097 31.5508 0.9214
    iNet 35.7257 0.9590 31.8738 0.9249 31.5954 0.9355

     | Show Table
    DownLoad: CSV
    Figure 6.  Profiles of the first CT slice.
    Figure 7.  Profiles of the second CT slice.
    Figure 8.  Profiles of the third CT slice.

    In this work, we proposed the iNet, which is inspired by the recently popular UD reconstruction method. Guided by the universal approximation theorem, we show that nonlinear functional relationships in iterative reconstruction algorithms can be simulated using a neural network model. The key to this framework is its potential to learn nonlinear maps in iterative reconstruction algorithms, which makes it suitable for extension to any iterative algorithm. In addition, the proposed method in this work brings new ideas to the research of interpretability of deep learning methods.

    As a first attempt, we evaluated the performance of iNET on a clinical CT dataset. This improvement over standard ADMM, FBP, and MLEM is significant and promises to be used for further applications and theoretical studies, which we leave for future work. The applications include the reconstruction of larger size images, and smaller and sparser real data. The theoretical research includes systematizing such methods and proving them in mathematical terms in the future.

    This work was supported by the Fundamental Research Funds for the Central Universities (N2019006 and N180719020).

    The authors declare there is no conflict of interest.



    [1] H. Zhang, B. Liu, H. Yu, B. Dong, MetaInv-Net: meta inversion network for sparse view CT image reconstruction, IEEE Trans. Med. Imaging, 40 (2020), 621–634. https://doi.org/10.1109/TMI.2020.3033541 doi: 10.1109/TMI.2020.3033541
    [2] Y. Wei, M. Zhao, M. Zhao, M. Lei, ADMM-based decoder for binary linear codes aided by deep learning, IEEE Commun. Lett., 24 (2020), 1028–1032. https://doi.org/10.1109/LCOMM.2020.2974199 doi: 10.1109/LCOMM.2020.2974199
    [3] Y. Yang, J. Sun, H. Li, Z. Xu, Deep ADMM-Net for compressive sensing MRI, Adv. Neural Inf. Process. Syst., 29 (2016), 10–18. https://dl.acm.org/doi/10.5555/3157096.3157098
    [4] Y. Yang, J. Sun, H. Li, Z. Xu, ADMM-CSNet: a deep learning approach for image compressive sensing, IEEE Trans. Pattern Anal. Mach. Intell., 42 (2018), 521–538. https://doi.org/10.1109/TPAMI.2018.2883941 doi: 10.1109/TPAMI.2018.2883941
    [5] L. Yang, H. Wang, H. Qian, An ADMM-ResNet for data recovery in wireless sensor networks with guaranteed convergence, Digital Signal Process., 111 (2021), 102956. https://doi.org/10.1016/j.dsp.2020.102956 doi: 10.1016/j.dsp.2020.102956
    [6] J. M. Ramirez, J. I. Martínez-Torre, H. Arguello, LADMM-Net: an unrolled deep network for spectral image fusion from compressive data, Signal Process., 189 (2021), 108239. https://doi.org/10.1016/j.sigpro.2021.108239 doi: 10.1016/j.sigpro.2021.108239
    [7] K. Gong, D. Wu, K. Kim, Y. Yang, G. EI Fakhri, Y. Seo, et al., EMnet: an unrolled deep neural network for PET image reconstruction, Med. Imaging 2019: Phys. Med. Imaging, 10948 (2019), 1203–1208. https://doi.org/10.1117/12.2513096 doi: 10.1117/12.2513096
    [8] Y. Liu, Q. Liu, M. Zhang, Q. Yang, S. Wang, D. Liang, IFR-Net: iterative feature refinement network for compressed sensing MRI, IEEE Trans. Comput. Imaging, 6 (2019), 434–446. https://doi.org/10.1109/TCI.2019.2956877 doi: 10.1109/TCI.2019.2956877
    [9] J. Zhang, B. Ghanem, STA-Net: interpretable optimization-inspired deep network for image compressive sensing, in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018 (2018), 1828–1837. https://doi.org/10.1109/CVPR.2018.00196
    [10] K. Jin, M. T. McCann, E. Froustey, M. Unser, Deep convolutional neural network for inverse problems in imaging, IEEE Trans. Image Process., 26 (2017), 4509–4522. https://doi.org/10.1109/TIP.2017.2713099 doi: 10.1109/TIP.2017.2713099
    [11] H. Chen, Y. Zhang, Y. Chen, J. Zhang, W. Zhang, H. Sun, et al., LEARN: learned experts' assessment-based reconstruction network for sparse-data CT, IEEE Trans. Med Imaging, 37 (2018), 1333–1347. https://doi.org/10.1109/TMI.2018.2805692 doi: 10.1109/TMI.2018.2805692
    [12] G. Cybenko, Approximation by superpositions of a sigmoidal function, Math. Control, Signals Syst., 2 (1989), 303–314. https://doi.org/10.1007/BF02551274 doi: 10.1007/BF02551274
    [13] K. Hornik, M. Stinchcombe, H. White, Multilayer feedforward networks are universal approximators, Neural Networks, 2 (1989), 359–366. https://doi.org/10.1016/0893-6080(89)90020-8 doi: 10.1016/0893-6080(89)90020-8
    [14] T. Chen, H. Chen, Approximations of continuous functionals by neural networks with application to dynamic systems, IEEE Trans. Neural Networks, 4 (1993), 910–918. https://doi.org/10.1109/72.286886 doi: 10.1109/72.286886
    [15] H. Mhaskar, N. Hahm, Neural networks for functional approximation and system identification, Neural Comput., 9 (1997), 143–159. https://doi.org/10.1162/neco.1997.9.1.143 doi: 10.1162/neco.1997.9.1.143
    [16] F. Rossi, B. Conan-Guez, Functional multi-layer perceptron: a non-linear tool for functional data analysis, Neural Networks, 18 (2005), 45–60. https://doi.org/10.1016/j.neunet.2004.07.001 doi: 10.1016/j.neunet.2004.07.001
    [17] D. P. Kingma, J. Ba, Adam: a method for stochastic optimization, preprint, arXiv: 1412.6980.
  • This article has been cited by:

    1. Dariusz Wójcik, Tomasz Rymarczyk, Bartosz Przysucha, Michał Gołąbek, Dariusz Majerek, Tomasz Warowny, Manuchehr Soleimani, Energy Reduction with Super-Resolution Convolutional Neural Network for Ultrasound Tomography, 2023, 16, 1996-1073, 1387, 10.3390/en16031387
  • Reader Comments
  • © 2022 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2658) PDF downloads(138) Cited by(1)

Figures and Tables

Figures(8)  /  Tables(1)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog