Literature DB >> 29712978

Artifact Removal using Improved GoogLeNet for Sparse-view CT Reconstruction.

Shipeng Xie1, Xinyu Zheng2, Yang Chen3,4, Lizhe Xie5, Jin Liu3,4, Yudong Zhang6, Jingjie Yan2, Hu Zhu2, Yining Hu3,4.   

Abstract

Sparse-view Reconstruction can be used to provide accelerated low dose CT imaging with both accelerated scan and reduced projection/back-projection calculation. Despite the rapid developments, image noise and artifacts still remain a major issue in the low dose protocol. In this paper, a deep learning based method named Improved GoogLeNet is proposed to remove streak artifacts due to projection missing in sparse-view CT reconstruction. Residual learning is used in GoogLeNet to study the artifacts of sparse-view CT reconstruction, and then subtracts the artifacts obtained by learning from the sparse reconstructed images, finally recovers a clear correction image. The intensity of reconstruction using the proposed method is very close to the full-view projective reconstructed image. The results indicate that the proposed method is practical and effective for reducing the artifacts and preserving the quality of the reconstructed image.

Entities:  

Year:  2018        PMID: 29712978      PMCID: PMC5928081          DOI: 10.1038/s41598-018-25153-w

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

X-ray Computed Tomography (CT) techniques have been widely utilized in clinical for diagnosis and intervention, including imaging, image-guided needle biopsy, image-guided intervention, and radiotherapy with noticeable benefits. However, with the broadened application of CT in clinical scenarios, the radiation risk issue is receiving more and more attention. As a result, the demand of radiation dose reduction is becoming more and more intense under the principle of ALARA (as low as reasonably achievable). Despite the rapid developments, image noise and artifacts still remains a major issue in the low dose protocol. Balancing image quality and x-ray dose level has become a well-known trade-off problem. Basically, low dose CT can be achieved reducing the tube currents (or voltage) or projection numbers. The approach of tube current (or voltage) reduction sacrifices image quality for dose reduction. Projection number reduction can be realized by applying sparse-view protocol for a given scanning trajectory. CT reconstruction with this approach is termed sparse-view CT reconstruction in this study. Compared to tube current or voltage reduction, sparse-view CT reconstruction does not suffer from the increased noise in projections and has the additional benefit of accelerated scan and projection/back projection calculation. Nevertheless, sparse-view CT reconstruction suffers from image quality deterioration caused by the increased streaking artifacts due to missing projections. A great effort has been devoted to improve sparse-view CT reconstruction in the past twenty years. Specifically, by accommodating measurement statistics, modeling data acquisition geometry, and enforcing physical constraints, regularized iterative reconstruction algorithms often produce superior image quality with highly noisy measurements, and hence having become increasingly popular. In 2006, Donoho proposed the concept of compressed sensing (CS) and proved that sparse signals or piecewise images could be satisfactorily reconstructed from far less sampling data than the requirement of the Nyquist sampling theorem. Base on the CS theory, a state of art solution, which is called as adaptive steepest descent projection onto convex sets (ASD-POCS) method[1], was invented by Sidky et al. by minimizing the total variation (TV) of the desired image for CT image reconstruction from sparse projection views. Recently, a more general term of TV minimisation, called adaptive-weighted total variation (AwTV) model[2], was proposed to improve the preservation of edge details by bringing the local information into the above conventional TV model. To eliminating the patchy artifacts and preserving subtle structures, Liu et al. proposed a total variation-stokes-projection onto convex sets (TVS-POCS) method[3] for the purpose of recovering possible missing information in the sparse-view data situation. Although these TV-based algorithms are successful in a number of cases, the power of the TV minimization constraint is still limited. Besides the TV-based method and its general case, a prior image-constrained compressed sensing (PICCS) method[4] and patch based nonlocal means (NLM)[5], tight wavelet frames, feature dictionary learning[6,7], low rank methods and so on, were introduced to further reduce the number of required projection views by incorporating prior images or patch information to the CS theory. Compared to TV based method, such approaches have the potential of achieving better performance in representing patch-wise structure features and leading to better CT image quality. Recently, Deep Learning techniques have recently been considered to improve CT reconstruction quality. H. C. Burger at al proposes a Multi-Layer Perceptron (MLP) machine based method to learn the mapping from the noisy images to the corresponding noise-free images and obtain an impressive performance in image restoration[8-10]. However, the application of MLP with fully connected layers is often limited by the requirement of fixed input/output size and the weight parameter explosion in network training. J.K. Batenburg and D. Pelt proposed introduce a new CT reconstruction method that improves the filtered back projection method by using a custom data-dependent filter that minimizes the projection error of the resulting reconstruction[11]. Li et al.[12] proposed a dictionary-based sinogram completion method to inpaint the missing sinogram data by applying K-SVD algorithm[13], with database composed by the patches from simulated CT sinogram. Chen et al.[14,15] proposed a new sinogram restoration approach (Sinogram Discriminative Feature Representation) to improve projection data inconsistency. Lee et al.[16] applied convolution neural network (CNN) to interpolate missing data of sinogram for sparse-view CT, by combining with residual learning for better convergence and patch-wisely training the network to avoid memory problem. Würf et al.[17] and Ma et al.[18] mapped FBP algorithm into a deep CNN architecture that allowed a data-driven approach for joint optimization of correction steps in projection domain and image domain. Cheng et al.[19] simulated the iterative process using a DL based leapfrogging strategy. The method was applied to speed up a penalized likelihood PET image reconstruction algorithm, block sequential regularized expectation maximization. In[20], a residual convolutional network architecture was designed to build the relationship between the wavelet coefficients of low-dose and high-dose CT images. Han et al.[21] proposed a U-net structured architecture with residual learning to predict the artifacts in sparse-angle reconstructed CT image. A residual learning of deep CNN method was reported in[22] for image de-noising. Jin et al.[23] proposed a deep convolutional network (FBPConvNet) that combines FBP with a multi-resolution CNN based on Unet[24] and residual learning[25].

Results

Experimental Design

In this section, the well-known filtered back-projection (FBP) reconstruction method is performed and the residual learning is used to remove artifacts generated during sparse-view reconstruction. There are 16000 slices of images for each types of the training data, and 1600 slices of images for each types of the test data. For the training data set, we use the FBP reconstruction using 60 and 120 projection views (full scan) as input x and the difference between the full-view (720 views) reconstruction and the sparse view reconstructions are used as label f. The architectural parameters are described in Table 1.
Table 1

Incarnation of the architecture.

TypePatch size/stride
convolution3 × 3/1
ReLu
convolution64 filters of size 3 × 3/1
BN + ReLu
Inception8 same inception model layers
convolution3 × 3 × 64/1
Incarnation of the architecture. The acquisition parameters of the experimental scans are defined in Table 2. To evaluate the imaging performance of the proposed method under realistic conditions, a set of clinical data and images were used. The image dataset contains 2000 full dose 512*512 CT images. The reference slices were generated using FBP method from 720 projection views. We calculate synthetic projections using fan-beam geometry and projections data were down-sampled to 60 and 120 views to simulate the few-view geometry.
Table 2

acquisition parameters of the experimental scans.

DataThorax
Distance Source to Detector988.00 mm
Distance Source to Patient560.00 mm
Scanner modeHelical
Tube voltage100 KVp
Tube current240 mA
Detector size313.89 mm × 313.89 mm
reconstruction512 × 512 × 640
Volume size0.61 × 0.61 × 0.3125 mm3
acquisition parameters of the experimental scans.

CT data analysis

In this section, sparse view CT reconstruction input images are generated using FBP from 120 (3° angle increment for the tube), and 60 (6° angle increment for the tube) projection views, respectively. The raw data are exported from clinical routine CT examinations. Artifact-free original images are generated by FBP which uses all 720 projection views. The data of the 720 projections are used as full-view projection. We assumed the reconstruction from the full-view data using FBP algorithm as our gold standard image. As we all know, after reducing the number of projections, there is a large number of stripe artifacts in the reconstructed image. As shown in Figs 1 and 2, the column (a) shows the full-view projection. The column (b) shows the sparse reconstruction of the 120-views and 60-views using FBP. The column (c) shows ADS-POCS method [1] of the 120-views and 60-views and the column (d) is our method.
Figure 1

The 512*512 CT images reconstructed by 120 projection views. Column (a) shows the reconstruction by the FBP method from the full projection views. Column (b) shows the reconstruction by the FBP method from the sparse projection views. Column (c) shows the reconstruction by the ADS-POCS method from the sparse projection views. Column (d) shows the reconstruction by the proposed method. All images display in the same window at row 1, 3 and 5. In order to compare the CT value with different reconstruction method, we use 2D atlas to display. The result is shown in row 2, 4 and 6.

Figure 2

The 512*512 CT images reconstructed by 60 projection views. Column (a) shows the reconstruction by the FBP method from the full projection views. Column (b) shows the reconstruction by the FBP method from the sparse projection views. Column (c) shows the reconstruction by the ADS-POCS method from the sparse projection views. Column (d) shows the reconstruction by the proposed method. All images display in the same window at row 1, 3 and 5. In order to compare the CT value with different reconstruction method, we use 2D atlas to display. The results are shown in row 2, 4 and 6.

The 512*512 CT images reconstructed by 120 projection views. Column (a) shows the reconstruction by the FBP method from the full projection views. Column (b) shows the reconstruction by the FBP method from the sparse projection views. Column (c) shows the reconstruction by the ADS-POCS method from the sparse projection views. Column (d) shows the reconstruction by the proposed method. All images display in the same window at row 1, 3 and 5. In order to compare the CT value with different reconstruction method, we use 2D atlas to display. The result is shown in row 2, 4 and 6. The 512*512 CT images reconstructed by 60 projection views. Column (a) shows the reconstruction by the FBP method from the full projection views. Column (b) shows the reconstruction by the FBP method from the sparse projection views. Column (c) shows the reconstruction by the ADS-POCS method from the sparse projection views. Column (d) shows the reconstruction by the proposed method. All images display in the same window at row 1, 3 and 5. In order to compare the CT value with different reconstruction method, we use 2D atlas to display. The results are shown in row 2, 4 and 6. Peak signal-to-noise ratio (PSNR) and stands for structural similarity (SSIM) index is related to evaluate quality of the reconstructed image. PSNR is based on the error between corresponding pixels. The unit of PSNR is dB, the larger value means smaller distortion. SSMI measures image similarity from three aspects of brightness, contrast and structure respectively. The range of SSIM is 0 to 1. Similarly, the larger value represents smaller image distortion. The average PSNR and SSIM between the results and full projection reconstructed images are calculated and shown in Table 3. As shown in Table 3, the PSNR values and the SSIM values have a very pleasant value. The stored data is 12 bits.
Table 3

Average values of PSNR and SSIM between proposed method and full-view FBP reconstruction for 512*512 CT images.

512*512120-view PSNR60-view PSNR120-view SSIM60-view SSIM
FBP29.3224.970.54300.3504
ADS-POCS40.6635.120.95570.8941
Our method 46.8 42.02 0.9880 0.9642
Average values of PSNR and SSIM between proposed method and full-view FBP reconstruction for 512*512 CT images.

Time consumption

This article uses MatConvNet toolkit in the training process, the running environment is MATLAB 2017a. In the train stage, the processor is Intel Xeon E5-2650, the memory is 128 GB. We use one GeForce GTX 1080TI GPU video card for train. In this configuration environment, it takes about 72 hours to train samples over the network for one type of data. In the test stage, the processor is Intel (R) Core (TM) i7 CPU@ 2.2 GHz, the memory is 16GB. In this stage configuration environment, the time consumption is show in Table 4.
Table 4

Time consumption of different methods.

120-view (unit: s)60-view (unit: s)
FBP0.210.10
ADS-POCS16.314.6
Our method 4.1 4.1
Time consumption of different methods.

Discussion

As shown in the zoomed ROI images at the top of Figs 1, 2 and 3, the proposed method displays a very good image quality. Compared to the gold standard image, the FBP result suffered from the artifacts in a high degree. The algorithms can greatly remove the artifacts than the ADS-POCS. The artifacts of reconstructed images after different sparse-view training is greatly removed by the proposed method.
Figure 3

The 256*256 CT images reconstructed by 120 projection views. Column (a) shows the reconstruction by the FBP method from the full projection views. Column (b) shows the reconstruction by the FBP method from the sparse projection views. Column (c) shows the reconstruction by the ADS-POCS method from the sparse projection views. Column (d) shows the reconstruction by the proposed method.

The 256*256 CT images reconstructed by 120 projection views. Column (a) shows the reconstruction by the FBP method from the full projection views. Column (b) shows the reconstruction by the FBP method from the sparse projection views. Column (c) shows the reconstruction by the ADS-POCS method from the sparse projection views. Column (d) shows the reconstruction by the proposed method. As shown in Fig. 3, 256*256 CT images are used to remove the artifacts. The PSNR and SSIM index can be better by using the lower resolution in the experiments. The average PSNR and SSIM between the results and full projection reconstructed images are calculated and shown in Table 5. As shown in Table 5, the PSNR values and the SSIM values have a very pleasant value.
Table 5

Average values of PSNR and SSIM between proposed method and full-view FBP reconstruction for 256*256 CT images.

256*256120-view PSNR60-view PSNR120-view SSIM60-view SSIM
FBP29.926.390.68900.4383
ADS-POCS41.7636.010.97450.9174
Our method 49.67 44.07 0.9920 0.9772
Average values of PSNR and SSIM between proposed method and full-view FBP reconstruction for 256*256 CT images. Does the multi-scale improved GoogLeNet works better than one scale? We use the one scale feed-forward convolutional neural network to compare. The one scale CNN’s flowchart is shown in Fig. 4.
Figure 4

The flowchart of the AFCNN method.

The flowchart of the AFCNN method. As shown in Fig. 5 and Table 6, we can clearly see the multi-scale improved GoogLeNet works better than the one scale CNN.
Figure 5

The image reconstructed by 60-views. Column (a) shows the reconstruction by the FBP method from the full projection views. Column (b) shows the reconstruction with artifacts by the one scale CNN method from the sparse projection views. Column (c) shows the artifacts-free reconstruction by the proposed method.

Table 6

Average values of PSNR and SSIM between proposed method and full-view FBP reconstruction for 256*256 CT images.

120-views PSNR60-views PSNR120-views SSIM60-views SSIM
One scale CNN48.6142.440.99050.9662
Multi-scale CNN 49.67 44.07 0.9920 0.9772
The image reconstructed by 60-views. Column (a) shows the reconstruction by the FBP method from the full projection views. Column (b) shows the reconstruction with artifacts by the one scale CNN method from the sparse projection views. Column (c) shows the artifacts-free reconstruction by the proposed method. Average values of PSNR and SSIM between proposed method and full-view FBP reconstruction for 256*256 CT images. In this paper, we use improved GoogLeNet for the artifacts learning. The sparse-view reconstructed images f is much more like the full view reconstructed images than the artifacts n(especially when the artifacts level is low). Thus, typical mapping function F(f) would be closer to an identity mapping than CNN’s mapping function μ(f), and the artifacts learning for improved GoogLeNet is more suitable for remove artifacts. We develop the GoogLeNet model and use it that outperforms the FBP method, ADS-POCS method and one scale CNN model. FBP method is the classical method. It has been used in the clinical CT system. But in the urtal-sparse view reconstruction, the artifacts were severe and covered all the information using FBP method. Although ASD-POCS preserved some structures, the details were heavily blurred and it is high computational complexity. As shown in Figs 1 and 2, our network preserved the details better than ASD-POCS. It also has the quickly computation speed. The intensity of reconstruction using the proposed method is very close to the full-view projective image. The proposed method approaches to the gold standard image. The results indicate that the improved GoogLeNet algorithm is practical and effective for reducing the streak artifacts caused by projection missing in sparse CT reconstruction and preserving the quality of the sparse-view reconstruction CT image.

Methods

Recently, the CNN has shown great success in handling various tasks. This work focuses on the design and learning of CNN for de-artifacts of the sparse-view CT reconstruction. In this paper, we apply a GoogLeNet[26] (GN) based post-processing approach to remove the artifacts in the sparse-view CT reconstruction. GoogleNet is a deep convolutional neural network architecture that achieves the new state of the art for classification and detection. The most important in GN is the inception modules, which has the multiscale convolutional kernels. We restricted the current incarnations of the inception architecture with different filter sizes and showed it in Fig. 6.
Figure 6

The flowchart of the proposed method.

The flowchart of the proposed method. This paper uses the residual learning for Improved GN to study the artifacts of sparse-view CT reconstruction, and then subtracts the artifacts which obtained by learning from the sparse reconstructed images, finally recovers a clear correction image. The model of sparse-view reconstructed image f = x + n, which include the artifacts n, is similar to the model of image with noise. A mapping function F(f) = x is learned to predict the clear image in the typical de-noising models such as MLP[10]. In this paper, we use the Improved GN to train a mapping function μ(f) ≈ n and then use it to get the residual image n. The averaged mean squared error between the artifacts which is the difference of reconstructed images and the artifacts of network training is used as the loss function to measure the recovery image, and the formula iswhich represents N pairs of images it contains artifacts and real images, μ(f) is the CNN’ mapping function. In order to ensure the quality of the recovery image, it is necessary to train the parameters Θ in the network to obtain the appropriate parameter values so that the mean square error is minimized. The artifacts image is obtained from the Improved GN, and then the final clear recovery image is obtained according to the algorithm x = f − n, which is shown in Fig. 7.
Figure 7

The process of removing artifacts from sparse reconstructed images in CT.

The process of removing artifacts from sparse reconstructed images in CT.
  13 in total

1.  Reducing abdominal CT radiation dose with adaptive statistical iterative reconstruction technique.

Authors:  Priyanka Prakash; Mannudeep K Kalra; Avinash K Kambadakone; Homer Pien; Jiang Hsieh; Michael A Blake; Dushyant V Sahani
Journal:  Invest Radiol       Date:  2010-04       Impact factor: 6.016

2.  Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising.

Authors:  Kai Zhang; Wangmeng Zuo; Yunjin Chen; Deyu Meng; Lei Zhang
Journal:  IEEE Trans Image Process       Date:  2017-02-01       Impact factor: 10.856

3.  Deep Convolutional Neural Network for Inverse Problems in Imaging.

Authors:  Michael T McCann; Emmanuel Froustey; Michael Unser
Journal:  IEEE Trans Image Process       Date:  2017-06-15       Impact factor: 10.856

4.  Improving filtered backprojection reconstruction by data-dependent filtering.

Authors:  Daniel M Pelt; Kees Joost Batenburg
Journal:  IEEE Trans Image Process       Date:  2014-07-22       Impact factor: 10.856

5.  Artifact suppressed dictionary learning for low-dose CT image processing.

Authors:  Yang Chen; Luyao Shi; Qianjing Feng; Jian Yang; Huazhong Shu; Limin Luo; Jean-Louis Coatrieux; Wufan Chen
Journal:  IEEE Trans Med Imaging       Date:  2014-07-10       Impact factor: 10.048

6.  A deep convolutional neural network using directional wavelets for low-dose X-ray CT reconstruction.

Authors:  Eunhee Kang; Junhong Min; Jong Chul Ye
Journal:  Med Phys       Date:  2017-10       Impact factor: 4.071

7.  Discriminative Feature Representation to Improve Projection Data Inconsistency for Low Dose CT Imaging.

Authors:  Jin Liu; Jianhua Ma; Yi Zhang; Yang Chen; Jian Yang; Huazhong Shu; Limin Luo; Gouenou Coatrieux; Wei Yang; Qianjin Feng; Wufan Chen
Journal:  IEEE Trans Med Imaging       Date:  2017-08-14       Impact factor: 10.048

8.  Adaptive-weighted total variation minimization for sparse data toward low-dose x-ray computed tomography image reconstruction.

Authors:  Yan Liu; Jianhua Ma; Yi Fan; Zhengrong Liang
Journal:  Phys Med Biol       Date:  2012-11-15       Impact factor: 3.609

9.  Image reconstruction in circular cone-beam computed tomography by constrained, total-variation minimization.

Authors:  Emil Y Sidky; Xiaochuan Pan
Journal:  Phys Med Biol       Date:  2008-08-13       Impact factor: 3.609

10.  Total variation-stokes strategy for sparse-view X-ray CT image reconstruction.

Authors:  Yan Liu; Zhengrong Liang; Jianhua Ma; Hongbing Lu; Ke Wang; Hao Zhang; William Moore
Journal:  IEEE Trans Med Imaging       Date:  2014-03       Impact factor: 10.048

View more
  16 in total

1.  U-net based metal segmentation on projection domain for metal artifact reduction in dental CT.

Authors:  Mohamed A A Hegazy; Myung Hye Cho; Min Hyoung Cho; Soo Yeol Lee
Journal:  Biomed Eng Lett       Date:  2019-04-29

2.  Learning to Reconstruct Computed Tomography Images Directly From Sinogram Data Under A Variety of Data Acquisition Conditions.

Authors:  Yinsheng Li; Ke Li; Chengzhu Zhang; Juan Montoya; Guang-Hong Chen
Journal:  IEEE Trans Med Imaging       Date:  2019-04-11       Impact factor: 10.048

3.  Enhancement of 4-D Cone-Beam Computed Tomography (4D-CBCT) Using a Dual-Encoder Convolutional Neural Network (DeCNN).

Authors:  Zhuoran Jiang; Zeyu Zhang; Yushi Chang; Yun Ge; Fang-Fang Yin; Lei Ren
Journal:  IEEE Trans Radiat Plasma Med Sci       Date:  2021-12-07

4.  Accurate and robust sparse-view angle CT image reconstruction using deep learning and prior image constrained compressed sensing (DL-PICCS).

Authors:  Chengzhu Zhang; Yinsheng Li; Guang-Hong Chen
Journal:  Med Phys       Date:  2021-09-13       Impact factor: 4.506

5.  Image restoration of motion artifacts in cardiac arteries and vessels based on a generative adversarial network.

Authors:  Fuquan Deng; Qian Wan; Yingting Zeng; Yanbin Shi; Huiying Wu; Yu Wu; Weifeng Xu; Greta S P Mok; Xiaochun Zhang; Zhanli Hu
Journal:  Quant Imaging Med Surg       Date:  2022-05

Review 6.  Deep Learning-Based Image Reconstruction for Different Medical Imaging Modalities.

Authors:  Feng Jinchao; Shahzad Ahmed; Muhammad Yaqub; Kaleem Arshid; Wenqian Zhang; Muhammad Zubair Nawaz; Tariq Mahmood
Journal:  Comput Math Methods Med       Date:  2022-06-16       Impact factor: 2.809

7.  Reconstruction of three-dimensional tomographic patient models for radiation dose modulation in CT from two scout views using deep learning.

Authors:  Juan C Montoya; Chengzhu Zhang; Yinsheng Li; Ke Li; Guang-Hong Chen
Journal:  Med Phys       Date:  2022-01-06       Impact factor: 4.506

Review 8.  Artificial intelligence in paediatric radiology: Future opportunities.

Authors:  Natasha Davendralingam; Neil J Sebire; Owen J Arthurs; Susan C Shelmerdine
Journal:  Br J Radiol       Date:  2020-09-17       Impact factor: 3.039

Review 9.  Applications of Deep Learning to Neuro-Imaging Techniques.

Authors:  Guangming Zhu; Bin Jiang; Liz Tong; Yuan Xie; Greg Zaharchuk; Max Wintermark
Journal:  Front Neurol       Date:  2019-08-14       Impact factor: 4.003

10.  Machine Friendly Machine Learning: Interpretation of Computed Tomography Without Image Reconstruction.

Authors:  Hyunkwang Lee; Chao Huang; Sehyo Yune; Shahein H Tajmir; Myeongchan Kim; Synho Do
Journal:  Sci Rep       Date:  2019-10-29       Impact factor: 4.379

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.