Our study attempts to reduce image artifacts in sparse-view CT images using deep learning. Sparse-view CT reduces the patient exposure dose, and artifact reduction is essential for applying this technique to clinical CT images. Therefore, deep learning with the CGAN model, which is expected to achieve accurate synthetic image generation, was evaluated for artifact correction using simulated sparse-view CT images.
As shown in Figs. 4 and 5, image quality degradation with sparse projection occurred in the simulated sparse-view CT, especially at decimation angles of 5° and 10°, as shown in the subtraction image. In the results of the AE model, restoration of the decoding process was insufficient because of the suppression of the artifact region with a relatively high-contrast resolution component; This was noticeable at the decimation angles of 5° and 10°. The U-Net model reduced artifacts while maintaining the image resolution at a decimation angle of up to 5°. However, partial over-smoothing was observed at the boundary between the adipose and muscle regions. This result shows the same tendency as that reported in previous studies [30, 31]. Moreover, the lung and air regions were different from the original image in the artifact-reduced images of AE and U-Net. The low-density CT value in the lung region tends to be smoothed with an excessively low CT value; therefore, the microvessel structure in the lung vanishes; This can be seen in the reconstructed image in Figs. 4 and 5, even for decimation angles of 1° and 2°. As shown in Tables 1 and 2, the ME and MAE had large differences in the lung and air regions from the original image. The AE and U-Net set the loss function of MSE compared with the training data, and learning with total variation regularization; therefore, these over-smoothing corrections and filling with a uniform value were possibly shown in the artifact reduction images of the AE and U-Net. This result could affect the accuracy of computational analyses using CT values in the image, such as computer-assisted detection/diagnosis and radiation treatment planning. In contrast, the artifacts of sparse projection were properly corrected, and over-smoothing of the pixel value did not occur in the results of the CGAN method. Therefore, the results of ME and MAE were also low values by almost under 5 HU in all regions, and the artifact correction image by the CGAN was suitable for applying the computational analysis image.
Table 3 shows the results of the SSIM and PSNR in sparse projection and correction images by each model. The highest values of SSIM and PSNR were shown in the results of CGAN. The SSIM over 0.8 and the PSNR over 20 dB were accomplished even for the decimation angles of 10 degrees. In the previous study for image correction of sparse-view CT, results of SSIM and PSNR reached about 0.8 and 30 dB by using dual CNN-based methods [32]. Our results of CGAN were comparable with previous studies; therefore, the CGAN model can synthesize accurate images of the sparse-view CT. On the other hand, the results of SSIM and PSNR were significantly degraded by the AE and U-Net models because ME and MAE had large values in these models.
The CGAN model significantly improved the image quality index in terms of SSIM and PSNR compared to the U-Net and AE models. With the addition of conditions and L1 norm regularization, the CGAN significantly improved artifact correction for sparse-view CT and restored the synthetic image close to the original image. In this study, up to a decimation angle of 2°, accurate restoration of image quality, including organ structure and CT values, was achieved using the CGAN model. Over a decimation angle of 5°, the details of the organ structures appear to be transformed, and accurate restoration is limited in the CGAN model. On the other hand, the image similarity index in the SSIM and PSNR were significantly improved using the CGAN correction compared with sparse projection image in all decimated angle cases. Therefore, the accurate restoration of pixel values in the lung region, soft tissues, and bones can enhance the accuracy of image registration using pixel value information and improve the accuracy of calculating the distribution of radiation dose in the radiation therapy. Many groups have explored sinogram synthesis methods based on CNNs in the projection domain and proposed filling in missing view data in sinograms [33, 34]. Our study applied the reconstructed images as training and evaluation data; therefore, artifact correction was performed on the reconstructed CT images. Therefore, because there is no sinogram-based correction, our study has the advantage of not being affected by filter characteristics such as high-frequency enhancement by the FBP. And, since the image reconstruction process can be speeded up with directly corrected in the reconstruction image, it is considered to be more practical in clinical practice. Moreover, implementing artifact correction directly on the reconstructed images is considered more practical and versatile, because the users cannot acquire the sinograms directly from clinical CT scanners. On the other hand, focusing on the details of the tumor contour, the details of the tumor structure were distorted, limiting the complete reconstruction of the structure using the CGAN model.
It is difficult to collect a large number of pixel-by-pixel paired CT images with sparse projections in a clinical CT unit because conventional CT equipment involves continuous rotation data acquisition. Therefore, in this study, many virtual sparse-view CT images were created from sufficiently projected CT images using computational simulation, and deep learning was accomplished using these images. The CGAN needs to add a conditional label using paired images, and the effectiveness of image quality improvement is expected. In this study, the evaluated image of the deep-learning model was an artifact image generated by the computational simulation of sparse projection, and the correction effect for the artifacts caused by the actual sparse projection was not verified. However, because sparse projection is not possible with current clinical CT units, modification of the CT data acquisition system is needed to apply artifact correction methods with deep learning. We believe that our research findings can contribute to reducing radiation exposure and shortening imaging time (by reducing the projection data per phase due to 4D reconstruction) in cone-beam CT images that can be acquired through sparse projection. Our study clarified the effect of image quality improvement for sparse-view CT using three deep learning models and revealed that the CGAN model can synthesize the most similar image, including consistency of CT values. For the clinical application of artifact correction of sparse-view CT images, it is necessary to evaluate the practicality of artifact correction by CGAN by verifying the accuracy of this learning model for actual sparse projection images in future studies.