Next Article in Journal
Targeting Microtubule-Associated Protein Tau in Chemotherapy-Resistant Models of High-Grade Serous Ovarian Carcinoma
Previous Article in Journal
Comparison of Selected Non-Coding RNAs and Gene Expression Profiles between Common Osteosarcoma Cell Lines
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Generation and Evaluation of Synthetic Computed Tomography (CT) from Cone-Beam CT (CBCT) by Incorporating Feature-Driven Loss into Intensity-Based Loss Functions in Deep Convolutional Neural Network

1
Department of Radiation Oncology, Yonsei Cancer Center, Heavy Ion Therapy Research Institute, Yonsei University College of Medicine, Seoul 03722, Korea
2
Medical Physics and Biomedical Engineering Lab. (MPBEL), Yonsei University College of Medicine, Seoul 03722, Korea
3
Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, TX 75390, USA
4
Oncosoft Inc., Seoul 03787, Korea
*
Authors to whom correspondence should be addressed.
Cancers 2022, 14(18), 4534; https://doi.org/10.3390/cancers14184534
Submission received: 12 August 2022 / Revised: 8 September 2022 / Accepted: 15 September 2022 / Published: 19 September 2022

Abstract

:

Simple Summary

Despite numerous benefits of cone-beam computed tomography (CBCT), its applications to radiotherapy were limited mainly due to degraded image quality. Recently, enhancing the CBCT image quality by generating synthetic CT image by deep convolutional neural network (CNN) has become frequent. Most of the previous works, however, generated synthetic CT with simple, classical intensity-driven loss in network training, while not specifying a full-package of verifications. This work trained the network by combining feature- and intensity-driven losses and attempted to demonstrate clinical relevance of the synthetic CT images by assessing both image similarity and dose calculating accuracy throughout a commercial Monte-Carlo.

Abstract

Deep convolutional neural network (CNN) helped enhance image quality of cone-beam computed tomography (CBCT) by generating synthetic CT. Most of the previous works, however, trained network by intensity-based loss functions, possibly undermining to promote image feature similarity. The verifications were not sufficient to demonstrate clinical applicability, either. This work investigated the effect of variable loss functions combining feature- and intensity-driven losses in synthetic CT generation, followed by strengthening the verification of generated images in both image similarity and dosimetry accuracy. The proposed strategy highlighted the feature-driven quantification in (1) training the network by perceptual loss, besides L1 and structural similarity (SSIM) losses regarding anatomical similarity, and (2) evaluating image similarity by feature mapping ratio (FMR), besides conventional metrics. In addition, the synthetic CT images were assessed in terms of dose calculating accuracy by a commercial Monte-Carlo algorithm. The network was trained with 50 paired CBCT-CT scans acquired at the same CT simulator and treatment unit to constrain environmental factors any other than loss functions. For 10 independent cases, incorporating perceptual loss into L1 and SSIM losses outperformed the other combinations, which enhanced FMR of image similarity by 10%, and the dose calculating accuracy by 1–2% of gamma passing rate in 1%/1mm criterion.

1. Introduction

Image-guided radiotherapy (IGRT) is a viable option in modern radiotherapy [1,2]. Such imaging systems as in-room megavoltage CT, CT-on-rails, and cone-beam CT (CBCT) are available for IGRT [3]. Of those, CBCT equipped with a flat-panel kV detector and kV radiation source became a dominant scanning system over the past decades [4,5,6]. The strength of the imaging system is to provide a 3D volumetric information between patient set-up and actual treatment with relatively low radiation exposure than the other options, followed by facilitating the image matching against the planning CT image.
Over the treatment, the anatomic change derived from a combination of treatment response, weight loss, and radiation effects on normal tissues is inevitable. The changes in internal organs and surface must be non-trivial especially in intensity modulated radiotherapy (IMRT), currently dominant radiation treatment scheme, that try to maximally conform the dose to the target volume, and avoid organ-at-risks (OARs) than the 3D conformal radiotherapy (3D-CRT) [7,8,9]. The notion of on-line and off-line adaptive radiation therapy (ART) [10] was introduced to take preemptive option on the worst case scenario. Volumetric image scanning on a daily fractional basis is also able to monitor the internal anatomic changes, and to help conduct ART in many aspects.
Despite many advantages and potentials above, the limitations of CBCT converge to the image quality. In contrast with conventional fan-beam CT scanners, the cone-shape image scanning increases the unwanted, unexpected scatters reaching to the flat-panel, finally leading to several types of imaging artifacts [11,12]. To overcome the shortcomings in image quality of CBCT, various techniques have been developed. Histogram matching (HM) method [13,14,15] was proposed to enhance dose calculating accuracy. However, matching the HU values of the CT, comprehending the cone-beam artifacts, to those of the planning CT could be problematic. The Monte Carlo (MC) simulation study [16,17,18] and iterative reconstruction method [19,20,21] showed enhancing CBCT image quality for ART at the substantial cost of computational expenses. Although the computation time was significantly reduced with modified MC simulation [22], the additional processing was still demanded to reduce the remaining cone-beam artifacts.
The advent of deep learning based on convolutional neural network (CNN) opened a new prospect in medical image applications, including image segmentation, reconstruction and translation. The image translation, especially, can be applied to enhancing the image quality of CBCT by generating the planning CT-like synthetic image. As most of the planning and dose calculating tasks for radiotherapy are conducted through planning CT images, generating the synthetic CT from CBCT is the rational trial to attempt. Previously, several studies [23,24] developed U-Net based model for generating synthetic CT (sCT) from CBCT with a comparison of image similarity only. Another studies implemented CycleGAN model for the sCT generation providing a promising results [25,26,27,28,29], while CycleGAN may not be an optimal option to take when the paired datasets are available [30].
The previous, recent works, briefly, paid attention to plugging the new network architectures to enhance the performance of reducing the cone-beam artifacts in synthetic CT generation. It did not provide radiation therapy applicable assessment, such as dose calculating accuracy beyond image similarity of the generated synthetic CT to the ground-truth CTs, which may not be sufficient to attain clinical applicability regarding radiotherapy (RT). Many of the studies were also oriented to training the network by intensity-based loss functions accustomed in image translations, possibly undermining the importance of promoting similarity in image features [31,32]. Thus, this study attempted to make contributions to improving the performance in synthetic CT generation as follows:
Incorporating the feature-driven operations into
  • defining the loss function in training the network for synthetic CT generation
  • assessing the image similarity for the generated image
Offering a full package of verifications for the generated synthetic CT images from CBCT with respect to image similarity throughout dosimetry accuracy
More specifically, this work combined a feature-driven loss function [33,34] with the intensity-based losses, such that it can promote the image feature similarity of the generated images in addition to the anatomical similarity (Section 2.1, Section 2.2 and Section 2.3), as implemented in some of the previous works for the image reconstruction and MR-to-CT image translation [35,36]. In evaluating image similarity, besides the conventional metrics, it employed a feature-oriented metrics to quantify the degree of similarity in image feature (Section 2.4). In an effort to provide a full verifying procedure, the dosimetry accuracy was also investigated for the synthetic CT images by dose calculation throughout a commercial Monte-Carlo algorithm (Section 2.4). Overall, this study was to explore which combination of loss functions for network training would yield the clinically feasible synthetic CT images from CBCT images for RT. To achieve so, it also focused on acquiring well-paired, consistent CT and CBCT images by deformable image registration and by obtaining a pair of CT-CBCT images from a single simulator and treatment unit.

2. Materials and Methods

2.1. Dataset

The patient cohort for DL model consisted of 65 brain and HN cancer patients with CBCT and CT images, where the patients were scanned in Canon Aquilion LB CT simulator (Canon Medical Systems Corporation, Otawara, Japan), and treated by Elekta Infinity (Elekta, Stockholm, Sweden) linear accelerators from January in 2019 to December in 2021. 52 cases of those were the patients with skull and brain cancers, while the remaining cases were the head-and-neck cancers. Table 1 specifies the characteristics of data used in this work. The 65 scans from 65 patients were divided into 50, 5 and 10 for training, validating, and testing the network, respectively. The input and output images of the network were the paired images of CBCT and planning CT images, which plays a role in generating synthetic CT from CBCT images. This work took a special care of pairing CT and CBCT images and maintaining the data consistency. Namely, to achieve so, we acquired the CT and CBCT images that were scanned and treated at the same imaging simulator and the same treatment unit. Thus, it adopted a general deep neural network, in contrast with cycling GAN architecture suited for unpaired dataset. Hence, he planning CT was deformably registered and resampled to CBCT, named deformed CT (dCT), such that it can enhance the image structural similarity between CBCT and planning CT images.

2.2. Loss Functions on FC DensNet

In generating the synthetic CT from CBCT, the training model used a couple of combinations of different loss functions: L1-loss, perceptual loss, and structural similarity (SSIM) loss, as illustrated in Figure 1A. The loss functions used in this work could be split into two types, where the L1 and SSIM loss referred to the anatomical difference in enhancing image similarity. Contrarily, the perceptual loss penalizes and updates the weights in network by comparing features on the images.
The L1-loss function, most frequently used function in DL models with images, measures the pixel-wise mean absolute difference between the output of the DL model and ground-truth. The L1 Loss is expressed in (1):
L L 1 x , G x = x G x 1
where x is the true planning CT image, and G(x) is the synthetic CT image generated from CBCT.SSIM has been widely employed as one of the metrics to evaluate image quality. It has the characteristic of preserving image contrast and luminance better than other losses. The SSIM loss is defined as in the following:
L S S I M x , G x = 2 μ x μ G x + c 1 2 σ x G x + c 2 μ x 2 + μ G x 2 + c 1 σ x 2 + σ G x 2 + c 2
where μ𝑥 is the average of x, μ𝐺(𝑥) is the average of G(x), σ𝑥 2 is the variance of x, σ𝐺(𝑥) 2 is the variance of G(x), and 𝜎𝑥𝐺(𝑥) is the covariance of x and G(x), 𝑐 is the stabilization variable. The perceptual loss function was first introduced as a VGG loss based on a pre-trained VGG network. The perceptual loss compensates for the perceptually unsatisfactory results of pixel-wise losses such as L1 Loss. For this, perceptual loss calculates the euclidean distance between feature maps extracted from a pre-trained VGG network. The definition of perceptual loss is given in (3):
L P e r c e p t u a l x , G x = V G G x V G G G x 2 2
The loss functions defined above were plugged into one of the fully convolutional network (FCN) structures, FC DenseNet [37]. The DenseNet was employed in this application as it t has shown a promising result for the paired image dataset. FCN is similar to common CNN, while the fully connected layers are removed from the end of the network and the output is generated by combining the output of pooling layers from different convolutional ones. As shown in Figure 1B, the FC-DenseNet consisted of down-sampling path, up-sampling path. Specifically, in down-sampling path, following the convolution layer, the transition down layers consists of batch normalization, exponential linear units (ELU) [38], 1 × 1 convolution, dropout (p = 0.2), and a 2 × 2 max-pooling operation. In up-sampling path, the transition up layers, before the convolution layer, consist of transposed convolution.

2.3. Implementation

The deformable image registration (DIR) between CBCT (target) and planning CT (moving) images was performed on MIM software (Mim Software Inc., Cleveland, OH, USA) and MATLAB (The MathWorks, Inc., Natick, MA, USA) with deformation and resampling, the resolution and imaging size of deformed planning CT image (dCT) are identical to those of CBCT (270 × 270 with 1 mm resolution). The FC-DenseNet was implemented using Python 3.8.3 and PyTorch 1.5.1 [39]. The training was performed on graphical processing units (GPUs) (NVIDIA TITAN RTX with 24 GB of memory). Each model was trained under the identical hyper-parameter setting, while varying the definition of loss functions. The number of epochs was 150. The DL architecture had an initial learning rate of 0.00002 and the architecture’s weights were optimized using Adam [40]. During the training, data augmentations (horizontal flip, random rotation, random blur) were applied on the fly randomly.
From our observations, the image similarity was enhanced when the perceptual and/or SSIM losses were combined with L1-loss. Thus, we composed the loss function by addition of L1-loss to perceptual loss and/or SSIM loss. The combinations exploited in this work were defined as follows: L1-loss only (L1), L1 + Perceptual loss (LP), L1 + SSIM loss (LS), and L1 + Perceptual + SSIM loss (LPS). In this work, the weights for the different losses were well-balanced as imposing unbalanced regularizing weights led to poorer performance from our observations.

2.4. Evaluation

2.4.1. Image Similarity

To evaluate the image similarity, the dCT images were used as the ground truth to evaluate the four different loss combination models. We compared the image similarity between ground-truth dCT and synthetic CT images in terms of such conventional similarity metrics as mean-absolute error (MAE), SSIM, and peak-signal-to-noise-ratio (PSNR).
Recently, interest in features of tomographic images such as CT, magnetic resonance (MR), or positron emission tomography (PET) images has increased [41]. There are SIFT, KAZE, ORB, etc. methods for mapping features from images. Although the above image similarity metrics are good evaluations, one step further, we propose a feature mapping ratio (FMR) based on the A-KAZE feature mapping algorithm [42], an algorithm that improved the SIFT algorithm, as an image similarity comparison metric. Figure 2 shows how FMR processes the feature points in the paired sCT and dCT images. It extracts feature points in each sCT and dCT images, following calculates binary descriptors, and matches descriptors. Next, the image quality was compared through the ratio between the detected features and the matched features. Specifically, in the step of extracting feature points, it computes a nonlinear scale space to extract feature points, generating a robust binary descriptor that exploits gradient information from the nonlinear scale space. In the following descriptor matching, a brute-force algorithm is used to match descriptors with a hamming distance less than a threshold, which was set to be 0.8 in this work.

2.4.2. Dose Calculation

We had an optimized VMAT plan for each testing case, designed to treat the patient on the given CT image. As the deformed CT images (dCT) were slightly different in registration, the same planning parameters were applied to the dCTs. Hence, the calculated dose on dCT was defined as the ground-truth dose distribution. The dose calculation was performed by MONACO treatment planning system from Elekta, in which the commercial Monte-Carlo (MC) dose calculation is available. The dose distributions on the synthetic CT images resulted from the different loss functions were calculated by the commercial MC algorithm with the same VMAT planning parameters as those for the dCT images. The dosimetric similarity was quantified by absolute dose difference, and gamma passing rate on 1%/1 mm, and 2%/2 mm criteria.

3. Results

3.1. Image Similarity Comparison

Figure 3 illustrates a couple of exemplary synthetic CT images of the 10 testing cases, generated from different loss functions. It shows that the synthetic CT images produced from LS and LPS loss functions tended to be analogous to the ground-truth CT images. As indicated by arrows, the synthetic CT images from the SSIM-associated loss functions well simulated the structural details of the true CT images, relative to those from L1-loss and LP loss functions.
Table 2 lists the numerical details of image similarity between the dCT and sCT images for the 10 testing cases. The synthetic CT images generated from different loss functions resulted in about 6 HU distance on average against the ground-truth CT images, which were fairly good. In SSIM, MAE, and PSNR, it turns out that the LPS loss produced slightly more accurate synthetic CT images than the other loss functions though the differences were not significant. In feature matching throughout the FMR, the difference across different loss function definitions became more explicit. Combining the SSIM loss with L1-loss increased the FMR from 0.532 to 0.569. With all loss function (L1, SSIM, and perceptual losses) combined, FMR reached out to 0.683. The perceptual loss resulted in the worse image similarity in the conventional metrics, even relative to L1-loss only. It is interesting to note that the LP loss produced higher value than L1-loss on average in feature matching. It could be associated with the fact that the perceptual loss was designed to put more emphasis on the feature representation.

3.2. Dose Distribution Comparison

Table 3 shows the dosimetric analysis with gamma passing rate and absolute dose difference across the 10 cases. It compared the reference and compute dose distributions calculated on dCT and sCTs from different loss functions. Though LP loss function produced greater gamma passing rate than L1-loss in 2%/2 mm criterion of gamma passing rate, LPS was followed by LP, L1 and LS in dosimetric accuracy. It was obvious that the dose distribution computed on the synthetic CT images from LPS had the lowest error to the ground-truth dose distribution with no exception. In 1%/1 mm criterion of gamma passing rate, the LPS loss function had 96.2% passing rate on average, which was about 1% and 1.5% greater than those from the L1 and LP loss functions.
Figure 4 illustrates the dosimetric comparison between dose distributions on dCT and sCTs. From absolute dose differences in the second row of Figure 4, the magnitude of dosimetric errors got smaller from left to right (from L1-loss to LPS loss functions). In addition, in the third row of Figure 4 representing gamma passing rate, it was seen that the region of the errors became narrower on the synthetic CT images from SSIM-associated loss functions.

4. Discussion

CBCT has many desirable features suitable for IGRT, whereas the degraded image quality relative to the planning CT images were pitfalls. The source of the degeneracy was mainly caused by cone-shape imaging with flat-panel detector, which was inherently susceptible to the photon scattering. The so-called cone-beam artifact reduction was made easier with an aid of an emerging framework, CNNs with deep learning. Many studies attempted to enhance the CBCT image quality by generating the synthetic CT images from CBCT images throughout deep convolutional neural networks. It was found that most of the studies applied the new network architectures to synthetic CT generation, while a few studies provided a full package including image similarity and dosimetric analysis to see if the algorithm is applicable to the clinic. In addition, most of the works utilized the intensity-based loss function in training the given network, overlooking the possibility of promoting similarity in image features.
To differentiate from the previous studies, this work derived the feature-driven quantifications in defining loss function by perceptual loss and evaluating image similarity by the means of feature mapping ratio (FMR). In training the network, in addition to the perceptual loss, we diversified the loss function with SSIM loss combined with L1-loss, which could also strengthen the anatomical and structural similarities. In assessing the dosimetric accuracy, the dose calculation was performed on the MC-based algorithm with actual VMAT planning parameters. To make it fully controlled, the other environmental variables were constrained, such that all CBCT and CT images used for training and testing the network were obtained from the only one RT treatment unit and the same CT simulator.
As seen in the results, the SSIM-associated loss function (L1 + SSIM, and L1 + perceptual + SSIM) produced the most similar synthetic CT images to the ground-truth CT images with respect to the image-similarity metrics. The perceptual loss did not lead to better results than L1-loss only in conventional similarity metrics. Contrarily, the synthetic CT images from LP (L1 + perceptual) loss were greater than those from L1-loss only on average. This implies that the feature driven training affected the feature-mapping accuracy constructively. The feature-based perceptual loss was demonstrated to be powerful if it is combined with secure structural similarity, in which the LPS loss (L1 + perceptual + SSIM)-based training enlarged the FMR significantly without compromising the structural similarity as carried out by the LP loss. The similar tendency was appeared in analyzing dosimetric accuracy. The synthetic CT images generated by the LPS loss outperformed the other images from different combinations in both gamma passing rate and absolute dose differences. The synthetic CT images generated from CBCT resulted in 96.2% and 99.6% gamma passing rates against the reference dose distribution in 1%/1 mm, and 2%/2 mm.
The low-grade image quality of CBCT has been able to affect the accuracy of IGRT. The reduction in cone-beam artifacts directly facilitates the image matching procedure in IGRT. Especially, the abovementioned results regarding dosimetric accuracy would be able to extend the border of CBCT. To be more specific, by generating synthetic CT from CBCT, the dose calculation on the synthetic CT would become available, which are potentially used for evaluating dose distribution and summation on a fraction basis. In addition, it could reduce the necessity of CT re-simulation by possible substitute CT images for synthetic CT images from CBCT. The synthetic CT images having greater image contrast than CBCT with reduced imaging artifacts would be beneficial for the deep learning-based auto-segmentation as well. Hence, these factors mostly associated with enhanced efficiency could eventually facilitate the realization of adaptive radiation therapy (ART).
Despite various advantages delivered from this work, there are a couple of limitations to be stated. To stand out the benefits of this work, we constrained the variability of CT/CBCT image data. From our perspectives, however, it was more important to examine which combination of loss functions can produce more qualified results with a constrained dataset. Another potential limitation might have been increase in inefficiency due to additional loss functions in the network training. In fact, the training time was increased by about 60% with additional 0.6 GB GPU VRAM usage when accompanying the perceptual loss that requires for running the VGG network. With respect to the inference after completing the training, which is considered more important in clinical aspects, there was no such a big difference in time, presenting only 1 s (5 s vs. 6 s). The body sites that we referred to were brain and head-and-neck regions, which were covered by CBCT images. In upper abdomen and pelvic regions, the maximum FOV of CBCT may not cover whole region of interest. In such conditions, the CBCT images were further influenced by the additional scattering effect adjacent to the marginal side of images. Last, the synthetic CT images from CBCT were to reduce so-called cone-beam artifacts, while it possibly has an inherent CT-possessing artifacts. There is room, thus, to further enhance the image quality and CBCT-based (adaptive) radiotherapy.

5. Conclusions

This study investigated the generation of synthetic CT from CBCT to reduce the cone-beam artifacts, thus enhancing the image quality of CBCT. We varied the definition of loss functions combining L1-loss with intensity and shape-based SSIM, and feature-based perceptual losses for the well registered, paired CBCT-CT dataset. With evaluating metrics including image similarity by feature mapping criterion, and dosimetric accuracy for the MC-simulated dose distributions, the SSIM-associated loss functions produced the qualified synthetic CT images. When incorporating the perceptual loss into L1- and SSIM losses, the resulting synthetic CT images yielded the best performance in both image similarity and dose calculating accuracy. The results would support a claim that CBCT, once being developed to reduce the artifacts, could be employed for radiation therapy in more constructive ways, such as for adaptive radiation therapy.

Author Contributions

Conceptualization, S.K.Y., H.K. and J.S.K.; methodology, S.K.Y., H.K. and J.S.K.; e-software, S.K.Y. and H.K.; validation, S.K.Y. and H.K.; formal analysis, S.K.Y. and H.K.; investigation, S.K.Y. and H.K.; resources, H.K. and J.S.K.; data curation, H.K. and J.S.K.; writing—original draft preparation, S.K.Y. and H.K.; writing—review and editing: B.S.C. and I.P., All authors; visualization, S.K.Y. and H.K.; supervision, H.K. and J.S.K.; project administration, H.K. and J.S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a National Research Foundation of Korea Grant funded by the Korean government (NRF-2021R1A2C2005824), and a Faculty Research Grant of Yonsei University of College of Medicine (6-2020-0125).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Review Board of the Severance Hospital (IRB #4-2022-0368).

Informed Consent Statement

The informed consent form was waived due to the retrospective nature of the study.

Data Availability Statement

The original contributions presented in the study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guckenberger, M.; Meyer, J.; Vordermark, D.; Baier, K.; Wilbert, J.; Flentje, M. Magnitude and clinical relevance of translational and rotational patient setup errors: A cone-beam CT study. Int. J. Radiat. Oncol. Biol. Phys. 2006, 65, 934–942. [Google Scholar] [CrossRef] [PubMed]
  2. Verellen, D.; Ridder, M.D.; Linthout, N.; Tournel, K.; Soete, G.; Storme, G. Innovations in image-guided radiotherapy. Nat. Rev. Cancer 2007, 7, 949–960. [Google Scholar] [CrossRef] [PubMed]
  3. Schwartz, D.L.; Garden, A.S.; Shah, S.J.; Chronowski, G.; Sejpal, S.; Rosenthal, D.I.; Chen, Y.; Zhang, Y.; Zhang, L.; Wong, P.-F. Adaptive radiotherapy for head and neck cancer—dosimetric results from a prospective clinical trial. Radiother. Oncol. 2013, 106, 80–84. [Google Scholar] [CrossRef] [PubMed]
  4. Tuy, H.K. An inversion formula for cone-beam reconstruction. SIAM J. Appl. Math. 1983, 43, 546–552. [Google Scholar] [CrossRef]
  5. Cho, P.S.; Johnson, R.H.; Griffin, T.W. Cone-beam CT for radiotherapy applications. Phys. Med. Biol. 1995, 40, 1863. [Google Scholar] [CrossRef]
  6. Jaffray, D.A.; Siewerdsen, J.H.; Wong, J.W.; Martinez, A.A. Flat-panel cone-beam computed tomography for image-guided radiation therapy. Int. J. Radiat. Oncol. Biol. Phys. 2002, 53, 1337–1349. [Google Scholar] [CrossRef]
  7. Hong, T.S.; Tomé, W.A.; Chappell, R.J.; Chinnaiyan, P.; Mehta, M.P.; Harari, P.M. The impact of daily setup variations on head-and-neck intensity-modulated radiation therapy. Int. J. Radiat. Oncol. Biol. Phys. 2005, 61, 779–788. [Google Scholar] [CrossRef]
  8. Hector, C.; Webb, S.; Evans, P. The dosimetric consequences of inter-fractional patient movement on conventional and intensity-modulated breast radiotherapy treatments. Radiother. Oncol. 2000, 54, 57–64. [Google Scholar] [CrossRef]
  9. Heffernan, P.B.; Robb, R.A. Image reconstruction from incomplete projection data: Iterative reconstruction-reprojection techniques. IEEE Trans. Biomed. Eng. 1983, BME-30, 838–841. [Google Scholar] [CrossRef]
  10. Castadot, P.; Lee, J.A.; Geets, X.; Grégoire, V. Adaptive radiotherapy of head and neck cancer. Semin. Radiat. Oncol. 2010, 20, 84–93. [Google Scholar] [CrossRef]
  11. Nagarajappa, A.K.; Dwivedi, N.; Tiwari, R. Artifacts: The downturn of CBCT image. J. Int. Soc. Prev. Community Dent. 2015, 5, 440. [Google Scholar]
  12. Tadinada, A.; Jalali, E.; Jadhav, A.; Schincaglia, G.P.; Yadav, S. Artifacts in Cone Beam Computed Tomography Image Volumes: An Illustrative Depiction. J. Mass. Dent. Soc. 2015, 64, 12–15. [Google Scholar] [PubMed]
  13. Arai, K.; Kadoya, N.; Kato, T.; Endo, H.; Komori, S.; Abe, Y.; Nakamura, T.; Wada, H.; Kikuchi, Y.; Takai, Y. Feasibility of CBCT-based proton dose calculation using a histogram-matching algorithm in proton beam therapy. Phys. Med. 2017, 33, 68–76. [Google Scholar] [CrossRef] [PubMed]
  14. Abe, T.; Tateoka, K.; Saito, Y.; Nakazawa, T.; Yano, M.; Nakata, K.; Someya, M.; Hori, M.; Sakata, K. Method for converting cone-beam CT values into Hounsfield units for radiation treatment planning. Int. J. Med. Phys. Clin. Eng. Radiat. Oncol. 2017, 6, 361–375. [Google Scholar] [CrossRef]
  15. Kidar, H.S.; Azizi, H. Enhancement of Hounsfield unit distribution in cone-beam CT images for adaptive radiation therapy: Evaluation of a hybrid correction approach. Phys. Med. 2020, 69, 269–274. [Google Scholar] [CrossRef]
  16. Zbijewski, W.; Beekman, F.J. Efficient Monte Carlo based scatter artifact reduction in cone-beam micro-CT. IEEE Trans. Med. Imaging 2006, 25, 817–827. [Google Scholar] [CrossRef]
  17. Mainegra-Hing, E.; Kawrakow, I. Variance reduction techniques for fast Monte Carlo CBCT scatter correction calculations. Phys. Med. Biol. 2010, 55, 4495. [Google Scholar] [CrossRef]
  18. Bootsma, G.; Verhaegen, F.; Jaffray, D. Efficient scatter distribution estimation and correction in CBCT using concurrent Monte Carlo fitting. Med. Phys. 2015, 42, 54–68. [Google Scholar] [CrossRef]
  19. Wang, J.; Li, T.; Xing, L. Iterative image reconstruction for CBCT using edge-preserving prior. Med. Phys. 2009, 36, 252–260. [Google Scholar] [CrossRef]
  20. Jia, X.; Dong, B.; Lou, Y.; Jiang, S.B. GPU-based iterative cone-beam CT reconstruction using tight frame regularization. Phys. Med. Biol. 2011, 56, 3787. [Google Scholar] [CrossRef]
  21. Gardner, S.J.; Mao, W.; Liu, C.; Aref, I.; Elshaikh, M.; Lee, J.K.; Pradhan, D.; Movsas, B.; Chetty, I.J.; Siddiqui, F. Improvements in CBCT image quality using a novel iterative reconstruction algorithm: A clinical evaluation. Adv. Radiat. Oncol. 2019, 4, 390–400. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Xu, Y.; Bai, T.; Yan, H.; Ouyang, L.; Pompos, A.; Wang, J.; Zhou, L.; Jiang, S.B.; Jia, X. A practical cone-beam CT scatter correction method with optimized Monte Carlo simulations for image-guided radiation therapy. Phys. Med. Biol. 2015, 60, 3567. [Google Scholar] [CrossRef] [PubMed]
  23. Chen, L.; Liang, X.; Shen, C.; Nguyen, D.; Jiang, S.; Wang, J. Synthetic CT generation from CBCT images via unsupervised deep learning. Phys. Med. Biol. 2021, 66, 115019. [Google Scholar] [CrossRef] [PubMed]
  24. Yuan, N.; Dyer, B.; Rao, S.; Chen, Q.; Benedict, S.; Shang, L.; Kang, Y.; Qi, J.; Rong, Y. Convolutional neural network enhancement of fast-scan low-dose cone-beam CT images for head and neck radiotherapy. Phys. Med. Biol. 2020, 65, 035003. [Google Scholar] [CrossRef]
  25. Liang, X.; Chen, L.; Nguyen, D.; Zhou, Z.; Gu, X.; Yang, M.; Wang, J.; Jiang, S. Generating synthesized computed tomography (CT) from cone-beam computed tomography (CBCT) using CycleGAN for adaptive radiation therapy. Phys. Med. Biol. 2019, 64, 125002. [Google Scholar] [CrossRef]
  26. Zhang, Y.; Yue, N.; Su, M.Y.; Liu, B.; Ding, Y.; Zhou, Y.; Wang, H.; Kuang, Y.; Nie, K. Improving CBCT quality to CT level using deep learning with generative adversarial network. Med. Phys. 2021, 48, 2816–2826. [Google Scholar] [CrossRef]
  27. Maspero, M.; Houweling, A.C.; Savenije, M.H.; van Heijst, T.C.; Verhoeff, J.J.; Kotte, A.N.; van den Berg, C.A. A single neural network for cone-beam computed tomography-based radiotherapy of head-and-neck, lung and breast cancer. Phys. Imaging Radiat. Oncol. 2020, 14, 24–31. [Google Scholar] [CrossRef]
  28. Deng, L.; Hu, J.; Wang, J.; Huang, S.; Yang, X. Synthetic CT generation based on CBCT using respath-cycleGAN. Med. Phys. 2022, 49, 5317–5329. [Google Scholar] [CrossRef]
  29. Zhang, Y.; Ding, S.-g.; Gong, X.-c.; Yuan, X.-x.; Lin, J.-f.; Chen, Q.; Li, J.-g. Generating synthesized computed tomography from CBCT using a conditional generative adversarial network for head and neck cancer patients. Technol. Cancer Res. Treat. 2022, 21, 15330338221085358. [Google Scholar] [CrossRef]
  30. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 29 October 2017; pp. 2223–2232. [Google Scholar]
  31. Vulli, A.; Srinivasu, P.N.; Sashank, M.S.K.; Shafi, J.; Choi, J.; Ijaz, M.F. Fine-Tuned DenseNet-169 for Breast Cancer Metastasis Prediction Using FastAI and 1-Cycle Policy. Sensors 2022, 22, 2988. [Google Scholar] [CrossRef]
  32. Singh, J.; Thakur, D.; Gera, T.; Shah, B.; Abuhmed, T.; Ali, F. Classification and analysis of android malware images using feature fusion technique. IEEE Access 2021, 9, 90102–90117. [Google Scholar] [CrossRef]
  33. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 14 October 2016; pp. 694–711. [Google Scholar]
  34. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  35. Olberg, S.; Zhang, H.; Kennedy, W.R.; Chun, J.; Rodriguez, V.; Zoberi, I.; Thomas, M.A.; Kim, J.S.; Mutic, S.; Green, O.L. Synthetic CT reconstruction using a deep spatial pyramid convolutional framework for MR-only breast radiotherapy. Med. Phys. 2019, 46, 4135–4147. [Google Scholar] [CrossRef] [PubMed]
  36. Yoo, G.S.; Luu, H.M.; Kim, H.; Park, W.; Pyo, H.; Han, Y.; Park, J.Y.; Park, S.-H. Feasibility of Synthetic Computed Tomography Images Generated from Magnetic Resonance Imaging Scans Using Various Deep Learning Methods in the Planning of Radiation Therapy for Prostate Cancer. Cancers 2021, 14, 40. [Google Scholar] [CrossRef]
  37. Jégou, S.; Drozdzal, M.; Vazquez, D.; Romero, A.; Bengio, Y. The one hundred layers tiramisu: Fully convolutional densenets for semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition workshops, Honolulu, HI, USA, 26 July 2017; pp. 11–19. [Google Scholar]
  38. Clevert, D.-A.; Unterthiner, T.; Hochreiter, S. Fast and accurate deep network learning by exponential linear units (elus). arXiv 2015, arXiv:1511.07289. [Google Scholar]
  39. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8026–8037. [Google Scholar]
  40. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  41. Gillies, R.J.; Kinahan, P.E.; Hricak, H. Radiomics: Images are more than pictures, they are data. Radiology 2016, 278, 563–577. [Google Scholar] [CrossRef] [Green Version]
  42. Alcantarilla, P.F.; Solutions, T. Fast explicit diffusion for accelerated features in nonlinear scale spaces. IEEE Trans. Patt. Anal. Mach. Intell. 2011, 34, 1281–1298. [Google Scholar]
Figure 1. Study design of our study. (A) Flowchart of the compare four different learning losses and obtain a cone-beam artifact-free CT image trained with each loss. (B) FC-DenseNet is used as the DL architecture for the eliminate cone-beam artifact in CBCT.
Figure 1. Study design of our study. (A) Flowchart of the compare four different learning losses and obtain a cone-beam artifact-free CT image trained with each loss. (B) FC-DenseNet is used as the DL architecture for the eliminate cone-beam artifact in CBCT.
Cancers 14 04534 g001
Figure 2. Feature mapping ratio (FMR). (A) Example of detect feature point. To detect the feature point, the difference image between adjacent Gaussian blurring images is used and the local extrema position is used as the feature point. (B) Example of binary descriptor. The binary number is calculated as 110(2) by points a, b, and c around the feature point p. 110(2) means information that b is brighter than a, c is brighter than b, and a is darker than c. (C) The binary descriptor determines the similarity using a Hamming distance.
Figure 2. Feature mapping ratio (FMR). (A) Example of detect feature point. To detect the feature point, the difference image between adjacent Gaussian blurring images is used and the local extrema position is used as the feature point. (B) Example of binary descriptor. The binary number is calculated as 110(2) by points a, b, and c around the feature point p. 110(2) means information that b is brighter than a, c is brighter than b, and a is darker than c. (C) The binary descriptor determines the similarity using a Hamming distance.
Cancers 14 04534 g002
Figure 3. Examples of DL output images. True CT, CBCT, synthetic CT(L1 loss), synthetic CT(LP loss), synthetic CT(LS loss), and synthetic CT(LPS loss) from axial plane from one of the patients in the 10 test set.
Figure 3. Examples of DL output images. True CT, CBCT, synthetic CT(L1 loss), synthetic CT(LP loss), synthetic CT(LS loss), and synthetic CT(LPS loss) from axial plane from one of the patients in the 10 test set.
Cancers 14 04534 g003
Figure 4. Example of dose distributions and dose differences. Ground-Truth, synthetic CT(L1 loss), synthetic CT(LP loss), synthetic CT(LS loss), and synthetic CT(LPS loss) from axial plane from Patient 3 in the 10 test set.
Figure 4. Example of dose distributions and dose differences. Ground-Truth, synthetic CT(L1 loss), synthetic CT(LP loss), synthetic CT(LS loss), and synthetic CT(LPS loss) from axial plane from Patient 3 in the 10 test set.
Cancers 14 04534 g004
Table 1. Patient Characteristics.
Table 1. Patient Characteristics.
VariablesTotal (65)Train (+Valid) Set (55)Test Set (10)
Age (years)
Median (range)56 (3–83)57.5 (3–78)55.5 (24–83)
Sex
Male33 (51)27 (49)6 (60)
Female32 (49)28 (51)4 (40)
Acquisition Diff. (days)
Median (range)13 (0–119)13 (0–119)15.5 (7–33)
Abbreviations: Acquisition Diff., acquisition date difference between CBCT images and CT images.
Table 2. SSIM, MAE (HU), and FMR performances for the DL outputs by generated from different learning losses for the 10 test set.
Table 2. SSIM, MAE (HU), and FMR performances for the DL outputs by generated from different learning losses for the 10 test set.
P1P2P3P4P5P6P7P8P9P10Average
SSIML10.9950.9800.9880.9860.9800.9860.9880.9910.9850.9910.987 (±0.045)
LP0.9930.9790.9880.9840.9780.9840.9870.9890.9840.9900.986 (±0.045)
LS0.9930.9800.9880.9860.9830.9850.9890.9900.9850.9880.987 (±0.035)
LPS0.9950.9830.9910.9870.9850.9880.9900.9920.9870.9920.989 (±0.034)
MAE (HU)L13.3248.9995.1017.3459.5506.5626.5554.6187.7904.5466.439 (±1.931)
LP3.7559.0494.9857.6449.8886.8256.6584.8318.3494.8286.681 (±1.947)
LS3.5138.6625.0157.3408.7546.6296.1994.8077.7505.1926.386 (±1.664)
LPS3.2117.9544.3856.5418.0166.0805.7244.2897.3094.4215.793 (±1.592)
PSNRL141.86233.78737.98236.36834.06735.81736.40339.43136.38338.65237.075 (±2.338)
LP41.40634.04838.56836.24933.89935.73436.79339.50636.35638.61637.118 (±2.271)
LS41.87434.65238.43536.65035.20235.90537.37439.34036.87937.59037.390 (±2.008)
LPS42.41834.99139.46631.17335.55336.15537.46940.02436.97338.66237.888 (±2.157)
FMR(0.8)L10.6050.4840.6010.5080.4440.4900.4590.6220.5260.5850.532 (±0.062)
LP0.6130.4700.6230.5250.4330.5020.4680.6050.5490.6010.539 (±0.066)
LS0.6420.5030.6290.5740.5100.5300.5130.6130.5800.5920.569 (±0.049)
LPS0.6440.5300.6460.5780.5340.5350.5360.6560.5620.6040.683 (±0.049)
Table 3. Gamma passing rate and dose difference analysis on CBCT and DL output by different learning losses compared with dose distribution on deformed CT for the 10 test set.
Table 3. Gamma passing rate and dose difference analysis on CBCT and DL output by different learning losses compared with dose distribution on deformed CT for the 10 test set.
P1P2P3P4P5P6P7P8P9P10Average
Gamma passing rate1 mm/1%L10.98850.91010.99090.97310.97270.89880.90940.97070.90270.93880.9456 (±0.036)
LP0.98690.89110.98530.97630.96850.89130.88510.96500.98400.92670.9370 (±0.041)
LS0.98360.91100.98560.98190.98210.92850.91900.97140.91230.93380.9510 (±0.031)
LPS0.99190.92790.99190.98940.99090.93900.94720.98010.92250.93740.9618 (±0.028)
2 mm/2%L10.99990.98690.99990.99540.99540.98510.98840.99820.98650.99330.9929 (±0.005)
LP0.99990.98490.99970.99800.99350.98500.98860.99800.99040.99530.9933 (±0.006)
LS0.99990.99190.99980.99930.99860.99110.99430.99780.99260.99330.9960 (±0.003)
LPS1.00000.99470.99990.99970.99980.99380.99680.99920.98850.99690.9969 (±0.004)
Dose differenceL10.01320.01460.00630.00810.00740.01970.01740.01000.01630.01250.0126 (±0.004)
LP0.01340.01550.00680.00800.00830.02070.01800.01040.01750.01290.0132 (±0.005)
LS0.01420.01450.00680.00770.00830.01740.01610.01020.01490.01340.0124 (±0.004)
LPS0.01250.01290.00600.00660.00810.01500.01410.00950.01510.01260.0112 (±0.003)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yoo, S.K.; Kim, H.; Choi, B.S.; Park, I.; Kim, J.S. Generation and Evaluation of Synthetic Computed Tomography (CT) from Cone-Beam CT (CBCT) by Incorporating Feature-Driven Loss into Intensity-Based Loss Functions in Deep Convolutional Neural Network. Cancers 2022, 14, 4534. https://doi.org/10.3390/cancers14184534

AMA Style

Yoo SK, Kim H, Choi BS, Park I, Kim JS. Generation and Evaluation of Synthetic Computed Tomography (CT) from Cone-Beam CT (CBCT) by Incorporating Feature-Driven Loss into Intensity-Based Loss Functions in Deep Convolutional Neural Network. Cancers. 2022; 14(18):4534. https://doi.org/10.3390/cancers14184534

Chicago/Turabian Style

Yoo, Sang Kyun, Hojin Kim, Byoung Su Choi, Inkyung Park, and Jin Sung Kim. 2022. "Generation and Evaluation of Synthetic Computed Tomography (CT) from Cone-Beam CT (CBCT) by Incorporating Feature-Driven Loss into Intensity-Based Loss Functions in Deep Convolutional Neural Network" Cancers 14, no. 18: 4534. https://doi.org/10.3390/cancers14184534

APA Style

Yoo, S. K., Kim, H., Choi, B. S., Park, I., & Kim, J. S. (2022). Generation and Evaluation of Synthetic Computed Tomography (CT) from Cone-Beam CT (CBCT) by Incorporating Feature-Driven Loss into Intensity-Based Loss Functions in Deep Convolutional Neural Network. Cancers, 14(18), 4534. https://doi.org/10.3390/cancers14184534

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop