| Literature DB >> 35789489 |
Branimir Rusanov1,2, Ghulam Mubashar Hassan1, Mark Reynolds1, Mahsheed Sabet1,2, Jake Kendrick1,2, Pejman Rowshanfarzad1,2, Martin Ebert1,2.
Abstract
The use of deep learning (DL) to improve cone-beam CT (CBCT) image quality has gained popularity as computational resources and algorithmic sophistication have advanced in tandem. CBCT imaging has the potential to facilitate online adaptive radiation therapy (ART) by utilizing up-to-date patient anatomy to modify treatment parameters before irradiation. Poor CBCT image quality has been an impediment to realizing ART due to the increased scatter conditions inherent to cone-beam acquisitions. Given the recent interest in DL applications in radiation oncology, and specifically DL for CBCT correction, we provide a systematic theoretical and literature review for future stakeholders. The review encompasses DL approaches for synthetic CT generation, as well as projection domain methods employed in the CBCT correction literature. We review trends pertaining to publications from January 2018 to April 2022 and condense their major findings-with emphasis on study design and DL techniques. Clinically relevant endpoints relating to image quality and dosimetric accuracy are summarized, highlighting gaps in the literature. Finally, we make recommendations for both clinicians and DL practitioners based on literature trends and the current DL state-of-the-art methods utilized in radiation oncology.Entities:
Keywords: AI; CT; adaptive radiotherapy; cone-beam CT; deep learning; image synthesis; synthetic CT
Mesh:
Year: 2022 PMID: 35789489 PMCID: PMC9543319 DOI: 10.1002/mp.15840
Source DB: PubMed Journal: Med Phys ISSN: 0094-2405 Impact factor: 4.506
Benefits and limitations of three common deep learning (DL) architectures: U‐Net, GAN (generative adversarial network), and cycle‐GAN
| Architecture | Strengths | Limitations |
|---|---|---|
| U‐Net |
Simplest implementation Stable convergence Fastest training |
Paired data only Anatomic misalignments reduce model accuracy and image realism |
| GAN |
Paired or unpaired training Improved image realism due to adversarial loss Model tunability |
Moderate implementation difficulty Unstable convergence Slower training Poor structure preservation for unpaired data |
| Cycle‐GAN |
Paired or unpaired training Model tunability Improved image realism due to adversarial loss Good structure preservation |
Complex implementation Unstable convergence Slowest training Highest hardware requirements |
FIGURE 1(a) The convolution output (feature map) results from element‐wise multiplication followed by summation between the filter and image. Note how image information is encoded into a reduced spatial dimension. (b) Depiction of the U‐Net architecture. Note how the input spatial dimensions are progressively reduced, whereas the feature dimension increases with network depth. (c) The GAN architecture comprising a generator and discriminator. Generators are typically U‐Net‐type architectures with encoder/decoder arms, whereas discriminators are encoder classifiers. (d) The Cycle‐GAN network comprising two generators and discriminators capable of unpaired image translation via computation of the cycle error. The orange arrows indicate the backward synthesis cycle path.
Summary of common image and dose based similarity metrics
| Metric | Formula | |
|---|---|---|
| Image similarity | MAE/ME ↓ |
|
| MSE/RMSE ↓ |
| |
| PSNR ↑ |
| |
| SSIM ↑ |
with | |
| DICE ↑ |
with | |
| Dosimetric similarity | DPR ↑ |
Fraction of voxels where DD ⩽ with |
| DVH ↑ | Cumulative histogram of dose–volume frequency distribution for a given volume | |
| GPR ↑ | Fraction of voxels where |
Note: Arrows indicate better result.
Abbreviations: DPR, dose difference pass rate; DVH, dose–volume histogram; GPR, gamma pass rate; MAE, mean absolute error; ME, mean error; MSE, mean squared error; PSNR, peak signal‐to‐noise ratio; RMSE, root mean square error; SSIM, structural similarity.
FIGURE 2Flowchart of study selection process
Summary of synthetic CT (sCT) generation methods
| Author and year | Anatomic site | Model | Loss function | Augmentation | Preprocessing | (train/val/test) | Training configuration | Image similarity (input CBCT) | Dose similarity |
|---|---|---|---|---|---|---|---|---|---|
| Kida et al. 2018 | Pelvis | U‐Net | MAE |
Voxels outside body set to −1000 HU Intra‐subject RR and DIR Masked CT to CBCT contour | 5‐CV (16/0/4) | Paired axial 2D | SSIM: 0.967 (0.928). PSNR: 50.9 (31.1). RMSE: 13 (232) | ||
| Xie et al. 2018 | Pelvis | Deep‐CNN | MSE |
Intra‐subject DIR 2D patches DIR patches | 15/0/5 | Paired axial patch 2D | PSNR: 8.823 (7.889). Anatomy ROI mean HU | ||
| Chen et al. 2019 | HN | U‐Net |
MAE SSIM |
Resample CT to CBCT Rescaled HU [0,1] Intra‐subject RR plan CT and CBCT Intra‐subject DIR replan CT and CBCT | 30/7/7 | Dual‐input paired axial 2D | MAE: 18.98 (44.38). PSNR: 33.26 (27.35). SSIM: 0.8911 (0.7109). RMSE: 60.16 (126.43) | ||
| Pelvis | 6/0/7 | Paired axial 2D | MAE: 42.40 (104.21). PSNR: 32.83 (27.59). SSIM: 0.9405 (0.8897). RMSE: 94.06 (163.71) | ||||||
| Harms et al. 2019 | Brain | Cycle‐GAN |
Adversarial loss Cycle loss (L1.5 norm) Synthetic loss (L1.5 norm) Gradient loss |
Resample CBCT to CT Intra‐subject RR Inter‐subject RR to common volume Air truncation 3D patches | LoO‐CV (23/0/1) | Paired patch 3D | MAE: 13.0 ± 2.2 (23.8 ± 5.1). PSNR: 37.5 ± 2.3 (32.3 ± 5.9) | ||
| Pelvis | LoO‐CV (19/0/1) | MAE: 16.1 ± 4.5 (56.3 ± 19.7). PSNR: 30.7 ± 3.7 (22.2 ± 3.4) | |||||||
| Kida et al. 2019 | Pelvis | Cycle‐GAN |
Adversarial loss Cycle loss Total variation loss Air loss Gradient loss Idempotent loss | Gaussian noise |
Voxels outside body set to −1000 HU Intra‐subject RR plan CT and CBCT Air truncation Intra‐subject DIR replan CT and CBCT HU clipped [−500, 200] HU rescaled [−1,1] | 16/0/4 | Unpaired axial 2D | Average ROI HU. Volume HU histograms. Self‐SSIM | |
| Kurz et al. 2019 | Pelvis | Cycle‐GAN |
Adversarial loss Cycle loss |
Random cropping Random left‐right flips |
Intra‐subject RR Voxels outside body set to −1000 HU CT/CBCT downsampled HU clipped [−1000,2071], rescaled 16 bit | 4‐CV (25/0/8) | Unpaired axial 2D | MAE: 87 (103) |
DD1: 89%. DD2: 100%. DVH < 1.5%.
|
| Lei et al. 2019 | Brain | Cycle‐GAN |
Adversarial loss Cycle loss Synthetic loss | Intra‐subject RR | LoO‐CV (11/0/1) | Paired patch 3D | MAE: 20.8 ± 3.4 (44.0 ± 12.6). PSNR: 32.8 ± 1.5 (26.1 ± 2.5) | ||
| Li et al. 2019 | Nasopharynx | U‐Net | MAE |
Random left‐right flips Random positional shifts |
Resample CT to CBCT Intra‐patient RR | 50/10/10 | Paired axial 2D | MAE: 6–27 (60–120). ME: −26–4 (−74–51) | DVH < 0.2% (0.8%). GPR1: 95.5% (90.8%) |
| Liang et al. 2019 | HN | Cycle‐GAN |
Adversarial loss Cycle loss Identity loss |
Resample CT to CBCT HU rescaled [−1,1] Intra‐patient DIR on test data | 81/9/20 | Unpaired axial 2D | MAE: 29.85 ± 4.94 (69.29 ± 11.01). RMSE: 84.46 ± 12.40 (182.8 ± 29.16). SSIM: 0.85 ± 0.03 (0.73 ± 0.04). PSNR: 30.65 ± 1.36 (25.28 ± 2.19) | GPR2: 98.40% ± 1.68% (91.37% ± 6.72%). GPR1: 96.26% ± 3.59% (88.22% ± 88.22%) | |
| Barateau et al. 2020 | HN | GAN |
Perceptual loss Adversarial loss | Random translations, rotations, shears | Intra‐patient RR and DIR | 30/0/14 | Paired axial 2D | MAE: 82.4 ± 10.6 (266.6 ± 25.8) | GPR2: 98.1% (91.0%). DVH (OAR) < 99 cGy. DVH (PTV) < 0.7% |
| Eckl et al. 2020 | HN | Cycle‐GAN |
Adversarial loss Cycle loss Synthetic loss |
Thorax and HN HU clipped [−1000,4000] Pelvis HU clipped [−1000,1000] HU rescaled [−1,1] Intra‐patient RR Images resampled 224 × 224 | 25/0/15 | Paired axial 2D | MAE: 77.2 ± 12.6 | GPR3: 98.6 ± 1.0%. GPR2: 95.0 ± 2.4%. DD2: 91.5 ± 4.3%. DVH < 1.7% | |
| Thorax | 53/0/15 | MAE: 94.2 ± 31.7 | GPR3: 97.8 ± 3.3%. GPR2: 93.8 ± 5.9%. DD2: 76.7 ± 17.3%. DVH < 1.7% | ||||||
| Pelvis | 205/0/15 | MAE: 41.8 ± 5.3 | GPR3: 99.9 ± 0.1%. GPR2: 98.5 ± 1.7%. DD2: 88.9 ± 9.3%. DVH < 1.1% |
Note: Dose similarity in italics suggests proton plans, otherwise photon plans. GPR3 = 3%/3 mm; GPR2 = 2%/2 mm.
Abbreviations: CBCT, cone‐beam CT; CNN, convolutional neural network; CV, cross‐validation; DIR, deformable image registration; DVH, dose–volume histogram; GAN, generative adversarial network; HN, head and neck; LoO‐CV, leave on out cross‐validation; MAE, mean absolute error; ME, mean error; MSE, mean squared error; P/C/GTV, planning/clinical/gross target volume; PSNR, peak signal‐to‐noise ratio; RMSE, root mean square error; ROI, regions of interest; RR, rigid registration; SSIM, structural similarity.
Image similarity metrics computed within body contour.
Summary of synthetic CT (sCT) generation methods
| Author and year | Anatomic site | Model | Loss function | Augmentation | Preprocessing | (train/val/test) | Training configuration | Image similarity (input CBCT) | Dose similarity |
|---|---|---|---|---|---|---|---|---|---|
| Liu et al. 2020 | Abdomen | Cycle‐GAN |
Adversarial loss Cycle loss Synthetic loss |
Intra‐patient RR and DIR CBCT resampled to CT | LoO‐CV (29/0/1) | Paired patch 3D | MAE: 56.89 ± 13.84 (81.06 ± 15.86) | DVH < 0.8% | |
| Maspero et al. 2020 |
HN Lung Breast | Cycle‐GAN |
Adversarial loss Cycle loss |
Random left‐right flipping Random 30 × 30 cropping |
Voxels outside largest circular mask on CBCT and CT set to −1000 HU Intra‐patient RR Images resampled 286 × 286 HU clipped [−1024,3071] HU rescaled [0,1] CT anatomy outside CBCT FOV stitched on |
15/8/10 15/8/10 15/8/10 | Unpaired axial 2D |
MAE: 51 ± 12 (195 ± 20) MAE: 86 ± 9 (219 ± 44) MAE: 67 ± 18 (152 ± 40) |
GPR3: 99.3 ± 0.4%. GPR2: 97.8 ± 1% GPR3: 98.2 ± 1%. GPR2: 94.9 ± 3% GPR3: 97 ± 4%. GPR2: 92. ± 8% |
| Park et al. 2020 | Lung | Cycle‐GAN |
Adversarial loss Cycle loss | CT and CBCT resampled to 384 × 384 | 8/0/2 | Unpaired sagittal and coronal 2D | PSNR: 30.60 (26.13). SSIM: 0.8977 (0.8173) | ||
| Thummerer et al. 2020 | HN | U‐Net | MAE |
Voxels outside body set to −1000 HU Intra‐patient RR and DIR CT and CBCT masks reduced to common voxels Slices containing shoulders removed | 3‐CV (16/2/9) | Paired axial, sagittal and coronal 2D | MAE: 40.2 ± 3.9 |
| |
| Thummerer et al. 2020 | HN | U‐Net | MAE |
Small translations Random left‐right mirroring |
Voxels outside body set to −1000 HU Intra‐patient RR and DIR CT and CBCT masks reduced to common voxels Slices containing shoulders removed | 3‐CV (11/11/11) | Paired axial, sagittal, coronal 2D | MAE: 36.3 ± 6.2 |
|
| Xie et al. 2020 | Pelvis | Deep‐CNN | Contextual loss | Random rotations | Intra‐patient DIR | 499/64/64 (slices) | Paired axial 2D | MAE: 46.01 ± 5.28 (51.01 ± 5.38). PSNR: 23.07 (22.66). SSIM: 0.8873 (0.8749) | |
| Yuan et al. 2020 | HN | U‐Net | MAE |
Intra‐patient RR Images cropped 256 × 256 Central 52 slices used | 5‐CV (40/5/10) | Paired axial 2D | MAE: 49.24 (167.46). SSIM: 0.85 (0.42) | ||
| Zhang et al. 2020 | Pelvis | GAN |
Feature matching MAE |
Random left–right flipping Random small angle rotation Background noise |
Intra‐patient DIR HU rescaled to mean of 0, STD of 1 (standardized) | 150/0/15 | Paired multi‐slice axial 2.5D | MAE: 23.6 ± 4.5 (43.8 ± 6.9). PSNR: 20.09 ± 3.4 (14.53 ± 6.7) | DVH < 1% |
| Dahiya et al. 2021 | Thorax | GAN |
Adversarial loss MAE | Geometric augmentation (scale, sheer, rotation) |
CBCT artifact injection into CT HU clipped [−1000, 3095] HU rescaled [−1, 1] Intra‐patient DIR Image resampled to 128 × 128 × 128 | 140/0/15 | Paired 3D | MAE: 29.31 ± 12.64 (162.77 ± 53.91). RMSE: 78.62 ± 78.62 (328.18 ± 84.65). SSIM: 0.92 ± 0.01 (0.73 ± 0.07). PSNR: 34.69 ± 2.41 (22.24 ± 2.40) | |
| Dai et al. 2021 | Breast | Cycle‐GAN |
Adversarial loss Cycle loss | 52/0/23 | MAE: 71.58 ± 8.78 (86.42 ± 10.12) | GPR3: 91.46 ± 4.63%. GPR2: 85.09 ± 6.28%. DVH (CTV) < 3.58% | |||
| Dong et al. 2021 | Pelvis | Cycle‐GAN |
Adversarial loss Cycle loss Identity loss |
Images resampled 1 × 1 × 1‐mm grid HU rescaled [−1, 1] Voxels outside body set to −1000 HU | 46/0/9 | Unpaired axial 2D | MAE: 14.6 ± 2.39 (49.96 ± 7.21). RMSE: 56.05 ± 13.05 (105.9 ± 11.52). PSNR: 32.5 ± 1.87 (26.82 ± 0.638). SSIM: 0.825 ± 1.92 (0.728 ± 0.36) | ||
| Gao et al. 2021 | Thorax | Cycle‐GAN |
Adversarial loss Cycle loss Identity loss |
Intra‐patient RR CT FOV cropped to CBCT HU clipped [−1000, 1500] HU rescaled [−1, 1] Images resampled 256 × 256 | 136/0/34 | Unpaired axial 2D | MAE: 43.5 ± 6.69 (92.8 ± 16.7). SSIM: 0.937 ± 0.039 (0.783 ± 0.063). PSNR: 29.5 ± 2.36 (21.6 ± 2.81) | GPR3: 99.7 ± 0.39% (92.8 ± 3.86%). GPR2: 98.6 ± 1.78% (84.4 ± 5.81%). GPR1: 91.4 ± 3.26% (50.1 ± 9.04%) | |
| Liu et al. 2021 | Thorax | Modified ADN |
Adversarial loss Attribute consistency loss Reconstruction loss Self‐reconstruction loss SSIM loss | Random horizontal flip |
Resample CT/CBCT to 1 × 1 × 1‐mm grid Resample to 384 × 384 Extract 256 × 256 image patches HU clipped [−1000, 2000] HU rescaled [−1, 1] Intra‐patient RR | 32/8/12 | Unpaired axial 2D patch | MAE: 32.70 ± 7.26 (70.56 ± 11.81). RMSE: 60.53 ± 60.53 (112.13 ± 17.91). SSIM: 0.86 ± 0.04 (0.64 ± 0.04). PSNR: 34.12 ± 1.32 (28.67 ± 1.41) |
Note: Dose similarity in italics suggests proton plans, otherwise photon plans. GPR3 = 3%/3 mm; GPR2 = 2%/2 mm.
Abbreviations: ADN, artifact disentanglement network; CBCT, cone‐beam CT; CNN, convolutional neural network; CV, cross‐validation; DIR, deformable image registration; DVH, dose–volume histogram; GAN, generative adversarial network; HN, head and neck; LoO‐CV, leave on out cross validation; MAE, mean absolute error; ME, mean error; P/C/GTV, planning/clinical/gross target volume; PSNR, peak signal‐to‐noise ratio; RMSE, root mean square error; RR, rigid registration; SSIM, structural similarity.
Image similarity metrics computed within body contour.
Summary of synthetic CT (sCT) generation methods
| Author and year | Anatomic site | Model | Loss function | Augmentation | Preprocessing | (train/val/test) | Training configuration | Image similarity (input CBCT) | Dose similarity |
|---|---|---|---|---|---|---|---|---|---|
| Qiu et al. 2021 | Thorax | Cycle‐GAN |
Adversarial loss Cycle loss Histogram matching loss Synthetic loss Gradient loss Perceptual loss |
Rotations Flips Rescaling Rigid deformations | Intra‐patient RR and DIR | 5‐CV (16/0/4) | Paired axial 2D | MAE: 66.2 ± 8.2 (110.0 ± 24.9) | |
| Rossi et al. 2021 | Pelvis | U‐Net | MAE |
Random 90° rotations Horizontal flip |
Voxels outside body set to −1000 HU HU clipped [−1024, 3200] HU rescaled [0, 1] Intra‐patient RR Image resampled to 256 × 256 | 4‐CV (42/0/14) | Paired axial 2D | MAE: 35.14 ± 13.19 (93.30 ± 59.60). PSNR: 30.89 ± 2.66 (26.70 ± 3.36). SSIM: 0.912 ± 0.033 (0.887 ± 0.048) | |
| Sun et al. 2021 | Pelvis | Cycle‐GAN |
Adversarial loss Cycle loss Gradient loss |
Intra‐patient RR Image resampled to 384 × 192 × 192 | 5‐CV (80/20/20) | Paired patch 3D | MAE: 51.62 ± 4.49. SSIM: 0.86 ± 0.03. PSNR: 30.70 ± 0.78 (27.15 ± 0.57) | GPR2: 97% | |
| Thummerer et al. 2021 | Thorax | U‐Net | MAE |
Intra‐patient RR and DIR Voxels outside body set to −1000 HU CT FOV cropped to CBCT | 3‐CV (22/0/11) | Paired axial, sagittal, coronal 2D | MAE: 30.7 ± 4.4 |
| |
| Tien et al. 2021 | Breast | Cycle‐GAN |
Adversarial loss Cycle loss Synthetic loss Identity loss Gradient loss |
Random cropping to 128 × 128 Random horizontal/vertical flips Random rotation |
Clipped images to 264 × 336 HU clipped [−950,500] HU rescaled [0,1] | 12/0/3 | Paired axial 2D | Average ROI HU. ROI MAE. ROI PSNR. ROI SSIM | |
| Uh et al. 2021 |
Abdomen Pelvis | Cycle‐GAN |
Adversarial loss Cycle loss |
Intra‐patient RR Voxels outside body set to −1000 HU Body normalization: lateral extent of anatomy scaled to 475 mm CBCT and CT resampled |
21/0/7 29/0/7 | Paired axial 2D |
MAE: 44 (141) MAE: 51 (105) |
| |
| Xue et al. 2021 | Nasopharynx | Cycle‐GAN |
Adversarial loss Cycle loss Identity loss |
Intra‐patient RR Voxels outside body set to −1000 HU HU clipped [−1000, 2000] HU rescaled [−1,1] | 135/0/34 | Paired axial 2D | MAE: 23.8 ± 8.6 (42.2 ± 17.4). RMSE: 79.7± 20.1 (134.3 ± 31.0). PSNR: 37.8 ± 2.1 (27.2 ± 1.9). SSIM: 0.96 ± 0.01 (0.91 ± 0.03) | GPR3 > 98.52% ± 3.09%. GPR2 > 96.82% ± 1.71% | |
| Zhao et al. 2021 | Pelvis | Cycle‐GAN |
Adversarial loss Cycle loss Idempotent loss Gradient loss | Added noise |
Voxels outside body set to −1000 HU Intra‐patient RR CBCT and CT resampled HU clipped [−1000,3095] HU rescaled [−1,1] | 100/0/10 | Unpaired axial 2D | MAE: 52.99 ± 12.09 (135.84 ± 41.59). SSIM: 0.81 ± 0.03 (0.44 ± 0.07). PSNR: 26.99 ± 1.48 (21.76 ± 1.95). | DVH < 50 cGy (< 350 cGy) |
| Wu et al. 2022 | Pelvis | Deep‐CNN |
Gradient loss MAE |
CBCT resampled to CT Intra‐patient DIR Voxels outside body set to −1000 HU Images cropped to 440 × 440 HU rescaled [0, 1] | 5‐CV (90/30/23) | Paired 2D | MAE: 52.18 ± 3.68 (352.56) | ||
| Lemus et al. 2022 | Abdomen | Cycle‐GAN |
Cycle loss Adversarial loss Gradient loss Idempotent loss Total Variation loss | Random 256 × 256 image sampling |
Intra patient RR (training) Intra patient DIR (testing) Images cropped to 480 × 384 | 10‐CV (11/0/6) | Paired 2D | MAE: 54.44 ± 16.39 (72.95 ± 6.63). RMSE: 108.765 ± 40.54 (137.29 ± 21.19) | DVH (PTV): 1.5% (3.6%). GPR3/2: 98.35% (96%) |
Note: Dose similarity in italics suggests proton plans, otherwise photon plans. GPR3 = 3%/3 mm; GPR2 = 2%/2 mm.
Abbreviations: CBCT, cone‐beam CT; CNN, convolutional neural network; CV, cross validation; DIR, deformable image registration; DVH, dose–volume histogram; GAN, generative adversarial network; HN, head and neck; LoO‐CV, leave on out cross validation; MAE, mean absolute error; ME, mean error; P/C/GTV, planning/clinical/gross target volume; PSNR, peak signal‐to‐noise ratio; RMSE, root mean square error; ROI, regions of interest; RR, rigid registration; SSIM, structural similarity.
Image similarity metrics computed within body contour.
Miscellaneous approaches for cone‐beam CT (CBCT) correction
| Author and year | Anatomic site | Model | Loss function | Augmentation | Preprocessing | (train/val/test) | Training configuration | Image similarity (input CBCT) | Dose similarity |
|---|---|---|---|---|---|---|---|---|---|
| Hansen et al. 2018 | Pelvis | U‐Net | MSE | Linear combination of two random inputs (Mixup) | A priori scatter correction for target projections | 15/8/7 | Paired projection 2D | MAE: 46 (144). ME: −3 (138) | GPR2: 100%. GPR1: 90%. |
| Jiang et al. 2019 | Pelvis | U‐Net | MSE | MC scatter correction for target CBCTs | 15/3/2 | Paired axial 2D | RMSE: 18.8 (188.4). SSIM: 0.9993 (0.9753) | ||
| Landry et al. 2019 | Pelvis | U‐Net 1 | MSE | Mixup | A priori scatter correction for target projections | 27/7/8 | Paired projection 2D | MAE: 51 (104). ME: 1 (30) |
DPR2 > 99.5%. DPR1 > 98.4%. GPR2 > 99.5%
|
| U‐Net 2 |
Random left‐right flips Random position shifts Random HU shifts |
Intra‐patient DIR CT resampled to CBCT Voxels outside body set to −1000 HU CT cropped to CBCT cylindrical FOV CT and CBCT cropped to remove conical ends of CBCT | Paired axial 2D | MAE: 88 (104). ME: 2 (30) |
DPR2 > 99.5%. DPR1 > 98.4%. GPR2 > 99.5%
| ||||
| U‐Net 3 | Voxels outside body set to −1000 HU | MAE: 58 (104). ME: 3 (30) | DPR2 > 99.5%. DPR1 > 98.4%. GPR2 > 99.5%. | ||||||
| Nomura et al. 2019 | HN | U‐Net | MAE |
Random left‐right flips Random 90° rotations |
MC simulation of training, validation and testing data Voxels outside body set to −1000 HU Anatomy segmented into air, adipose, soft tissue, muscle, rib bone |
Training: 5 phantoms. Validation: HN phantom. Testing: 1 HN, 1 Thorax patient | Paired projections 2D | MAE: 17.9 ± 5.7 (21.8 ± 5.9). SSIM: 0.9997 ± 0.0003 (0.9995 ± 0.0003). PSNR: 37.2 ± 2.6 (35.6 ± 2.3) | |
| Thorax | MAE: 29.0 ± 2.5 (32.5 ± 3.2). SSIM: 0.9993 ± 0.0002 (0.9990 ± 0.0003). PSNR: 31.7 ± 0.8 (30.6 ± 0.9) | ||||||||
| Lalonde et al. 2020 | HN | U‐Net | MAPE | Vertical and horizontal flips |
Projections downsampled to 256 × 256 Projection intensities normalized against flood field | 29/9/10 | Paired projections 2D | MAE: 13.41 (69.64). ME: −0.801 (−28.61) |
|
| Rusanov et al. 2021 | HN | U‐Net | MAE | Random vertical/horizontal flips | Bowtie filter removal via projection normalization using flood field scan | 4/0/2 | Paired projection 2D | MAE: 74 (318) |
Note: GPR3 = 3%/3 mm; GPR2 = 2%/2 mm; GPR1 = 1%/1‐mm criteria. DPR2 = 2% DD threshold; DPR1 = 1% DD threshold.
Abbreviations: DIR, deformable image registration; HN, head and neck; MAE, mean absolute error; MC, Monte Carlo; ME, mean error; MSE, mean squared error; PSNR, peak signal‐to‐noise ratio; RMSE, root mean square error; SSIM, structural similarity.
Image similarity metrics computed within body contour.
FIGURE 3Distribution of total and per architecture investigations per year
FIGURE 4Pie chart of distribution of anatomic sites investigated
FIGURE 5Percent mean absolute error (MAE) improvement per network for studies utilizing common data
FIGURE 6Percent mean absolute error (MAE) improvement for cycle‐generative adversarial network (GAN) models trained with paired or unpaired datasets, controlling for pelvic, head and neck (HN), and all anatomical regions, as well as training set sizes within four patients
Mean cohort size and model performance statistics for all publications
| Training size (no patients) | Testing size (no. patients) | CBCT MAE (HU) | sCT MAE (HU) | % MAE improvement | |
|---|---|---|---|---|---|
| All studies | 47.74 ± 47.11 | 11.02 ± 7.90 | – | 46.83 ± 22.23 | – |
| All studies* | 51.27 ± 44.27 | 11.77 ± 8.77 | 114.66 ± 75.84 | 45.13 ± 22.01 | 54.60 ± 17.90 |
| Pelvis* | 62.63 ± 43.20 | 10.88 ± 6.11 | 117.47 ± 93.73 | 41.58 ± 22.73 | 57.97 ± 19.68 |
| HN* | 45.63 ± 39.38 | 12.13 ± 10.17 | 106.59 ± 84.63 | 36.13 ± 21.76 | 58.67 ± 10.75 |
| Thorax* | 59.00 ± 56.18 | 14.17 ± 9.46 | 134.52 ± 49.45 | 54.12 ± 20.47 | 57.54 ± 12.65 |
| Abdomen* | 20.33 ± 7.36 | 4.67 ± 2.62 | 98.34 ± 30.35 | 51.78 ± 5.59 | 41.33 ± 19.51 |
Note: * indicates studies which reported CBCT and sCT MAE values.
Abbreviations: CBCT, cone‐beam CT; HN, head and neck; MAE, mean absolute error; sCT, synthetic CT.
FIGURE 7Scatter plot demonstrating the relationship between training cohort size and percent mean absolute error (MAE) improvement
FIGURE 8Absolute synthetic CT (sCT) mean absolute error (MAE) ordered from highest to lowest compared against training set size. Publication format describes: (model architecture/supervision type + additional loss functions and/or 3D training) | anatomical region. +, additional loss functions/3D input; A, abdomen; ADN, artifact disentanglement network; C, cycle‐GAN; CNN, convolutional neural network; D, deep CNN; G, GAN; GAN, generative adversarial network; HN, head and neck; P, paired training; P, pelvis; T, thorax; U, U‐Net; Un, unpaired training.
FIGURE 9Percentage mean absolute error (MAE) improvement ordered from lowest to highest compared against training set size. Publication format describes: (model architecture/supervision type + additional loss functions and/or 3D training) | anatomical region. +, additional loss functions/3D input; A, abdomen; ADN, artifact disentanglement network; C, cycle‐GAN; CNN, convolutional neural network; D, deep CNN; G, GAN; GAN, generative adversarial network; HN, head and neck; P, paired training; P, pelvis; T, thorax; U, U‐Net; Un, unpaired training
FIGURE 10Percentage structural similarity (SSIM) improvement ordered from lowest to highest compared against training set size. Publication format describes: (model architecture/supervision type + additional loss functions and/or 3D training) | anatomical region. *, low dose CBCT; +, additional loss functions/3D input; A, abdomen; ADN, artifact disentanglement network; C, cycle‐GAN; CNN, convolutional neural network; D, deep CNN; G, GAN; GAN, generative adversarial network; HN, head and neck; P, paired training; P, pelvis; T, thorax; U, U‐Net; Un, unpaired training
Studies reporting mean gamma pass rates for different anatomical regions and radiation modalities
| Head and neck | Pelvis | Abdomen | Breast | Lung | |||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
| Photon | [2, 0] | [5, 0] | [2, 0] | [1, 0] | [2, 0] | – | [1, 0] | – | – | [1, 1] | [0, 1] | – | [3, 0] | [1, 2] | – |
| Proton | [2, 0] | [2, 0] | – | [1, 0] | [2, 0] | – | – | [1, 0] | – | – | – | – | [1, 0] | [0, 1] | – |
Note: Mean gamma rates above 95% are considered clinically acceptable. Reporting format: [N > 95%, N < 95%] with N = number of evaluations. γ 3 = 3%/3 mm; γ 2 = 2%/2 mm; γ 1 = 1%/1 mm; γ 3/2 = 3%/2 mm. Light green = 1 validation; medium green = 2 validations; dark green = 3+ validations.