Next Article in Journal
Evaluation of Color Difference Models for Wide Color Gamut and High Dynamic Range
Next Article in Special Issue
The Neural Frontier of Future Medical Imaging: A Review of Deep Learning for Brain Tumor Detection
Previous Article in Journal
PAS or Not PAS? The Sonographic Assessment of Placenta Accreta Spectrum Disorders and the Clinical Validation of a New Diagnostic and Prognostic Scoring System
Previous Article in Special Issue
MOTH: Memory-Efficient On-the-Fly Tiling of Histological Image Annotations Using QuPath
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Toward Closing the Loop in Image-to-Image Conversion in Radiotherapy: A Quality Control Tool to Predict Synthetic Computed Tomography Hounsfield Unit Accuracy

by
Paolo Zaffino
1,2,*,
Ciro Benito Raggio
3,
Adrian Thummerer
2,4,
Gabriel Guterres Marmitt
2,
Johannes Albertus Langendijk
2,
Anna Procopio
1,
Carlo Cosentino
1,
Joao Seco
5,6,
Antje Christin Knopf
7,8,
Stefan Both
2,† and
Maria Francesca Spadea
3,†
1
Department of Experimental and Clinical Medicine, Magna Graecia University, viale Europa, 88100 Catanzaro, Italy
2
Department of Radiation Oncology, University Medical Center Groningen, University of Groningen, 9712 CP Groningen, The Netherlands
3
Institute of Biomedical Engineering, Karlsruhe Institute of Technology (KIT), 76131 Karlsruhe, Germany
4
Department of Radiation Oncology, University Hospital, LMU Munich, 81377 Munich, Germany
5
Department of Biomedical Physics in Radiation Oncology, Deutsches Krebsfoschungszentrum (DKFZ), 69120 Heidelberg, Germany
6
Department of Physics and Astronomy, Heidelberg University, 69120 Heidelberg, Germany
7
Institute for Medical Engineering and Medical Informatics, School of Life Science FHNW, 4132 Muttenz, Switzerland
8
Department of Radiotherapy and Radiation Oncology, Faculty of Medicine, University Hospital Carl Gustav Carus, Technische Universität Dresden, 01307 Dresden, Germany
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2024, 10(12), 316; https://doi.org/10.3390/jimaging10120316
Submission received: 13 November 2024 / Revised: 3 December 2024 / Accepted: 7 December 2024 / Published: 10 December 2024

Abstract

:
In recent years, synthetic Computed Tomography (CT) images generated from Magnetic Resonance (MR) or Cone Beam Computed Tomography (CBCT) acquisitions have been shown to be comparable to real CT images in terms of dose computation for radiotherapy simulation. However, until now, there has been no independent strategy to assess the quality of each synthetic image in the absence of ground truth. In this work, we propose a Deep Learning (DL)-based framework to predict the accuracy of synthetic CT in terms of Mean Absolute Error (MAE) without the need for a ground truth (GT). The proposed algorithm generates a volumetric map as an output, informing clinicians of the predicted MAE slice-by-slice. A cascading multi-model architecture was used to deal with the complexity of the MAE prediction task. The workflow was trained and tested on two cohorts of head and neck cancer patients with different imaging modalities: 27 MR scans and 33 CBCT. The algorithm evaluation revealed an accurate HU prediction (a median absolute prediction deviation equal to 4 HU for CBCT-based synthetic CTs and 6 HU for MR-based synthetic CTs), with discrepancies that do not affect the clinical decisions made on the basis of the proposed estimation. The workflow exhibited no systematic error in MAE prediction. This work represents a proof of concept about the feasibility of synthetic CT evaluation in daily clinical practice, and it paves the way for future patient-specific quality assessment strategies.

1. Introduction

Recently, several methods able to convert Magnetic Resonance (MR) and Cone Beam Computed Tomography (CBCT) images into synthetic CT (sCT) have been proposed [1,2,3,4]. This image-to-image (I2I) translation strategy can help when it is not possible or not convenient to acquire a real CT. In particular, radiotherapy is one of the main fields where sCT can play a fundamental role in improving treatment planning, allowing MR and CBCT to be integrated into the imaging pipeline without the need for additional CT scans [5,6,7,8,9,10,11]. Indeed, MR offers better soft tissue contrast to contour the target tissue and the organs at risk during treatment simulation. However, MR intensities do not reveal electron density properties, and additional CT acquisitions are necessary for dose calculation purposes. CBCT, on the other hand, is widely used for assessing inter-fraction geometric deviations in adaptive radiotherapy. But, again, CBCT cannot be used to calculate the dose due to the poor image quality; this forces the clinician to re-plan the treatment on new CT scans when it is necessary. The conversion of MR and CBCT in sCT images can reduce radiation toxicity among patients and decrease the time required for the treatment [2,12]. To date, the best-performing strategies to remap image intensities rely on supervised artificial intelligence algorithms [3,4], in particular those based on deep learning (DL). However, the accuracy of this kind of algorithm strongly depends on the dataset used for the training, and they can underperform in case of strong image artifacts and atypical anatomy. Moreover, the translation quality is assessed by comparing the sCT with the real CT, which represents the ground truth (GT). One of the most used metrics to assess conversion accuracy is the pixel-to-pixel Mean Absolute Error (MAE) [3], defined as Equation (1).
M A E = i = 1 n G T i s C T i n
Mean Absolute Error formula . i ranges from 1 to n , number of pixels / voxels both for GT and sCT .
In the case of clinical use, the absence of GT makes it impossible to evaluate the conversion accuracy, which has a direct impact on dose calculation. As a consequence, clinicians have to trust the pre-trained DL algorithms, being unable, except for eye-catching macroscopic mistakes, to detect conversion errors that could dramatically affect the treatment. For the clinical use of sCT for treatment adaptation [13], the implementation of online quality control procedures is mandatory. Some recent works [14,15,16] propose a DL model for synthetic CT generation with uncertainty predictions. The conversion pipeline provides an sCT alongside a predicted uncertainty map. A shortcoming of this approach is the information generation within one algorithm. In case of conversion failure, the uncertainty estimation could also be unreliable. Moreover, it is worth underlining the fundamental difference between “uncertainty” and “error”: a system (as well as a human) can make totally wrong predictions while being, at the same time, absolutely certain about them. For this reason, it is important to predict, preferably by using an independent tool, the conversion error of a process rather than its confidence. In this paper, we implemented a strategy to estimate the quality of sCT in terms of MAE without the need for the corresponding GT and independent of the translation model. The method is based on deep learning algorithms that were trained and tested on two patient cohorts including MR- and CBCT-based sCT images.

2. Materials and Methods

2.1. Dataset

Two datasets of head and neck cancer patients, previously described in Thummerer et al. [17,18], were used in this study. All patients were treated with intensity-modulated proton therapy at the University Medical Center Groningen, The Netherlands. The first set included 33 patients scanned both with CBCT and CT. The second set included 27 patients with MR and corresponding CT images. All CBCT/CT and MR/CT pairs were carefully aligned using deformable image registration to match the anatomy. For each patient of both datasets, an sCT was generated by using a basic version of the algorithm described in Spadea et al. [19]. The original Spadea et al. method, relying on an encoder/decoder architecture to execute 2D-based image translation, takes advantage of a multi-plane approach, in which volumes obtained along axial, sagittal, and coronal directions are voted to generate the final image. In this work, only the synthetic CTs reconstructed by stacking the translated 2D images along the craniocaudal axis were used. From now on, the sCTs derived from CBCT will be referred to as s C T C B C T and the ones obtained from MR as s C T M R .

2.2. General Pipeline

The proposed quality control approach is based on a 2D DL model architecture pipeline. The main idea is to provide an axial sCT slice as the input to the DL scheme and to obtain the predicted MAE for that image as the output (see Figure 1). To achieve this, the DL model is trained on retrospective data where ground truth CT and the corresponding sCT are available. Having GT images, the actual MAE can be computed slice-by-slice according to equation 1. During the training, the pipeline receives the sCT slice (input) and the actual MAE (output to predict). Afterwards, during deployment, the trained models only receive the sCT slices and predict an MAE scalar for the synthetic 2D image.
More specifically, as depicted in Figure 2, the workflow consists of a cascade of 2 series of DL models, used in a sequential way to handle the complexity of the MAE prediction task. The rationale is to first execute a raw prediction of the MAE interval (classification), followed by a more precise estimation of the MAE value (regression). This approach was inspired by a similar and successful strategy to predict segmentation accuracy in multi-atlas-based segmentation [20]. This strategy allows the initial problem to be split into two subproblems, both of which are easier to solve than the first one.

2.3. Output Visualization

The output of a single prediction is a scalar value representing the estimated MAE for the entire given axial slice. In envisioning a clinical use, it is fundamental to visualize the predicted conversion error slice-by-slice and overlaid on the sCT. To achieve this, a volume (named “Predicted MAE volume”, p M A E v o l u m e ), having the same shape as the corresponding sCT, is generated. During the sCT evaluation step, the predicted MAE of the i-th axial slice is assigned to the i-th axial slice of p M A E v o l u m e , which will be visualized by taking advantage of a dedicated color map. Figure 3 shows an example of an sCT overlaid to its p M A E v o l u m e .

2.4. Pipeline Details

As already reported, the first step for MAE estimation is a raw classification of the conversion error. More specifically, 4 classes are used in the proposed pipeline (low, low-medium, medium-high, and high MAE). In order to define the MAE range for each interval, the actual MAE distribution of the entire dataset is computed. Then, the first class of the prediction is associated with an MAE ranging from 0 to the 1st quartile of the total distribution, the second class from the 1st to the 2nd quartile, the third class from the 2nd to the 3rd quartile, and finally the fourth class from the 3rd percentile to the highest MAE values (in order to deal with an open interval, this last bin was saturated to an upper bound HU value). This allows for the balancing of the number of examples per class. The same values are also used as an initialization to define the MAE boundaries for the regression step (the second one of the pipeline), where a separate model is trained for each of the four MAE classes. More specifically, each of the four regression models is trained by showing only the slices having an MAE value included in the same range of the connected classification class. However, in order to take into account possible mistakes in the classification step, the boundaries of the regression models are relaxed by also including cases with GT MAE slightly lower/higher than the quartiles (±5 HU). Table 1 reports the computed bin ranges, both for the classification and regression steps. In total, 5 models have to be trained (1 for classification and 4 for regression), but, in the inference mode, each sCT slice will be processed just by 2 of them (classification followed by the class-specific regression). For the sake of clarity, the entire prediction pipeline is shown in Figure 4.

2.5. DL Architecture and Training

The DL architecture to be used in this work was chosen after testing several models, both from VGG [21] and ResNet [22] families. Our tests revealed that VGG-16 provides the best prediction accuracies, so it was used as the baseline architecture for all the models. VGG (Visual Geometry Group) is a family of high-performance convolutional neural networks developed for classification and regression tasks. Even though it was created for general-purpose computer vision applications, it has also been widely used in the biomedical environment. The backbone of VGG-16 includes 13 convolutional layers (that serve as feature extraction engines), followed by 3 fully connected layers (that are used to obtain the output label starting from the computed feature map). The loss functions to optimize for classification and regression are, respectively, cross entropy and mean squared error. The best epoch for the classification task is not chosen on the basis of the cross entropy value but by considering the accuracy weighted according to the misclassification magnitude. In practical terms, this means that a classification error of 1 class is preferred over 2 and 3 classes of misclassification. This choice is made to let the regression step, trained on slightly larger MAE boundaries than the connected classification bin, recover the potential classification error. The entire workflow is written in Python by taking advantage of PyTorch library [23]. The learning rate is set to 5 × 10 5 , and L 1 and L 2 regularizations are included (both weights equal to 4 × 10 4 ). The image augmentation scheme includes mirroring and translations, as defined by Spadea et al [19]. The dataset is split into 75% training, 10% validation, and 15% testing. K-fold cross-validation is used to evaluate the model generalization performance on the entire dataset.

2.6. Experiments

To assess the learning generalization capability of the approach toward different image modalities, 3 pipelines are implemented:
  • Models trained only with s C T C B C T as input (named P i p e l i n e C B C T ),
  • Models trained only with s C T M R as input ( P i p e l i n e M R ),
  • Models trained with both s C T C B C T and s C T M R as input ( P i p e l i n e M I X E D ).
Once all the models of each pipeline are trained, multiple testing schemes are run to evaluate the prediction performance. In particular, the following experiments are executed:
  • P i p e l i n e C B C T is used to predict the MAE only for s C T C B C T data
  • P i p e l i n e M R is used to predict the MAE only for s C T M R data
  • P i p e l i n e M I X E D is used to predict the MAE only for s C T C B C T data
  • P i p e l i n e M I X E D is used to predict the MAE only for s C T M R data
Table 1 shows the MAE binning used to train both classification and regression models in each pipeline.

2.7. Prediction Pipeline Evaluation

The accuracy of the prediction workflow is quantified in terms of signed and unsigned deviation from the GT MAE. More specifically, both the Prediction Deviation (PD) and the Absolute Prediction Deviation (APD) are computed for each slice as defined in Equations (2) and (3):
P D = M A E G T M A E p r e d i c t e d
Signed predicted deviation between GT and inferred MAE .
A P D = M A E G T M A E p r e d i c t e d
Absolute predicted deviation between GT and inferred MAE .
Both M A E G T and M A E p r e d i c t e d are saturated to the upper bound values reported in Table 1.

3. Results

Figure 5 and Figure 6 show, for each DL pipeline, the PD and APD distributions computed over all the 2D slices of all the patients.
Tables containing the median, the 5th, the 25th, the 75th, and the 95th percentiles of both PD and APD are reported in Appendix A. To assess whether a statistical difference exists between the predictions obtained by using the mixed-trained workflow ( P i p e l i n e M I X E D ) versus the specific modality pipeline ( P i p e l i n e C B C T and P i p e l i n e M R ), the Wilcoxon rank-sum test is run over all the pairs of MAE predictions obtained for all the patients fed into the different pipelines. These tests reveal that both single-modality options performed significantly better than the mixed solution (p-values < 0.01). Figure 3 shows an example of the overlay of the sCT and its p M A E v o l u m e . As can be seen, for slices depicting small and movable anatomical structures (e.g., nasal cavities), higher MAE values are predicted. This behavior was expected and desirable, confirming the effectiveness of the proposed strategy.
Despite the necessity for a more extended training phase to enable the training of 5 different models, the inference step is completed in a few seconds. As a result, the required time to evaluate an sCT is totally compatible with a clinical scenario and will not slow down the radiotherapy planning workflow.

4. Discussion

In this work, we introduce a new method to predict the accuracy of sCT generation when the GT is not available (i.e. use of sCT in a clinical setting). The technique is based on using a DL model that is independent from any other model used to generate the sCT.
Being the first implemented workflow for the autonomous MAE prediction of DL-based sCT, there are no closely related studies to compare it with. The most similar articles in the literature are those that generate uncertainty maps alongside sCT [14,15,16]. The idea behind such works is to have a distribution of values for each voxel rather than a single intensity (by using Bayesian neural networks or by executing multiple inferences with active dropout layers). The averages of the obtained distributions represent the final sCT, while the variabilities are informative about the uncertainty of the prediction (with the uncertainty map correlating with the GT intensity and dose errors). However, in this regard, it is fundamental to underline the conceptual difference existing between our proposed workflow (conversion error estimation) and the uncertainty prediction proposed by others, since being confident about something does not necessarily mean being accurate. The main concern about these strategies is the possibility that the conversion model provides, with high assurance, a synthetic image that is actually wrong. Because of this, the main pillars of the proposed workflow are the separation between the image translation step and its quality control, and the prediction of the actual error rather than the confidence. Predicting the conversion error of sCT, in fact, is of paramount importance in contexts such as radiotherapy, where wrong HU assignments lead to wrong dose estimation to the target and organs at risk. Specifically, for the clinical introduction of MR-guided radio and proton therapy [24], an autonomous and independent quality control tool (an information with an additional level of redundancy and with a rationale totally different from the uncertainty assessment produced by the conversion algorithm itself) would be required to ensure reliable clinical decision-making based on deep-learning generated medical image data. We would like to highlight that the aim of the proposed tool is to intercept conversion errors that are not expected and not detectable by the human operator. Clinicians, in fact, well know the anatomical regions that can be more likely affected by translation error and are able to identify evident anomalous conversions, so our intention is to catch unforeseeable failures. The findings of our experiments show that it is possible to predict the accuracy of sCT in terms of MAE, with small deviations between inferred and GT MAE values. In fact, considering the HU range, an absolute prediction deviation of 4 ± 3 H U (for CBCT-derived sCT) and 6 ± 3.5 H U (for MR-derived sCT) has no impact on the dose estimation, even in the case of proton therapy, known to be extremely sensitive to HU variations. Signed prediction deviations demonstrate the absence of systematic errors. Results are also consistent with real-case expectations. Referring to Figure 3, worse conversion error is predicted for axial slices included between the nasal cavities and the mandibular bone. This result is very common in sCT generation, since the image quality in this anatomical area is deteriorated by the presence of dental fillings and motion artifacts. In a typical clinical setting, the user would receive a warning on which part of the sCT is reliable and which is not. In light of the magnitude of the predicted error and its spatial localization with respect to the tumor and the tissues to be spared, clinicians may elect to exclude the use of the sCT. The comparison between models trained by using single-modality versus mixed-modality dataset reveals that better results are obtained with single-modality training. The mixed setup was implemented to test the capability of the workflow to be agnostic to the initial image modality. The entire workflow is based on the VGG-16 model, which has been demonstrated to be the most effective architecture for predicting the MAE in our tests. In addition to its high predictive accuracy, this model requires fewer computational resources than deeper networks, allowing for fast sCT evaluation even in the absence of dedicated high-end GPUs, which is a crucial advantage in a real clinical setting.
Regarding the potential error propagation between the classification and regression steps, the proposed workflow is also robust to this possibility. In fact, as can be deduced from the results (Figure 5 and Figure 6), the prediction deviations are much smaller than the differences that would be generated if the raw MAE bin was misclassified and if this error was not recovered in the regression phase.
The proposed approach could be further improved by replacing the global axial MAE scalars currently predicted as quality control metrics with a more informative index from a clinical point of view. Such an index could, for example, include dosimetric and/or voxel-level prediction. In the future, it would be interesting to evaluate the proposed workflow on images collected in different institutions and converted by using different translation strategies (including cycle-GAN architectures, to get rid of the error introduced by the image registration step). We envision pursuing work in that direction, executing more complex and more computationally demanding experiments. The proposed idea, in addition to enabling a real-time synthetic image evaluation in the clinic, can be also used to implement synthetic image generation algorithms, when CT-MRI or CT-CBCT paired data are not available. The pipeline described here, in fact, can directly replace the computation of the MAE-based loss when the ground truth CT is not available.

5. Conclusions

In conclusion, we demonstrated that an independent prediction of the performance of an algorithm for sCT generation is possible and, most importantly, we hope to start a debate about usable strategies in real clinical environments for assessing the quality of synthetic medical images.

Author Contributions

P.Z.: Conceptualization, Investigation, Methodology, Software, Visualization, Formal analysis, and Writing—original draft; C.B.R.: Software, Writing—review and editing; A.T.: Data curation, Resources, and Writing—review and editing; G.G.M.: Data curation, Resources, and Writing—review and editing; J.A.L.: Supervision, Writing—review and editing; A.P.: Visualization, Writing—review and editing; C.C.: Supervision, Visualization, and Writing—review and editing; J.S.: Supervision, Writing—review and editing; A.C.K.: Supervision, Visualization, and Writing—review and editing; S.B.: Supervision, Resources, and Writing—review and editing; M.F.S.: Supervision, Methodology, Formal analysis, and Writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

All the patients were enrolled in the UMCG Radiation Oncology standardized follow-up program, approved by the medical ethics committee. Being a retrospective study, no additional images were acquired.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets used in this article are not readily available because they contain sensitive information. Requests to access the datasets should be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1 and Table A2 show, for each DL pipeline, the median and the percentiles of PD and APD distributions computed over all the 2D slices of all the patients.
Table A1. Statistics about the PD for each conducted experiment.
Table A1. Statistics about the PD for each conducted experiment.
Prediction Deviation (PD)
Experiment5th Percentile
[HU]
25th Percentile
[HU]
Median
[HU]
75th Percentile
[HU]
95th Percentile
[HU]
P i p e l i n e C B C T predicting on s C T C B C T −10−40520
P i p e l i n e M R predicting on s C T M R −12−41716
P i p e l i n e M I X E D predicting on s C T C B C T −23−6−1421
P i p e l i n e M I X E D predicting on s C T M R −11−341532
Table A2. Statistics about the APD for each conducted experiment.
Table A2. Statistics about the APD for each conducted experiment.
Absolute Prediction Deviation (APD)
Experiment5th Percentile
[HU]
25th Percentile
[HU]
Median
[HU]
75th Percentile
[HU]
95th Percentile
[HU]
P i p e l i n e C B C T predicting on s C T C B C T 024821
P i p e l i n e M R predicting on s C T M R 1361017
P i p e l i n e M I X E D predicting on s C T C B C T 0351129
P i p e l i n e M I X E D predicting on s C T M R 1481632

References

  1. Arabi, H.; Dowling, J.A.; Burgos, N.; Han, X.; Greer, P.B.; Koutsouvelis, N.; Zaidi, H. Comparative study of algorithms for synthetic CT generation from MRI: Consequences for MRI-guided radiation planning in the pelvic region. Med. Phys. 2018, 45, 5218–5233. [Google Scholar] [CrossRef] [PubMed]
  2. Johnstone, E.; Wyatt, J.J.; Henry, A.M.; Short, S.C.; Sebag-Montefiore, D.; Murray, L.; Kelly, C.G.; McCallum, H.M.; Speight, R. Systematic review of synthetic computed tomography generation methodologies for use in magnetic resonance imaging–only radiation therapy. Int. J. Radiat. Oncol. Biol. Phys. 2018, 100, 199–217. [Google Scholar] [CrossRef] [PubMed]
  3. Spadea, M.F.; Maspero, M.; Zaffino, P.; Seco, J. Deep learning based synthetic-CT generation in radiotherapy and PET: A review. Med. Phys. 2021, 48, 6537–6566. [Google Scholar] [CrossRef] [PubMed]
  4. Boulanger, M.; Nunes, J.C.; Chourak, H.; Largent, A.; Tahri, S.; Acosta, O.; De Crevoisier, R.; Lafond, C.; Barateau, A. Deep learning methods to generate synthetic CT from MRI in radiotherapy: A literature review. Phys. Medica 2021, 89, 265–281. [Google Scholar] [CrossRef] [PubMed]
  5. Dinkla, A.M.; Florkow, M.C.; Maspero, M.; Savenije, M.H.; Zijlstra, F.; Doornaert, P.A.; van Stralen, M.; Philippens, M.E.; van den Berg, C.A.; Seevinck, P.R. Dosimetric evaluation of synthetic CT for head and neck radiotherapy generated by a patch-based three-dimensional convolutional neural network. Med. Phys. 2019, 46, 4095–4104. [Google Scholar] [CrossRef]
  6. Liu, Y.; Lei, Y.; Wang, Y.; Wang, T.; Ren, L.; Lin, L.; McDonald, M.; Curran, W.J.; Liu, T.; Zhou, J.; et al. MRI-based treatment planning for proton radiotherapy: Dosimetric validation of a deep learning-based liver synthetic CT generation method. Phys. Med. Biol. 2019, 64, 145015. [Google Scholar] [CrossRef]
  7. Liu, Y.; Lei, Y.; Wang, T.; Fu, Y.; Tang, X.; Curran, W.J.; Liu, T.; Patel, P.; Yang, X. CBCT-based synthetic CT generation using deep-attention cycleGAN for pancreatic adaptive radiotherapy. Med. Phys. 2020, 47, 2472–2483. [Google Scholar] [CrossRef]
  8. Maspero, M.; Savenije, M.H.; Dinkla, A.M.; Seevinck, P.R.; Intven, M.P.; Jurgenliemk-Schulz, I.M.; Kerkmeijer, L.G.; Van Den Berg, C.A. Dose evaluation of fast synthetic-CT generation using a generative adversarial network for general pelvis MR-only radiotherapy. Phys. Med. Biol. 2018, 63, 185001. [Google Scholar] [CrossRef]
  9. Maspero, M.; Bentvelzen, L.G.; Savenije, M.H.; Guerreiro, F.; Seravalli, E.; Janssens, G.O.; van den Berg, C.A.; Philippens, M.E. Deep learning-based synthetic CT generation for paediatric brain MR-only photon and proton radiotherapy. Radiother. Oncol. 2020, 153, 197–204. [Google Scholar] [CrossRef]
  10. Dai, X.; Lei, Y.; Wynne, J.; Janopaul-Naylor, J.; Wang, T.; Roper, J.; Curran, W.J.; Liu, T.; Patel, P.; Yang, X. Synthetic CT-aided multiorgan segmentation for CBCT-guided adaptive pancreatic radiotherapy. Med. Phys. 2021, 48, 7063–7073. [Google Scholar] [CrossRef]
  11. Gao, L.; Xie, K.; Wu, X.; Lu, Z.; Li, C.; Sun, J.; Lin, T.; Sui, J.; Ni, X. Generating synthetic CT from low-dose cone-beam CT by using generative adversarial networks for adaptive radiotherapy. Radiat. Oncol. 2021, 16, 202. [Google Scholar] [CrossRef] [PubMed]
  12. Kazemifar, S.; McGuire, S.; Timmerman, R.; Wardak, Z.; Nguyen, D.; Park, Y.; Jiang, S.; Owrangi, A. MRI-only brain radiotherapy: Assessing the dosimetric accuracy of synthetic CT images generated using a deep learning approach. Radiother. Oncol. 2019, 136, 56–63. [Google Scholar] [CrossRef] [PubMed]
  13. van Elmpt, W.; Taasti, V.T.; Redalen, K.R. Current and future developments of synthetic computed tomography generation for radiotherapy. Phys. Imaging Radiat. Oncol. 2023, 28, 100521. [Google Scholar] [CrossRef]
  14. Hemsley, M.; Chugh, B.; Ruschin, M.; Lee, Y.; Tseng, C.L.; Stanisz, G.; Lau, A. Deep generative model for synthetic-CT generation with uncertainty predictions. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2020: 23rd International Conference, Lima, Peru, 4–8 October 2020; Proceedings, Part I 23. Springer: Cham, Switzerland, 2020; pp. 834–844. [Google Scholar]
  15. Li, X.; Bellotti, R.; Meier, G.; Bachtiary, B.; Weber, D.; Lomax, A.; Buhmann, J.; Zhang, Y. Uncertainty-aware MR-based CT synthesis for robust proton therapy planning of brain tumour. Radiother. Oncol. 2024, 191, 110056. [Google Scholar] [CrossRef] [PubMed]
  16. Galapon, A.V., Jr.; Thummerer, A.; Langendijk, J.A.; Wagenaar, D.; Both, S. Feasibility of Monte Carlo dropout-based uncertainty maps to evaluate deep learning-based synthetic CTs for adaptive proton therapy. Med. Phys. 2024, 51, 2499–2509. [Google Scholar] [CrossRef]
  17. Thummerer, A.; Zaffino, P.; Meijers, A.; Marmitt, G.G.; Seco, J.; Steenbakkers, R.J.; Langendijk, J.A.; Both, S.; Spadea, M.F.; Knopf, A.C. Comparison of CBCT based synthetic CT methods suitable for proton dose calculations in adaptive proton therapy. Phys. Med. Biol. 2020, 65, 095002. [Google Scholar] [CrossRef]
  18. Thummerer, A.; De Jong, B.A.; Zaffino, P.; Meijers, A.; Marmitt, G.G.; Seco, J.; Steenbakkers, R.J.; Langendijk, J.A.; Both, S.; Spadea, M.F.; et al. Comparison of the suitability of CBCT-and MR-based synthetic CTs for daily adaptive proton therapy in head and neck patients. Phys. Med. Biol. 2020, 65, 235036. [Google Scholar] [CrossRef]
  19. Spadea, M.F.; Pileggi, G.; Zaffino, P.; Salome, P.; Catana, C.; Izquierdo-Garcia, D.; Amato, F.; Seco, J. Deep convolution neural network (DCNN) multiplane approach to synthetic CT generation from MR images—Application in brain proton therapy. Int. J. Radiat. Oncol. Biol. Phys. 2019, 105, 495–503. [Google Scholar] [CrossRef]
  20. Zaffino, P.; Ciardo, D.; Raudaschl, P.; Fritscher, K.; Ricotti, R.; Alterio, D.; Marvaso, G.; Fodor, C.; Baroni, G.; Amato, F.; et al. Multi atlas based segmentation: Should we prefer the best atlas group over the group of best atlases? Phys. Med. Biol. 2018, 63, 12NT01. [Google Scholar] [CrossRef]
  21. Simonyan, K. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2016; pp. 770–778. [Google Scholar]
  23. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, Canada, 8–14 December 2019; pp. 8026–8037. [Google Scholar]
  24. Hoffmann, A.; Oborn, B.; Moteabbed, M.; Yan, S.; Bortfeld, T.; Knopf, A.; Fuchs, H.; Georg, D.; Seco, J.; Spadea, M.F.; et al. MR-guided proton therapy: A review and a preview. Radiat. Oncol. 2020, 15, 129. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Representation of the general MAE prediction pipeline. An axial sCT slice is given as input, and the associated MAE scalar for the image slice is predicted by using a DL pipeline.
Figure 1. Representation of the general MAE prediction pipeline. An axial sCT slice is given as input, and the associated MAE scalar for the image slice is predicted by using a DL pipeline.
Jimaging 10 00316 g001
Figure 2. A more detailed graphical representation of the MAE prediction pipeline. The final MAE prediction is obtained as a result of two DL steps: First a raw MAE interval classification is performed, followed by a more precise MAE estimation based on a regression algorithm.
Figure 2. A more detailed graphical representation of the MAE prediction pipeline. The final MAE prediction is obtained as a result of two DL steps: First a raw MAE interval classification is performed, followed by a more precise MAE estimation based on a regression algorithm.
Jimaging 10 00316 g002
Figure 3. Exemplary s C T C B C T overlaid with its p M A E v o l u m e . In addition to the 2D views (axial, sagittal, and coronal planes), the 3D representation is also shown.
Figure 3. Exemplary s C T C B C T overlaid with its p M A E v o l u m e . In addition to the 2D views (axial, sagittal, and coronal planes), the 3D representation is also shown.
Jimaging 10 00316 g003
Figure 4. Detailed workflow of MAE prediction. A single sCT axial slice is fed firstly into a DL model that classifies it as belonging to a specific MAE class. According to this prediction, the 2D image is then provided as input to a connected DL regression model, specifically trained to operate on a restricted range of MAE values. As a result, the MAE of a single sCT slice can be forecasted. In order to train the different models with a GT MAE, the ground truth CT is needed (dashed lines are needed only to train the models).
Figure 4. Detailed workflow of MAE prediction. A single sCT axial slice is fed firstly into a DL model that classifies it as belonging to a specific MAE class. According to this prediction, the 2D image is then provided as input to a connected DL regression model, specifically trained to operate on a restricted range of MAE values. As a result, the MAE of a single sCT slice can be forecasted. In order to train the different models with a GT MAE, the ground truth CT is needed (dashed lines are needed only to train the models).
Jimaging 10 00316 g004
Figure 5. PD distributions for modality-specific and mixed pipelines. Results for s C T C B C T and s C T M R are reported, respectively, in the left and in the right panel.
Figure 5. PD distributions for modality-specific and mixed pipelines. Results for s C T C B C T and s C T M R are reported, respectively, in the left and in the right panel.
Jimaging 10 00316 g005
Figure 6. APD distributions for modality-specific and mixed pipelines. Results for s C T C B C T and s C T M R are reported, respectively, in the left and in the right panel.
Figure 6. APD distributions for modality-specific and mixed pipelines. Results for s C T C B C T and s C T M R are reported, respectively, in the left and in the right panel.
Jimaging 10 00316 g006
Table 1. MAE ranges for classification and regression, both for MR, CBCT, and MIXED pipelines.
Table 1. MAE ranges for classification and regression, both for MR, CBCT, and MIXED pipelines.
PipelineLow MAEMedium-Low MAEMedium-High MAEHigh MAE
MR0–47
(classification)
47–54
(classification)
54–68
(classification)
68–100
(classification)
0–52
(regression)
42–59
(regression)
49–73
(regression)
63–100
(regression)
CBCT0–27
(classification)
27–32
(classification)
32–42
(classification)
42–70
(classification)
0–32
(regression)
22–37
(regression)
27–47
(regression)
37–70
(regression)
MIXED0–32
(classification)
32–44
(classification)
44–56
(classification)
56–90
(classification)
0–37
(regression)
27–49
(regression)
39–61
(regression)
51–90
(regression)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zaffino, P.; Raggio, C.B.; Thummerer, A.; Marmitt, G.G.; Langendijk, J.A.; Procopio, A.; Cosentino, C.; Seco, J.; Knopf, A.C.; Both, S.; et al. Toward Closing the Loop in Image-to-Image Conversion in Radiotherapy: A Quality Control Tool to Predict Synthetic Computed Tomography Hounsfield Unit Accuracy. J. Imaging 2024, 10, 316. https://doi.org/10.3390/jimaging10120316

AMA Style

Zaffino P, Raggio CB, Thummerer A, Marmitt GG, Langendijk JA, Procopio A, Cosentino C, Seco J, Knopf AC, Both S, et al. Toward Closing the Loop in Image-to-Image Conversion in Radiotherapy: A Quality Control Tool to Predict Synthetic Computed Tomography Hounsfield Unit Accuracy. Journal of Imaging. 2024; 10(12):316. https://doi.org/10.3390/jimaging10120316

Chicago/Turabian Style

Zaffino, Paolo, Ciro Benito Raggio, Adrian Thummerer, Gabriel Guterres Marmitt, Johannes Albertus Langendijk, Anna Procopio, Carlo Cosentino, Joao Seco, Antje Christin Knopf, Stefan Both, and et al. 2024. "Toward Closing the Loop in Image-to-Image Conversion in Radiotherapy: A Quality Control Tool to Predict Synthetic Computed Tomography Hounsfield Unit Accuracy" Journal of Imaging 10, no. 12: 316. https://doi.org/10.3390/jimaging10120316

APA Style

Zaffino, P., Raggio, C. B., Thummerer, A., Marmitt, G. G., Langendijk, J. A., Procopio, A., Cosentino, C., Seco, J., Knopf, A. C., Both, S., & Spadea, M. F. (2024). Toward Closing the Loop in Image-to-Image Conversion in Radiotherapy: A Quality Control Tool to Predict Synthetic Computed Tomography Hounsfield Unit Accuracy. Journal of Imaging, 10(12), 316. https://doi.org/10.3390/jimaging10120316

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop