Next Article in Journal
Methylated Circulating Tumor DNA in Blood as a Tool for Diagnosing Lung Cancer: A Systematic Review and Meta-Analysis
Next Article in Special Issue
Downfield Proton MRSI at 3 Tesla: A Pilot Study in Human Brain Tumors
Previous Article in Journal
Vaccination against Extracellular Vimentin for Treatment of Urothelial Cancer of the Bladder in Client-Owned Dogs
Previous Article in Special Issue
The Impact of ioMRI on Glioblastoma Resection and Clinical Outcomes in a State-of-the-Art Neuro-Oncological Setup
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Fully Automated Post-Surgical Brain Tumor Segmentation Model for Radiation Treatment Planning and Longitudinal Tracking

1
Department of Radiation Oncology, Emory University School of Medicine, Atlanta, GA 30322, USA
2
Department of Biomedical Engineering, Emory University and Georgia Institute of Technology, Atlanta, GA 30332, USA
3
Siemens Healthineers, Atlanta, GA 30329, USA
4
Department of Radiation Oncology, Johns Hopkins University, Baltimore, MD 21218, USA
5
Department of Radiation Oncology, Sylvester Comprehensive Cancer Center, Miller School of Medicine, University of Miami, Miami, FL 45056, USA
6
Winship Cancer Institute, Emory University School of Medicine, Atlanta, GA 30322, USA
7
Department of Radiology and Imaging Sciences, Emory University School of Medicine, Atlanta, GA 30322, USA
*
Author to whom correspondence should be addressed.
Cancers 2023, 15(15), 3956; https://doi.org/10.3390/cancers15153956
Submission received: 29 June 2023 / Revised: 26 July 2023 / Accepted: 2 August 2023 / Published: 3 August 2023
(This article belongs to the Special Issue Advanced Imaging in Brain Tumor Patient Management)

Abstract

:

Simple Summary

With several previous efforts to segment pre-surgical brain tumor lesions from MRI, we sought to shine a different light on the problem. Radiation treatment planning post-surgery still relies heavily on manual contouring of T1-weighted contrast-enhanced and T2-weighted fluid-attenuated inversion recovery MRI. This is one of the first attempts to segment post-surgical brain lesions through deep learning approaches for radiation treatment planning and longitudinal tracking. Our best-performing model segments an overwhelming majority of lesions with at least 70% accuracy and has already been integrated into a web application heavily used by physicians and researchers for longitudinal tracking.

Abstract

Glioblastoma (GBM) has a poor survival rate even with aggressive surgery, concomitant radiation therapy (RT), and adjuvant chemotherapy. Standard-of-care RT involves irradiating a lower dose to the hyperintense lesion in T2-weighted fluid-attenuated inversion recovery MRI (T2w/FLAIR) and a higher dose to the enhancing tumor on contrast-enhanced, T1-weighted MRI (CE-T1w). While there have been several attempts to segment pre-surgical brain tumors, there have been minimal efforts to segment post-surgical tumors, which are complicated by a resection cavity and postoperative blood products, and tools are needed to assist physicians in generating treatment contours and assessing treated patients on follow up. This report is one of the first to train and test multiple deep learning models for the purpose of post-surgical brain tumor segmentation for RT planning and longitudinal tracking. Post-surgical FLAIR and CE-T1w MRIs, as well as their corresponding RT targets (GTV1 and GTV2, respectively) from 225 GBM patients treated with standard RT were trained on multiple deep learning models including: Unet, ResUnet, Swin-Unet, 3D Unet, and Swin-UNETR. These models were tested on an independent dataset of 30 GBM patients with the Dice metric used to evaluate segmentation accuracy. Finally, the best-performing segmentation model was integrated into our longitudinal tracking web application to assign automated structured reporting scores using change in percent cutoffs of lesion volume. The 3D Unet was our best-performing model with mean Dice scores of 0.72 for GTV1 and 0.73 for GTV2 with a standard deviation of 0.17 for both in the test dataset. We have successfully developed a lightweight post-surgical segmentation model for RT planning and longitudinal tracking.

1. Introduction

Due to its heavy reliance on imaging, treatment and follow up of brain tumors is an area that could benefit heavily from artificial intelligence (AI) physician guidance. For example, glioblastoma (GBM) is a severe form of brain cancer with a median survival of 15–16 months [1,2,3,4,5]. Standard treatment of GBM first requires resection of as much contrast-enhancing tumor on contrast-enhanced T1-weighted (CE-T1w) MRI as possible. The remaining tumor is treated with adjuvant temozolomide chemotherapy coupled with radiation therapy that targets a high dose of radiation to any residual enhancing tumor in CE-T1w and a lower dose towards hyperintense edema on T2-weighted fluid-attenuated inversion recovery (T2w/FLAIR) MRI. Each of these steps is highly dependent on imaging interpretation, and clinically assistive AI tools could help physicians track and treat GBM patients more accurately and more quickly.

1.1. Related Work

While there have been several efforts to segment brain lesions from MRIs, such as through the Brain Tumor Segmentation Challenge (BraTS), those efforts have focused on pre-surgical brain MRIs, and have yet to make a substantial translation into clinically useful tools [6,7,8,9]. There have been minimal efforts to develop segmentation algorithms to assist RT planning after surgery. Post-surgical MRIs have altered morphology that includes a cavity with blood product and heterogenous hyperintensity on FLAIR. In Figure 1, we show an example patient with CE-T1w and T2w/FLAIR MRIs prior to surgery, shortly after completing RT, and three months after completing RT. The CE-T1w MRIs often only contain trace amounts of residual enhancing tumor post-surgery. The cavity coupled with the small remaining lesion leads to a more difficult task for segmentation models to capture the lesion and the entire cavity. Overlaid on each MRI is the segmentation from the highest-performing Swin-UNETR model trained on the BraTS dataset [10]. We only show these segmentations not to critique the model or approach, but to suggest that previous efforts of brain tumor segmentation solve a different problem and may not be generally applicable to treated tumors.
In standard clinical practice, radiation oncologists generate gross tumor volume (GTV) targets that encompass the residual enhancing lesion and cavity in CE-T1w MRI (GTV2) and edema and cavity on T2w/FLAIR (GTV1). Due to the infiltrative nature of GBM, these targets usually have a margin added to generate clinical target volumes (CTVs) and planning treatment volumes (PTVs). There have been publications that focused on cavity segmentation and performing CTV segmentations with GTVs as input [11,12,13]. However, this is one of the first efforts to directly train a model for GTV segmentation. Furthermore, previous brain tumor segmentation approaches utilized four sets of images (T1w-pre contrast, CE-T1w, T2w, and T2w/FLAIR) for segmenting lesions in CE-T1w and T2w/FLAIR. In an effort to develop a lighter-weight model, we are attempting to use only CE-T1w and T2w/FLAIR to perform GTV segmentations. Such models that require fewer inputs could be more clinically translatable and require less effort due to lesser data requirements for segmentation by physicians and researchers.

1.2. Purpose

Due to the unique visual morphology of post-surgical lesions, and a clinically unmet segmentation approach for these brain MRIs, we sought to develop a lightweight deep learning algorithm that could assist radiation oncologists in generating RT volumes. Finally, with a sophisticated deep learning segmentation algorithm for post-operative brain tumors, we hope to integrate our algorithm within our web application, the Longitudinal Imaging Tracker (BrICS-LIT), to track post-treatment brain tumors and assign automated structured reporting scores [14]. In Table 1, we summarize the added value of our proposed effort compared to previous approaches of brain lesion segmentation. With BrICS-LIT, physicians can visualize changes in brain lesions over time, as well as assign structured reporting scores such as the Response Assessment in Neuro-Oncology (RANO) and Brain Tumor Reporting and Data System (BT-RADS) to classify their findings by patient disease-state [15,16,17,18,19,20,21].

2. Materials and Methods

2.1. Preparation of Training Data

Imaging from a de-identified database of 225 GBM patients who were treated at Emory University with intensity modulated radiation therapy over the past 10 years was used to train our segmentation model. For each patient, post-operative T2w/FLAIR and CE-T1w MRIs were collected as well as their corresponding GTV1 and GTV2 contours. In Figure 2, we show an example patient in our training dataset where the extent of GTV1 includes not only the hyperintense lesion, but also cavity and blood product. In the same figure, GTV2 in CE-T1w MRI includes the cavity and residual enhancing lesion. As per standard-of-care, GTV2 typically receives a higher radiation dose of 60 Gy, while GTV1 receives a lower dose of 50–51 Gy [2,3]. Due to the range of possible margins physicians add to their GTV targets, we had a radiation oncologist resident tighten the treatment contours to only include visible lesions in both sets of images. These tightened contours were then reviewed by an independent radiation oncologist and neuroradiologist with any differences resolved by consensus until agreement was reached.
For image preprocessing, T2w/FLAIR MRIs were registered and interpolated to their CE-T1w counterparts. All images were skull-stripped and resampled to a (256,256,160) volume before nonzero voxels underwent zero-mean, unit-variance normalization. For GTV1 segmentation, models were fed a 3-channel input with skull-stripped T2w/FLAIR MRI as the first channel, skull-stripped CE-T1w MRI as the second channel, and with-skull T2w/FLAIR MRI as the third channel. For GTV2 segmentation, CE-T1w MRI was used for the first and third channels, while skull-stripped T2w/FLAIR was used for the middle channel (Figure 3A). A third channel with skull was included to help the model delineate the extent of resection cavities that reside close to the skull. For 2D segmentation, slices of size (256,256) were fed in batches for training, while 3D segmentation models trained on volumes of size (128,128,128). The 225-patient dataset was split into a training set of 202 patients and a validation set of 23 patients. For the training data, on-the-fly augmentation was performed with random flipping and rotation with 0.5 probability. An independent test dataset containing T2w/FLAIR and CE-T1w MRIs with corresponding GTVs for 30 post-operative GBM patients from three different sites (Emory University, Johns Hopkins University, and University of Miami) was used to evaluate segmentation performance (Figure 3A).

2.2. Segmentation Models and Training

We compared 2D and 3D segmentation approaches with the goal of incorporating the most robust models in clinical workflows. Three of the most popular 2D segmentation models for medical image segmentation are the standard U-net, ResUnet, and Shifted-Window (Swin)-Unet [22,23,24]. The ResUnet model has the structure of Unet but with added residual connections in the encoder and decoder blocks to prevent vanishing gradient problems, as well as atrous convolutions and pyramidal pooling to garner more information from larger receptive fields. The 2D Swin-Unet structure is similar to Unet, but replaces standard convolutional neural network (CNN)-based encoder and decoder blocks with Swin transformers, which use tokenized image patches as input and attention-based shifted-window vision transformers to garner global image features [22,25]. Finally, we compared the CNN-based 3D Unet with Swin-UNETR—both 3D segmentation models that have found great success with pre-operative brain tumor segmentation [9,10,26]. The 3D Unet is similar in structure to Unet but uses 3D filters to convolve over image volumes as inputs (Figure 3B), while Swin-UNETR uses 3D Swin transformers as encoders and standard CNNs for decoding and generating the segmentations.
For the 2D models and 3D Unet, a five-fold cross validation procedure was used to determine the optimal hyperparameters for each model [27]. We used the same hyperparameters as the authors of Swin-UNETR for brain tumor segmentation [10]. A table of key hyperparameters optimized during cross validation are in Supplemental Table S1. The Dice similarity coefficient was used to evaluate segmentation accuracy between model predictions and physician-generated segmentations and Dice loss was used during training [28]. During training, a learning rate scheduler and early stopping criteria were used to reduce learning rate based on validation accuracy to fine-tune optimization, and to stop training when validation accuracy had not improved for 10 epochs. Three-dimensional models were trained on an NVIDIA RTX 6000 with GPU memory of 48 GB, while 2D models were trained on an NVIDIA V100 with GPU memory of 32 GB. Models were then tested on the independent 30-patient test dataset to calculate test Dice scores, Hausdorff distances, and Jaccard coefficients for comparison and evaluation.

2.3. Application of Segmentation in Longitudinal Tracking

We incorporated our best-performing segmentation algorithms into our web application BrICS-LIT for longitudinal tracking and calculated lesion volumes for an example patient. Due to our model including the resection cavity in CE-T1w lesion segmentation, we use Otsu thresholding to cluster GTV2 segmentations into four different groups before removing the lowest intensity cluster to only include the residual enhancing lesion for longitudinal tracking purposes [29]. We then used the same volumetric percent change cutoffs used to predict RANO scores from Kickingereder et al. for BT-RADS prediction [16]. More specifically, if a T2w/FLAIR lesion changed in volume by 100% or greater or a CE-T1w lesion increased in volume by 40% or greater, then tumor had recurred and BT-RADS score of 4 was assigned. If CE-T1w lesion increased in volume by 20% or greater or T2w/FLAIR lesion volume increased by 50% or greater, then imaging was considered “worsened” and BT-RADS scores of 3 or higher were assigned. Otherwise, imaging was considered improved or stable and scores of 1 or 2 were given. Finally, if for two consecutive visits a patient’s imaging had worsened by less than the tumor recurrence cutoff, a score of 3c was assigned to the latest patient visit, suggesting they are highly likely to experience tumor recurrence. By using these cutoffs in a decision-tree-based approach, along with the lesion volumes and relevant clinical and medication information retrieved from REDCap, BrICS-LIT makes automated disease-state classifications.

3. Results

3.1. Segmentation for RT Planning

In Table 2, we compare our 2D and 3D fully automated segmentation models. When comparing the purely CNN-based 2D Unet and Resunet models to the 2D Swin-Unet model, we see a similar segmentation performance during training for both GTVs. However, the Resunet outperforms the 2D Swin-Unet model for GTV2 segmentation in the test dataset with a Dice score of 0.57. The 2D Swin-Unet model outperforms the other 2D approaches for GTV1 segmentation with a test Dice score of 0.64, which is comparable to even the 3D approaches.
While both 3D Unet and Swin-UNETR have found great success in brain tumor segmentation tasks in the past, the 3D Unet outperformed 3D Swin-UNETR for our task of segmenting radiation treatment contours and, importantly, this was including the surgical resection cavity. Interestingly, with the early stopping criteria we had in place for each training task, the Swin-UNETR model converged at a lower training Dice score of 0.64 for GTV1 and 0.65 for GTV2, whereas the 3D Unet model converged at higher GTV1 and GTV2 training Dice scores of 0.77 and 0.79, respectively. The difference in Dice scores between training and testing was at most 0.06 for the 3D models. The table also includes the Hausdorff distance and Jaccard coefficient for the test dataset. A higher Hausdorff distance indicates a greater disagreement between prediction and ground truth as it calculates the largest distance between a point in the ground truth and its closest point in the prediction, while the Jaccard coefficient generally scales with the Dice score but penalizes poor predictions more, and is therefore consistently lower in the table.
In Figure 4, we show example GTV segmentations from our best-performing model, the 3D Unet. The first two columns contain CE-T1w MRIs and the second two columns contain T2w/FLAIR MRIs for three patients from three different institutions. We show a 2D contour in axial orientation as well as a 3D volume rendering of the overlap between GTV prediction and clinician-derived contours in sagittal orientation. Our model’s GTV predictions are almost identical to those created by clinical experts, and we believe these examples illustrate the unique task that our model was trained to solve. For both GTVs, the resection cavity was included, as it typically receives a higher dose of radiation. During radiation treatment planning, clinicians can use the segmentations from CE-T1w to generate high-dose GTV2 contours and subtract the GTV2 contour from the GTV1 contour to generate a lower dose targeted towards edema and infiltrative disease.
When evaluating the performance of the 3D Unet model further, about 75% of the test patients had a Dice score greater than 0.70 for GTV1 and greater than 0.67 for GTV2. This suggests that outliers in roughly a quarter of the test dataset caused the average performance of the model to decrease. To that end, Figure 5 shows some of the poorer-performing examples. In Figure 5A, while the model appears to segment the T2w/FLAIR hyperintensity and blood product quite well, it completely misses the small enhancing lesion in CE-T1w MRI, possibly missing it due to a lack of cavity nearby or its relatively small size. Figure 5B is an example where GTV2 is segmented in CE-T1w MRI, but the model underestimates the extent of T2w/FLAIR hyperintensity. This is possibly due to the relatively low contrast between abnormal and normal tissue on some of the 3D T2w/FLAIR acquisitions. Figure 5C is a case where the model confused the ventricle for a cavity as there was hyperintense T2w/FLAIR lesion in the same area.

3.2. Application of Segmentation in Longitudinal Tracking

With a fully automated segmentation model trained on post-operative MRIs, we sought to test our model in our web application BrICS-LIT for automated longitudinal tracking. To that end, Figure 6 shows an example patient’s pre-operative/pre-RT and post-RT follow-up MRIs. Our 3D Unet model performed segmentations of T2w/FLAIR and CE-T1w lesions, and after Otsu thresholding, the cavity was removed from the CE-T1w segmentations. Then, volumetric cutoffs that were used for RANO classification were applied to this patient’s lesion volumes. As per BT-RADS, the baseline score at post-surgery, on the far right, received a score of 0. In the next follow-up visit after the completion of RT, the CE-T1w lesion grew by almost 25% while the FLAIR lesion grew by slightly over 25%. A score of 3A was automatically given as imaging had worsened, but the visit was shortly after the completion of radiation treatment, and the increased volumes were attributed to radiation-related inflammation. In the following visit, lesion volumes increased by 37% and 27% for T2w/FLAIR and CE-T1w lesions, respectively. Due to the lesion volumes increasing for consecutive dates, a score of 3C was given, suggesting the lesion was highly likely to experience tumor recurrence. In the next visit, the T2w/FLAIR lesion volume had decreased but enhancement was similar or slightly larger. These mixed findings resulted in assigning the indeterminate BT-RADS score of 3B. On the latest visit, the CE-T1w lesion volume increased by 70%, well over the 40% cutoff, immediately causing BrICS-LIT to assign a score of 4, highly probable for tumor recurrence. This also meets the criteria for RANO progressive disease.
Figure 7 shows a graph of changes in T2w/FLAIR and CE-T1w lesion over time, as well as corresponding BT-RADS scores for the same patient generated by BrICS-LIT for physician viewing. The sharp increase in both lesion volumes is apparent 6 months after surgery, signaling true tumor recurrence.

4. Discussion

With the advent of artificial intelligence in many domains of our daily lives, a question remains as to when it will be included in standard-of-care medical treatments. While there has been an entrepreneurial boom in integrating deep learning and machine learning to the domain of personalized medicine, there are still many challenges in including AI-based tools in clinical imaging workflows [30,31,32]. Some of these challenges include the need for physicians to have efficient image processing software that can easily pre-process clinical imaging, handle computationally heavy algorithms, and perform inference, all while requiring minimal effort by the physician [33]. Furthermore, visualization tools can help physicians understand the reasoning behind predictions and decisions that AI algorithms make [33]. To that end, we trained the first deep learning segmentation algorithm for post-operative brain tumors for the purpose of radiation treatment planning and longitudinal tracking. With reliable GTV1 and GTV2 segmentation models that can perform segmentations in less than a minute, radiation oncologists can expediently generate GTV targets and add margins depending on their preferences to generate larger CTVs and PTVs for radiation treatment. These tools could potentially be integrated into radiation planning software commonly used by radiation oncologists. By using volumetric cutoffs, we can suggest completely automated disease-state classifications for the longitudinal tracking of post-treatment brain tumor patients, which we implemented using the BT-RADS rating scale in our online BrICS-LIT platform.
In comparing our 2D and 3D segmentation approaches, it is evident that the 2D models are overtrained, as the training Dice scores were above 0.90 while the testing Dice scores ranged between 0.43–0.64. We set early stopping criteria which would cause our models to stop training when the validation accuracy failed to improve after 10 epochs. This suggests that the 2D models were still learning some generalizable information, even if the gap between training and validation accuracy was steadily increasing. The benefit of using 3D approaches is that while their training accuracy was lower than the 2D models, the testing accuracy was somewhat aligned with the training accuracy, leading us to believe the 3D approaches are more robust with test images.
Our best-performing segmentation model was the CNN-based 3D Unet, with test Dice scores of 0.72 and 0.73 for GTV1 and GTV2, respectively. While the distribution of Dice scores in the test dataset suggests that there were a few cases that severely decreased the mean Dice score, we believe this gives us room for improvement. The models struggled with small residual enhancing lesions in CE-T1w with minimal cavities, which can easily be confused for blood vessels. They also struggled with the task of differentiating cavity from ventricle, especially when there was T2w/FLAIR hyperintense lesion in the same area. Finally, the model appears to struggle with capturing the entire extent of the lesion boundary in 3D T2w/FLAIR, where the acquisitions prioritize higher spatial resolution over contrast. Although we implemented several augmentation techniques such as random intensity shifting and random k-space sampling to mimic various levels of spatial resolution, we found that augmentation did not improve our models’ ability to generalize to the test data. Future efforts will involve curating a more diverse training dataset that also includes smaller residual enhancing lesions and a higher number of 3D T2w/FLAIR acquisitions. Ensuring that models have good-quality, high-contrast 3D FLAIR acquisitions can also help with segmentation.
Many brain tumor segmentation efforts use four sets of images to help segment lesions—T1w pre-contrast MRI, CE-T1w, T2w MRI, and T2w/FLAIR. The Swin-UNETR model achieved high Dice scores with the pre-operative BRATS dataset by utilizing all four sets of images. The encoder region of the model utilizes Swin transformers, which in theory would help the model determine affinities between different regions of the image. We believe it may have underperformed with our smaller training dataset with fewer image channels due to its complexity. A model of that complexity may have overfit the dataset, and with our early stopping criteria, converged too quickly even with small learning rates compared to the 3D Unet model. We chose to train our models with only CE-T1w and T2w/FLAIR MRI in order for our models to fit within clinical workflows with minimal effort, since CE-T1w and T2w/FLAIR are the sequences most heavily relied on by radiologists, radiation oncologists, and neurosurgeons to identify tumors. When housing our models in web applications like BrICS-LIT or even attempting to integrate them in clinical software like Velocity (Varian Medical Systems), we wanted our models to have as few requirements as possible to enable easy integration. For example, our goal with BrICS-LIT is the longitudinal tracking of brain tumor patients in clinical and research settings. It is far easier for researchers and clinicians to use our models in BrICS-LIT if the models only require the imaging that the clinicians were currently viewing as inputs. We do acknowledge that extra imaging, and most importantly the T1w pre-contrast MRI, could potentially improve our models’ segmentation accuracy. To address this concern, efforts are currently underway to train generative adversarial networks (GANs) on the BraTS 2021 dataset to artificially generate T1w-pre contrast and T2w MRIs from CE-T1w and T2w/FLAIR MRIs [34,35,36]. With a complete dataset from four imaging modalities, we plan to use previous BraTS competition-winning models that are trained on pre-operative MRIs and perform transfer learning on our dataset to finetune the task of including cavities, blood product, and the entire extent of hyperintense edemas.
As demonstrated in Figure 6, we have already integrated our model in our web application (BrICS-LIT) that is actively used by clinicians and researchers. By performing automated lesion segmentation and using quantitative volumetric cutoffs, we have shown our post-operative segmentation algorithms can help us generate automated disease-state classifications for BT-RADS and RANO. While borrowing volumetric cutoffs used for RANO classification have led to interesting BT-RADS classifications, we acknowledge that both structured reporting criteria may require different volumetric cutoffs that are more suitable for each. To mitigate this issue, we have a de-identified, independent database of over 150 patients with clinician-assigned BT-RADS scores that we will treat as the gold standard to help determine new volumetric cutoffs specific to BT-RADS. Furthermore, simply using percentage changes in lesion volume has its own limitations. For example, a patient could have a residual enhancing lesion that is only 0.4 cc. In their next visit, that lesion could grow by 50% to 0.6 cc. While our automated algorithm would likely say the tumor has recurred purely based on percent change, that 0.2 cc increase in lesion volume could simply be a rounding error in our segmentation algorithm. To address this point, we are investigating the use of classification and regression tree-based machine-learning approaches to predict disease-state classifications using lesion volumes and other relevant clinical information.

5. Conclusions

While there have been several successful efforts in the past to segment brain tumor lesions in an automated manner from MRIs, they tend to focus on pre-operative brain MRIs, which may limit their usefulness in clinical use and in clinical trials. Here, we have trained popular medical image segmentation algorithms for post-operative, residual brain tumors for the purpose of radiation treatment planning and longitudinal tracking. The best-performing 3D Unet can segment lesions with over 70% accuracy for most of our test cases, and has already been integrated into our BrICS-LIT web application for clinician and research use.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers15153956/s1, Table S1: Hyperparameters for every trained model.

Author Contributions

Conceptualization, K.K.R., H.S. and B.D.W.; Data curation, K.M.X., A.G.T., L.R.K., E.A.M., H.-K.G.S., H.S. and B.D.W.; Formal analysis, K.K.R.; Methodology, K.K.R., V.H., V.K.S., H.-K.G.S. and H.S.; Software, K.K.R.; Validation, A.G.T. and H.S.; Visualization, K.K.R.; Writing—original draft, K.K.R.; Writing—review & editing, K.K.R., A.G.T., V.H., V.K.S., L.R.K., E.A.M., H.-K.G.S., H.S. and B.D.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the NIH U01CA264039 (H.S., E.A.M.), NIH R01CA214557 (H.S., L.R.K., H-G.S.), and pre-doctoral fellowship NIH F31CA247564 (K.K.R.).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Review Board (or Ethics Committee) of Emory University.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors would like to acknowledge the amazing work and dedication of our Institutional Review Board staff, as well as thank all our patients and their caregivers, who believed in our study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gilbert, M.R.; Wang, M.; Aldape, K.D.; Stupp, R.; Hegi, M.; Jaeckle, K.A.; Armstrong, T.S.; Wefel, J.S.; Won, M.; Blumenthal, D.T.; et al. RTOG 0525: A randomized phase III trial comparing standard adjuvant temozolomide (TMZ) with a dose-dense (dd) schedule in newly diagnosed glioblastoma (GBM). J. Clin. Oncol. 2011, 29, 2006. [Google Scholar] [CrossRef] [Green Version]
  2. Stupp, R.; Hegi, M.E.; Mason, W.P.; van den Bent, M.J.; Taphoorn, M.J.; Janzer, R.C.; Ludwin, S.K.; Allgeier, A.; Fisher, B.; Belanger, K.; et al. Effects of radiotherapy with concomitant and adjuvant temozolomide versus radiotherapy alone on survival in glioblastoma in a randomised phase III study: 5-year analysis of the EORTC-NCIC trial. Lancet Oncol. 2009, 10, 459–466. [Google Scholar] [CrossRef]
  3. Stupp, R.; Mason, W.P.; van den Bent, M.J.; Weller, M.; Fisher, B.; Taphoorn, M.J.; Belanger, K.; Brandes, A.A.; Marosi, C.; Bogdahn, U.; et al. Radiotherapy plus concomitant and adjuvant temozolomide for glioblastoma. N. Engl. J. Med. 2005, 352, 987–996. [Google Scholar] [CrossRef] [Green Version]
  4. Stupp, R.; Taillibert, S.; Kanner, A.; Read, W.; Steinberg, D.; Lhermitte, B.; Toms, S.; Idbaih, A.; Ahluwalia, M.S.; Fink, K.; et al. Effect of Tumor-Treating Fields Plus Maintenance Temozolomide vs Maintenance Temozolomide Alone on Survival in Patients With Glioblastoma: A Randomized Clinical Trial. JAMA 2017, 318, 2306–2316. [Google Scholar] [CrossRef] [Green Version]
  5. Stupp, R.; Taillibert, S.; Kanner, A.A.; Kesari, S.; Steinberg, D.M.; Toms, S.A.; Taylor, L.P.; Lieberman, F.; Silvani, A.; Fink, K.L.; et al. Maintenance Therapy With Tumor-Treating Fields Plus Temozolomide vs Temozolomide Alone for Glioblastoma: A Randomized Clinical Trial. JAMA 2015, 314, 2535–2543. [Google Scholar] [CrossRef] [PubMed]
  6. Balwant, M.K. A Review on Convolutional Neural Networks for Brain Tumor Segmentation: Methods, Datasets, Libraries, and Future Directions. IRBM 2022, 43, 521–537. [Google Scholar] [CrossRef]
  7. Ghosh, A.; Thakur, S. Review of Brain Tumor MRI Image Segmentation Methods for BraTS Challenge Dataset. In Proceedings of the 2022 12th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Virtual, 27–28 January 2022; pp. 405–410. [Google Scholar]
  8. Tillmanns, N.; Lum, A.E.; Cassinelli, G.; Merkaj, S.; Verma, T.; Zeevi, T.; Staib, L.; Subramanian, H.; Bahar, R.C.; Brim, W.; et al. Identifying clinically applicable machine learning algorithms for glioma segmentation: Recent advances and discoveries. Neuro-Oncol. Adv. 2022, 4, vdac093. [Google Scholar] [CrossRef]
  9. Wang, F.; Jiang, R.; Zheng, L.; Meng, C.; Biswal, B. 3D U-Net Based Brain Tumor Segmentation and Survival Days Prediction. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries, Cham, Switzerland, 4 October 2020; pp. 131–141. [Google Scholar]
  10. Hatamizadeh, A.; Nath, V.; Tang, Y.; Yang, D.; Roth, H.R.; Xu, D. Swin UNETR: Swin Transformers for Semantic Segmentation of Brain Tumors in MRI Images. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries, Cham, Switzerland, 4 October 2022; pp. 272–284. [Google Scholar]
  11. Shusharina, N.; Söderberg, J.; Edmunds, D.; Löfman, F.; Shih, H.; Bortfeld, T. Automated delineation of the clinical target volume using anatomically constrained 3D expansion of the gross tumor volume. Radiother. Oncol. 2020, 146, 37–43. [Google Scholar] [CrossRef]
  12. Sadeghi, S.; Farzin, M.; Gholami, S. Fully automated clinical target volume segmentation for glioblastoma radiotherapy using a deep convolutional neural network. Pol. J. Radiol. 2023, 88, 31–40. [Google Scholar] [CrossRef]
  13. Ermiş, E.; Jungo, A.; Poel, R.; Blatti-Moreno, M.; Meier, R.; Knecht, U.; Aebersold, D.M.; Fix, M.K.; Manser, P.; Reyes, M.; et al. Fully automated brain resection cavity delineation for radiation target volume definition in glioblastoma patients using deep learning. Radiat. Oncol. 2020, 15, 100. [Google Scholar] [CrossRef]
  14. Ramesh, K.; Gurbani, S.S.; Mellon, E.A.; Huang, V.; Goryawala, M.; Barker, P.B.; Kleinberg, L.; Shu, H.G.; Shim, H.; Weinberg, B.D. The Longitudinal Imaging Tracker (BrICS-LIT):A Cloud Platform for Monitoring Treatment Response in Glioblastoma Patients. Tomography 2020, 6, 93–100. [Google Scholar] [CrossRef]
  15. Chukwueke, U.N.; Wen, P.Y. Use of the Response Assessment in Neuro-Oncology (RANO) criteria in clinical trials and clinical practice. CNS Oncol. 2019, 8, CNS28. [Google Scholar] [CrossRef] [Green Version]
  16. Kickingereder, P.; Isensee, F.; Tursunova, I.; Petersen, J.; Neuberger, U.; Bonekamp, D.; Brugnara, G.; Schell, M.; Kessler, T.; Foltyn, M.; et al. Automated quantitative tumour response assessment of MRI in neuro-oncology with artificial neural networks: A multicentre, retrospective study. Lancet Oncol. 2019, 20, 728–740. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Weinberg, B.; Ramesh, K.; Gurbani, S.; Schreibmann, E.; Kleinberg, L.; Shu, H.-K.; Shim, H. NIMG-23. Brain tumor reporting and data system (BT-rads) and quantitative tools to guide its implementation. Neuro Oncol. 2019, 21, vi166. [Google Scholar] [CrossRef]
  18. Gore, A.; Hoch, M.J.; Shu, H.G.; Olson, J.J.; Voloschin, A.D.; Weinberg, B.D. Institutional Implementation of a Structured Reporting System: Our Experience with the Brain Tumor Reporting and Data System. Acad. Radiol. 2019, 26, 974–980. [Google Scholar] [CrossRef]
  19. Kim, S.; Hoch, M.J.; Cooper, M.E.; Gore, A.; Weinberg, B.D. Using a Website to Teach a Structured Reporting System, the Brain Tumor Reporting and Data System. Curr. Probl. Diagn. Radiol. 2021, 50, 356–361. [Google Scholar] [CrossRef]
  20. Lee, S.J.; Weinberg, B.D.; Gore, A.; Banerjee, I. A Scalable Natural Language Processing for Inferring BT-RADS Categorization from Unstructured Brain Magnetic Resonance Reports. J. Digit. Imaging 2020, 33, 1393–1400. [Google Scholar] [CrossRef] [PubMed]
  21. Zhang, J.Y.; Weinberg, B.D.; Hu, R.; Saindane, A.; Mullins, M.; Allen, J.; Hoch, M.J. Quantitative Improvement in Brain Tumor MRI Through Structured Reporting (BT-RADS). Acad. Radiol. 2020, 27, 780–784. [Google Scholar] [CrossRef] [PubMed]
  22. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-Unet: Unet-Like Pure Transformer for Medical Image Segmentation. In Proceedings of the Computer Vision—ECCV 2022 Workshops, Cham, Switzerland, 23–27 October 2022; pp. 205–218. [Google Scholar]
  23. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  24. Yang, C.; Guo, X.; Wang, T.; Yang, Y.; Ji, N.; Li, D.; Lv, H.; Ma, T. Automatic Brain Tumor Segmentation Method Based on Modified Convolutional Neural Network. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Ann ual International Conference, Berlin, Germany, 23–27 July 2019; pp. 998–1001. [Google Scholar] [CrossRef]
  25. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  26. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016: 19th International Conference, Athens, Greece, 17–21 October 2016; pp. 424–432. [Google Scholar]
  27. Fushiki, T. Estimation of prediction error by using K-fold cross-validation. Stat. Comput. 2011, 21, 137–146. [Google Scholar] [CrossRef]
  28. Dice, L.R. Measures of the Amount of Ecologic Association Between Species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  29. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef] [Green Version]
  30. Barragán-Montero, A.; Javaid, U.; Valdés, G.; Nguyen, D.; Desbordes, P.; Macq, B.; Willems, S.; Vandewinckele, L.; Holmström, M.; Löfman, F.; et al. Artificial intelligence and machine learning for medical imaging: A technology review. Phys. Medica 2021, 83, 242–256. [Google Scholar] [CrossRef] [PubMed]
  31. Giansanti, D.; Di Basilio, F. The Artificial Intelligence in Digital Radiology: Part 1: The Challenges, Acceptance and Consensus. Healthcare 2022, 10, 509. [Google Scholar] [CrossRef]
  32. Saw, S.N.; Ng, K.H. Current challenges of implementing artificial intelligence in medical imaging. Phys. Med. 2022, 100, 12–17. [Google Scholar] [CrossRef] [PubMed]
  33. Martín-Noguerol, T.; Paulano-Godino, F.; López-Ortega, R.; Górriz, J.M.; Riascos, R.F.; Luna, A. Artificial intelligence in radiology: Relevance of collaborative work between radiologists and engineers for building a multidisciplinary team. Clin. Radiol. 2021, 76, 317–324. [Google Scholar] [CrossRef]
  34. Choi, Y.; Choi, M.; Kim, M.; Ha, J.-W.; Kim, S.; Choo, J. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8789–8797. [Google Scholar]
  35. Xin, B.; Hu, Y.; Zheng, Y.; Liao, H. Multi-modality generative adversarial networks with tumor consistency loss for brain mr image synthesis. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 4–7 April 2020; pp. 1803–1807. [Google Scholar]
  36. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
Figure 1. FLAIR and CE-T1w segmentation overlays (in red) from a BraTS competition-winning segmentation model over MRIs prior to surgery, prior to receiving radiation therapy, and 3 months after completing radiation therapy.
Figure 1. FLAIR and CE-T1w segmentation overlays (in red) from a BraTS competition-winning segmentation model over MRIs prior to surgery, prior to receiving radiation therapy, and 3 months after completing radiation therapy.
Cancers 15 03956 g001
Figure 2. Example contour overlays from our training dataset of GTV1 (yellow contour) overlaid on T2w/FLAIR MRI and GTV2 (red contour) on CE-T1w MRI that includes cavity and blood product.
Figure 2. Example contour overlays from our training dataset of GTV1 (yellow contour) overlaid on T2w/FLAIR MRI and GTV2 (red contour) on CE-T1w MRI that includes cavity and blood product.
Cancers 15 03956 g002
Figure 3. (A) Training pipeline for our segmentation models. During the pre-processing step, a 3-channel input was used with the first two channels using skull-stripped T2w/FLAIR and CE-T1w MRIs and the third channel including skull to assist in delineating cavity margins. The model weights that performed the best with the validation data were tested on the independent test dataset. (B) Network architecture for 3D Unet, which has a depth of 3, and was trained on our post-operative MRIs after pre-processing. The number of filters is listed above each block.
Figure 3. (A) Training pipeline for our segmentation models. During the pre-processing step, a 3-channel input was used with the first two channels using skull-stripped T2w/FLAIR and CE-T1w MRIs and the third channel including skull to assist in delineating cavity margins. The model weights that performed the best with the validation data were tested on the independent test dataset. (B) Network architecture for 3D Unet, which has a depth of 3, and was trained on our post-operative MRIs after pre-processing. The number of filters is listed above each block.
Cancers 15 03956 g003
Figure 4. Example GTV2 (CE-T1w target) and GTV1 (T2w/FLAIR target) segmentations for patients from three different institutions with our best-performing 3D Unet model. For GTV2 and GTV1, we have included an axial view of the clinician GTVs (red) and the prediction (blue) as well as a volume-rendered view of the contour overlap in the sagittal orientation. Dice scores are included between the 2D and volume rendering for each example.
Figure 4. Example GTV2 (CE-T1w target) and GTV1 (T2w/FLAIR target) segmentations for patients from three different institutions with our best-performing 3D Unet model. For GTV2 and GTV1, we have included an axial view of the clinician GTVs (red) and the prediction (blue) as well as a volume-rendered view of the contour overlap in the sagittal orientation. Dice scores are included between the 2D and volume rendering for each example.
Cancers 15 03956 g004
Figure 5. The three worst-performing cases of our 3D Unet model. In (A), the small CE-T1w lesion is missed, possibly being confused for blood vessel. In (B), the model has difficulty capturing the entire extent of hyperintensity from the 3D T2w/FLAIR acquisition, and in (C), the model confuses the ventricle for the shape of a cavity, especially when there are T2w/FLAIR edema in the same region.
Figure 5. The three worst-performing cases of our 3D Unet model. In (A), the small CE-T1w lesion is missed, possibly being confused for blood vessel. In (B), the model has difficulty capturing the entire extent of hyperintensity from the 3D T2w/FLAIR acquisition, and in (C), the model confuses the ventricle for the shape of a cavity, especially when there are T2w/FLAIR edema in the same region.
Cancers 15 03956 g005
Figure 6. The 3D Unet segmentation model implemented into our web application BrICS-LIT, displayed for an example patient’s post-operative and post-RT follow-up MRIs. Automated lesion segmentations (red overlay) and lesion volume cutoffs assist in generating automated structured reporting scores (bottom row) for post-treatment longitudinal tracking.
Figure 6. The 3D Unet segmentation model implemented into our web application BrICS-LIT, displayed for an example patient’s post-operative and post-RT follow-up MRIs. Automated lesion segmentations (red overlay) and lesion volume cutoffs assist in generating automated structured reporting scores (bottom row) for post-treatment longitudinal tracking.
Cancers 15 03956 g006
Figure 7. Interactive chart tracking changes in CE-T1w lesion volume (ENH), T2w/FLAIR lesion volume, and structured reporting scores for the patient shown in Figure 6. A drastic increase in lesion volume is seen 6 months after surgery.
Figure 7. Interactive chart tracking changes in CE-T1w lesion volume (ENH), T2w/FLAIR lesion volume, and structured reporting scores for the patient shown in Figure 6. A drastic increase in lesion volume is seen 6 months after surgery.
Cancers 15 03956 g007
Table 1. A comparison of strengths and weaknesses between previous efforts in brain lesion segmentation and our proposed effort.
Table 1. A comparison of strengths and weaknesses between previous efforts in brain lesion segmentation and our proposed effort.
Brain Lesion SegmentationStrengthsWeaknesses
Previous
Efforts
  • Utilizes pre-contrast T1w and T2w MRIs.
  • High segmentation performance for pre-surgical lesions.
  • BraTS includes a larger training dataset from multiple institutions.
  • Requires more imaging for each segmentation.
  • Cannot segment post-operative cavity.
Proposed
Effort
  • First effort to segment post-operative lesions (including cavity) for RT planning.
  • Model can also be used for post-RT longitudinal tracking.
  • Longitudinal lesion volumes can be used to generate automated disease-state classifications.
  • Ground truth contours generated and reviewed by radiation oncologists and a neuro-radiologist.
  • By only using CE-T1w and T2/FLAIR, models utilize less information for prediction, leading to potentially lower segmentation performance.
  • Training dataset is smaller than BraTS, but larger than other published efforts.
Table 2. A comparison of 2D and 3D approaches to post-surgical GTV segmentation. The CNN-based 3D Unet outperforms the other models on our test dataset with the highest Dice scores (bolded).
Table 2. A comparison of 2D and 3D approaches to post-surgical GTV segmentation. The CNN-based 3D Unet outperforms the other models on our test dataset with the highest Dice scores (bolded).
ModelGTV1GTV2
Train
(Dice)
Test
(Dice)
Test
(Hausdorff)
Test
(Jaccard)
Train
(Dice)
Test
(Dice)
Test
(Hausdorff)
Test
(Jaccard)
2D Unet0.930.4378.500.190.920.5675.410.34
2D Resunet0.930.5858.500.360.910.5735.570.35
2D Swin-Unet0.890.6460.710.440.860.5135.630.31
3D Unet0.770.7212.770.510.790.7310.750.58
3D Swin-UNETR0.640.6038.320.360.650.6423.330.44
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ramesh, K.K.; Xu, K.M.; Trivedi, A.G.; Huang, V.; Sharghi, V.K.; Kleinberg, L.R.; Mellon, E.A.; Shu, H.-K.G.; Shim, H.; Weinberg, B.D. A Fully Automated Post-Surgical Brain Tumor Segmentation Model for Radiation Treatment Planning and Longitudinal Tracking. Cancers 2023, 15, 3956. https://doi.org/10.3390/cancers15153956

AMA Style

Ramesh KK, Xu KM, Trivedi AG, Huang V, Sharghi VK, Kleinberg LR, Mellon EA, Shu H-KG, Shim H, Weinberg BD. A Fully Automated Post-Surgical Brain Tumor Segmentation Model for Radiation Treatment Planning and Longitudinal Tracking. Cancers. 2023; 15(15):3956. https://doi.org/10.3390/cancers15153956

Chicago/Turabian Style

Ramesh, Karthik K., Karen M. Xu, Anuradha G. Trivedi, Vicki Huang, Vahid Khalilzad Sharghi, Lawrence R. Kleinberg, Eric A. Mellon, Hui-Kuo G. Shu, Hyunsuk Shim, and Brent D. Weinberg. 2023. "A Fully Automated Post-Surgical Brain Tumor Segmentation Model for Radiation Treatment Planning and Longitudinal Tracking" Cancers 15, no. 15: 3956. https://doi.org/10.3390/cancers15153956

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop