Next Article in Journal
The Effect of Genomic DNA Contamination on the Detection of Circulating Long Non-Coding RNAs: The Paradigm of MALAT1
Next Article in Special Issue
Contrast Administration Impacts CT-Based Radiomics of Colorectal Liver Metastases and Non-Tumoral Liver Parenchyma Revealing the “Radiological” Tumour Microenvironment
Previous Article in Journal
Post-Mortem Detection of SARS-CoV-2 RNA in Long-Buried Lung Samples
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Does Anatomical Contextual Information Improve 3D U-Net-Based Brain Tumor Segmentation?

by
Iulian Emil Tampu
1,2,*,
Neda Haj-Hosseini
1,2 and
Anders Eklund
1,2,3
1
Department of Biomedical Engineering, Linköping University, 581 83 Linköping, Sweden
2
Center for Medical Image Science and Visualization, Linköping University, 581 83 Linköping, Sweden
3
Department of Computer and Information Science, Linköping University, 581 83 Linköping, Sweden
*
Author to whom correspondence should be addressed.
Diagnostics 2021, 11(7), 1159; https://doi.org/10.3390/diagnostics11071159
Submission received: 24 May 2021 / Revised: 22 June 2021 / Accepted: 23 June 2021 / Published: 25 June 2021
(This article belongs to the Special Issue AI as a Tool to Improve Hybrid Imaging in Cancer)

Abstract

:
Effective, robust, and automatic tools for brain tumor segmentation are needed for the extraction of information useful in treatment planning. Recently, convolutional neural networks have shown remarkable performance in the identification of tumor regions in magnetic resonance (MR) images. Context-aware artificial intelligence is an emerging concept for the development of deep learning applications for computer-aided medical image analysis. A large portion of the current research is devoted to the development of new network architectures to improve segmentation accuracy by using context-aware mechanisms. In this work, it is investigated whether or not the addition of contextual information from the brain anatomy in the form of white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF) masks and probability maps improves U-Net-based brain tumor segmentation. The BraTS2020 dataset was used to train and test two standard 3D U-Net (nnU-Net) models that, in addition to the conventional MR image modalities, used the anatomical contextual information as extra channels in the form of binary masks (CIM) or probability maps (CIP). For comparison, a baseline model (BLM) that only used the conventional MR image modalities was also trained. The impact of adding contextual information was investigated in terms of overall segmentation accuracy, model training time, domain generalization, and compensation for fewer MR modalities available for each subject. Median (mean) Dice scores of 90.2 (81.9), 90.2 (81.9), and 90.0 (82.1) were obtained on the official BraTS2020 validation dataset (125 subjects) for BLM, CIM, and CIP, respectively. Results show that there is no statistically significant difference when comparing Dice scores between the baseline model and the contextual information models (p > 0.05), even when comparing performances for high and low grade tumors independently. In a few low grade cases where improvement was seen, the number of false positives was reduced. Moreover, no improvements were found when considering model training time or domain generalization. Only in the case of compensation for fewer MR modalities available for each subject did the addition of anatomical contextual information significantly improve (p < 0.05) the segmentation of the whole tumor. In conclusion, there is no overall significant improvement in segmentation performance when using anatomical contextual information in the form of either binary WM, GM, and CSF masks or probability maps as extra channels.

1. Introduction

Generally, patients diagnosed with brain tumor undergo radical treatment which can include a combination of surgical tumor resection, radiotherapy, and chemotherapy [1]. In case of surgery, a major factor that influences patient survival and postoperative morbidity is the extent of the resection [1,2]. Treatment planning depends extensively on diagnostic radiology images for the identification of the tumor, key information for balancing the extent of the treatment target with the collateral effects.
MR imaging modalities, such as T1-weighted (T1w), T1-weighted with post-contrast gadolinium enhancement (T1Gd), T2-weighted (T2w), and T2 fluid attenuated inversion recovery (FLAIR), are commonly used for the identification of the tumor [3]. Reliable tools for the extraction of relevant information from the MR images are needed. For this, manual annotation of brain tumors is commonly practiced in clinical routine [4]; however, this is a time consuming and labor-intensive task. Moreover, manual annotation is not objective, with poor agreement between specialists [5]. Automatic methods could overcome these limitations, providing a faster and objective identification of the tumor sub-regions.
Automatic segmentation of brain tumor structures in MR images is challenging and has attracted a great research interest. Among the proposed methods [6], convolutional neural networks (CNNs) have shown state-of-the-art performance, ranking first in the multimodal Brain Tumor Image Segmentation Benchmark (BraTS) challenge during recent years [7]. Given the automatic feature extraction of CNNs [8], the majority of the research is focused on designing network architectures which provide better accuracy for the segmentation task. One of the most popular CNN architectures is U-Net [9], which introduced skip connections between the layers in the network. A plethora of U-Net-like architectures have since then been developed, including, among others, laborious skip connections strategies [10] and attention mechanisms [11]. However, Isensee et al. [12], who obtained top performance using a well-trained U-Net, showed that improving segmentation performance is not only a matter of adjusting the network architecture. The choice of loss function, training strategy, augmentation, and post-processing showed to have a large impact on the segmentation performance.
An emerging topic in artificial intelligence applications, including computer-aided medical interventions, is context-awareness that will allow algorithms to use the information from the surrounding and perform segmentation of images considering the anatomy context [13] and, thus, potentially improve the outcome. The latest literature describes different approaches for context-aware deep learning including auto-context strategies, changing the model architecture, and/or providing additional contextual information during training. Examples of auto-context strategies used to improve model segmentation performance can be seen in References [14,15]. In particular, Reference [15] implemented auto-context in their 3D symmetric fully convolutional neural network by combining multi modal MR images with 3D Haar features with the purpose of improving brain tumor segmentation.
A number of attempts have been made to evaluate the impact of the introduction of context-aware blocks in the model architecture on brain tumor segmentation [16,17,18,19,20]. For example, Pei et al. [19] used a context-aware deep neural network which thanks to a context encoding module between the encoder and the decoder part of the network, helped in overcoming the class imbalance problem that challenges brain tumor segmentation. However, such an implementation does not allow a comparison between the model accuracies with and without the context encoding module; thus, the contribution of context information cannot be assessed.
Another approach for achieving context-awareness is to provide the network with more information [21,22,23,24]. Wachinger et al. [21] included brain spectral coordinates information while training a patch-based deep CNN for brain anatomy segmentation. The authors argued that patches lose context information when propagating through the network, with possible confusion due to the symmetry of the brain. During training, each patch was supplemented with location information obtained from a spectral-based parameterization of the brain volume. Interestingly, the additional information was provided intermediately in the network, concatenating the context information to the feature maps of the initial convolutional layers. In two studies, Shen et al. [22,23] instead added four channels to the input of their fully convolutional network in addition to the four conventional MR modalities. The additional information consisted of symmetry maps computed on all the MR modalities, describing the asymmetry between the brain hemispheres introduced by the tumor. Kao et al. [24] included brain parcellation information during the training of a 3D U-Net as means of location information. In their work, the authors registered the MNI152 1 mm brain parcellation atlas [25] to each BraTS subject, obtaining a mapping of each voxel into one of the defined 69 brain structures.
The aim of this study is to expand this line of thought and investigate if using anatomical contextual information as additional input channels can improve brain tumor segmentation performance considering four aspects: (1) segmentation accuracy when training on multimodal MR images, (2) model training time, (3) compensation for fewer MR modalities available for each subject, and (4) domain generalization. All four aspects are studied also with respect to low grade glioma (LGG) and high grade glioma (HGG) cases independently. Anatomical contextual information is defined in this study as white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF) masks or probability maps obtained automatically using an automatic segmentation tool.

2. Materials and Methods

2.1. Dataset

The BraTS2020 dataset was used in this work [7,26,27], which contains 369 preoperative multimodal (T1w, T1Gd, T2w, and T2-FLAIR) 3D MR images, from 19 different centers, of both HGG (n = 293) and LGG (n = 79). Manual annotations of three tumor sub-regions for each case are provided with the dataset identifying the necrotic (NCR) and the non-enhancing tumor core (NET), the enhancing tumor (ET) and the peritumoral edema tissue (ED). The combination of the above annotations, namely the tumor core (TC = NCR ∪ NET ∪ ET), the ET, and the whole tumor (WT = TC ∪ ED), are targets of the segmentation task. A complete description of the BraTS2020 dataset is available in Reference [7]. Only the BraTS2020 dataset was used in this study, since it is a standardized and widely used dataset that allows comparison between methods and literature.

2.2. Anatomical Contextual Information

Contextual information in the form of binary WM, GM, and CSF masks and probability maps was obtained using FMRIB’s automated segmentation tool (FAST) [28] applied on the T1w MR volumes, each with normalized and zero-centered intensity. The difference between the FAST masks obtained from the raw T1w and the intensity normalized and zero-centered T1w volumes was minor. Of the total 369 subjects, 92% showed less than 10% difference in voxel classification (WM, GM, or CSF). The intensity normalized and zero-centered volumes were used instead of the raw data, since a preliminary investigation of the proposed method indicated that segmentation quality was lower when using contextual information from raw T1w data compared to when it was obtained from the intensity normalized and zero-centered volumes. As Tudorascu et al. described [29], methods that use spatial priors during the brain anatomy segmentation, such as the methods in the Statistical Parametric Mapping (SPM) [30] or FreeSurfer [31] softwares, may perform poorly on diseased brains that contain deformations. Brain tumors can induce substantial deformation of the brain structures, making the intensity-based FAST tool more suitable for obtaining the contextual information used in this study, even if it is not specificately designed for patients with tumors. An initial qualitative investigation on obtaining the anatomical segmentation through SPM showed that WM, GM, and CSF masks and probability maps lacked detail and were distorted. Given the qualitatively assessed higher quality of the soft tissue masks, and that no spatial priors are used during the anatomical segmentation, FAST was used in this study.

2.3. Model

Figure 1 shows an overview of the methodology, where the nnU-Net deep learning framework [32] was used. There are two reasons for this choice: (1) nnU-Net’s repeated success in the BraTS challenges in the recent years shows the reliability of this framework, which could be difficult to achieve with an in-house model, and (2) this allows reproducibility of the presented investigation. nnU-Net is built upon the 3D U-Net architecture and automatically the tunes network hyperparameters based on the training dataset and hardware available. Among others, the framework tunes the number of convolutional layers, input patch size and batch size [32]. In this study, the 3D full resolution U-Net configuration was adopted, and four NVIDIA Tesla V100 GPUs (32 GB memory) were used for the training. During training, the sum of Dice and cross-entropy loss was minimized using stochastic gradient descent with Nesterov momentum ( μ = 0.99 ). The number of training epochs was automatically set to 1000 by nnU-Net, without any early stopping strategies. Each epoch consisted of 250 mini-batches.

2.4. Evaluation and Statistical Methods

To investigate if the addition of contextual information has an impact on glioma segmentation performance, three models were trained that differed in the use or not of the contextual information: a baseline model (BLM) with input channels chosen among the four conventional MR modalities provided by BraTS, and two contextual information models both with three additional channels compared to BLM to accommodate the extra information obtained from FAST. One contextual information model used binary WM, GM, and CSF masks (CIM), while the other model used the WM, GM, and CSF probability maps (CIP). A 3-fold cross validation scheme was used to train each setting described below. After training, the segmentation of the test subjects was obtained as an ensemble of the predictions of the three models trained through cross validation. Dice score [26] and 95% Hausdorff distance (HD) [26] on the segmentation targets were obtained through the automatic validation system, on both the official BraTS2020 validation dataset (125 cases) and an independent test dataset (36 cases), described in Section 2.5. The non-parametric Wilcoxon signed-rank test was used to test the null hypothesis of no difference between the baseline model and the contextual information models, at a significance level of 5%. Statistical analysis was performed in IBM® SPSS® (Version 27.0, Armonk, NY, USA, IBM Corp).

2.5. Multimodal MR Model Training

To study the effect of anatomical contextual information on segmentation performance, all four conventional MR modalities were used as input to the three models, with CIM and CIP additionally using the anatomical contextual information, as described above. In addition to the official BraTS2020 validation dataset, 36 subjects containing an equal number of HGGs and LGGs were randomly selected from the training dataset as the independent test dataset, with the remaining 333 subjects used for training. The choice of an independent test dataset with control over the tumor grades allows us to investigate the effect that anatomical contextual information has on the segmentation of LGGs and HGGs independently. Moreover, to understand the impact of contextual information on the model training time, the validation loss curves saved by nnU-Net were analyzed a posteriori for these models. Training was considered finished when the validation loss did not improve over 50 epochs.

2.6. Compensation for Fewer MR Modalities

To explore if anatomical contextual information could compensate for the missing information when only one MR modality is used as input, the three models were trained and tested, similarly as in Section 2.5, with only T1Gd provided instead of four MR images per subject. T1Gd was selected, among the other MR modalities, given that (1) it provides contrast of the tumor core region compared to the surrounding healthy tissue [7] and (2) because T1w is already used by FAST to obtain the anatomical contextual information.

2.7. Domain Generalization

Finally, to investigate if the addition of contextual information improves domain generalization, the three models were trained on BraTS cases from a single institute and tested on data from a variety of institutes. In particular, a total of 69 (35 LGGs and 34 HGGs), among the 369 cases, were identified to originate from one of the 19 institutes that contributed to the BraTS2020 dataset. Identification of the institutes was possible using the information from References [7,27] and the BraTS name mapping information. Models were trained using all conventional MR modalities, and the 69 cases were excluded from the independent test dataset.

3. Results

An example of contextual information obtained using FAST can be seen in Figure 2, where the cross-section of CSF, GM, and WM masks are shown for two subjects. By visually inspecting the FAST results, the soft tissue segmentations are descriptive of the brain WM, GM, and CSF structures, with the masks and probability maps being distorted only in the regions where the tumor is located or proximal to it.

3.1. Segmentation Accuracy for Multimodal MR Model Training

An example of segmentation results in an axial slice for one of the independent test samples for BLM, CIM, and CIP models when using all the available MR modalities is shown in Figure 3. Performance of the three models on the official validation dataset (125 subjects) is shown as boxplots in Figure 4, where Dice scores and 95% HD are reported for all segmentation targets. Dice scores’ median values (mean) across target regions were 90.15 (81.85), 90.17 (81.87), and 90.04 (82.06), for BLM, CIM, and CIP, respectively. Table 1 summarizes the median Dice scores and and 95% HD obtained on the independent test dataset for LGG and HGG cases separately, showing that HGGs are overall better segmented than LGGs. When comparing CIM and CIP to BLM across the different tumor regions, no statistically significant difference (p > 0.05) in Dice scores was found when analyzing the results from both the independent test dataset and the official BraTS2020 validation set. Moreover, no significant difference (p > 0.05) was observed when comparing the effect of contextual information in segmenting LGG and HGG cases separately, with HGG showing slightly lower p-values. When looking at the cases that showed at least 5% improvement in mean Dice score when using anatomical contextual information, it could be seen that the enhancing tumor region was better segmented. Among the subjects in the independent test dataset, all of those with improved mean Dice score (5.6% of the total subjects) were LGGs, with the contextual information models avoiding false positives for the enhancing tumor region. Table 2 summarizes mean Dice scores across tumor regions for studies that implemented context-awareness by means of auto-context, architectural changes, or additional contextual information. The mean Dice scores obtained in this study are in the same range as results previously reported in literature. Note that the disparity in segmentation performance of the result presented here compared to Reference [32] is due to the fact that, in this work, only the 3D full resolution U-Net configuration was used, instead of a combination of 2D U-Net, 3D low resolution, and 3D full resolution U-Net, that can be trained and ensembled using the nnU-Net framework, at the cost of a longer training time.

3.2. Model Training Time for Multimodal MR Model Training

From the a posteriori analysis of the validation loss curves, the baseline model trained 12 and 5 h (46 and 9 epochs) faster than CIM and CIP, respectively, when looking at the average values across the three folds. Average training times and epochs for the three models are summarized in Table 3.

3.3. Compensation for Fewer MR Modalities

Segmentation performance results on the BraTS2020 validation set (125 subjects) for BLM, CIM, and CIP when trained using only T1Gd as MR modality are summarized in Figure 5. Dice score values for TC and ET regions are similar when compared to the models trained on all the four MR modalities. The whole tumor region segmentation, on the other hand, shows a decrease in performance that can be attributed to the lack of the contrast in the T1Gd between edema region and surrounding tissue, that is present in FLAIR.
When comparing the models trained only on T1Gd on the BraTS validation dataset, the Dice score for the whole tumor region is significantly improved (p < 0.05) for both contextual information models compared to the baseline model (also after Bonferroni correction for multiple comparisons). Considering the results for LGG and HGG cases separately computed on the independent test dataset (36 subjects), no significant difference could be found between the models, not even with respect to the whole tumor region. Median Dice and 95% HD for the independent test set are summarized in Table 4.

3.4. Domain Generalization

Dice scores and 95% HD values on the official BraTS validation dataset obtained for the models trained on data from a single institute are summarized in Figure 6. Compared to the baseline model trained on data from all the institutes, performance is lower especially for the TC and ET tumor regions. The drop in performance of all the models trained only on single-center data shows the impact of domain shift between the training and test datasets. The addition of anatomical contextual information does not help in this aspect, since the three models trained on single-center data have similar performances, which are all significantly lower (p < 0.05) than the one of the model trained on all the MR data available. Results on the independent test dataset, summarized in Table 5, show a similar trend for both LGG and HGG cases.

4. Discussion

The effect of anatomical contextual information on brain tumor segmentation was investigated with respect to segmentation performance, model training, model generalization, and compensation for fewer MR modalities.

4.1. Segmentation Accuracy and Training Time for Multimodal MR Model Training

Glioma segmentation performance in this study showed no significant improvement when comparing models trained with the addition of anatomical contextual information as input channels along with the conventional MR modalities. A possible reason for the observed results may be found in how the WM, GM, and CSF information are computed. FAST uses pixel intensity and spatial information for the segmentation. Arguably, this is very similar to what a U-Net architecture is using when trained for semantic segmentation. Thus, it is possible that the network is independently creating a representation of WM, GM, and CSF at some stage during training from the conventional MR modalities, nullifying the additional information. However, providing such information already as input channels did not speed up model training, given that BLM trained faster compared to the contextual information models based on the a posteriori analysis of the validation loss curves. The addition of the contextual information does not improve segmentation performance but instead increases the model convergence time, suggesting that the extra information is not used and makes the segmentation problem harder to solve.
Direct comparison between the obtained results with other works is partially possible, given the differences in testing datasets. The results presented here are in the same range of reported findings of studies that used additional contextual information during model training. Overall, the reported results in literature and the ones obtained in this study show that the inclusion of context-awareness, by means of model architecture changes or additional information as input to the network, has marginal or no improvement on glioma segmentation [15,16,17,18,19,20,22,23,24]. This should not discourage future research on the topic, but instead promote studies that exploit contextual information for brain tumor segmentation by other approaches and perhaps a combination of the currently implemented methods, i.e., context-aware blocks and additional contextual information as input to the network.

4.2. Quality of the Anatomical Contextual Information

Another reason for why the model does not use the additional anatomical information may be found in the quality of the WM, GM, and CSF binary masks and probability maps. As shown in Figure 2, the WM, GM, and CSF masks are distorted in the brain region containing the tumor, which may not help the network. One possible way of investigating this aspect is to compare model performance with anatomical contextual information obtained automatically or from manual annotations. However, the amount of time that would be needed for the annotation of WM, GM, and CSF of each subject is exceedingly large making this comparison unfeasible. Another possible approach is to obtain the anatomical information from quantitative MRI (qMRI) [34]. By quantitatively measuring the relaxation times of tissues in the brain, qMRI can provide probability maps for WM, GM, and CSF. In contrast to the automatically generated probability maps used in this study, the ones obtained through qMRI are not derived quantities; thus, the information given in each of the input channels is unique and is not a different version of the same information. This does increase the amount of information that the model can actually use for the segmentation task. Given that the BraTS2020 dataset does not provide qMRI data, this approach remains open for investigation.

4.3. Compensation for Fewer MR Modalities

Reducing the MR modalities needed to be acquired could have a positive impact on patient hospitalization experience and on healthcare economy, since a shorter time would be needed for the patient to be in the MR scanner and more patients could be scanned [35]. For this reason, here, it was investigated whether or not the addition of anatomical information could compensate for the decrease in segmentation performance caused by using only one MR modality (T1Gd) as input to the model. Results show that only the segmentation of the whole tumor region is affected by the lack of the excluded MR modalities. This is not surprising since the WT includes the ED region, which is not visible in T1Gd. However, the addition of contextual information marginally improves WT segmentation, suggesting that the WM, GM, and CSF masks help the model to better identify the edema region.

4.4. Domain Generalization

Domain shift is a challenge that today’s deep learning models in general have to address when intended for real world applications [36]. In the context of medical image segmentation, models trained on data from a single center or scanner often struggle to retain segmentation performance when tested or used on data that originates from different centers or scanners. As described by Zhou et al. [36], domain generalization is still a challenge that hinders the expansion of deep learning methods for real world applications. Anatomical contextual information has shown no impact on domain generalization, given that the models trained on single center data and with WM, GM, and CSF as extra input channels have suffered from a similar drop in the performance compared to the baseline performance as the model trained without it. A possible reason for this, as discussed above, is that the model is not using the additional anatomical information.

4.5. Effect of Contextual Information on LGG and HGG Cases

Overall, anatomical contextual information does not improve segmentation performance when considering high and low grade cases separately. Even if no statistically significant improvement could be found between the baseline and the contextual information models, in some of the LGG cases, the addition of contextual information reduced the number of false positives for the enhancing tumor region. The low number of LGGs in the training dataset, compared to HGGs, could bias the model in always segmenting an enhancing tumor region. However, in LGG cases, this tumor sub-region is missing, which leads to a high number of false positives when the model segments it. When using anatomical contextual information, the model could better discriminate HGG and LGG cases, thus avoiding segmenting the enhancing tumor region that is not present in LGGs. Yet, the statistical comparison does not support this hypothesis, since no statistically significant difference was found between the baseline model and the contextual information models for the cases in this study. Nevertheless, a higher number of LGG test samples are needed to study this effect.

4.6. Future Perspectives

A possible approach to investigate is providing contextual information not as additional channels to the input, but after the initial convolutional layers. Wachinger et al. [21] showed improved neuroanatomy segmentation performance when concatenating context information intermediately in the convolutional network. Moreover, the authors also showed that different types of context information affect the performance differently; a combination of spectral and cartesian-based parametrization of the brain volume yielded a better performance than when only one of the two was used, suggesting that they might contain complementary information. Thus, investigation could focus on finding different types of contextual information and their combinations. Future research could also address if using additional MR modalities (e.g., diffusion MR imaging) would improve brain tumor segmentation. The major factor hindering such investigation at the current moment is the lack of a standardized open access dataset which includes extra MR modalities.

5. Conclusions

Anatomical contextual information in the form of binary WM, GM, and CSF and probability maps was obtained in this study using the automatic FAST segmentation tool. The addition of anatomical contextual information as extra channels to the network shows no statistically significant difference in tumor segmentation performance when using a standardized 3D U-Net architecture and conventional multimodal MR images as dataset. Only in the case of using one conventional MR modality (T1Gd) did the addition of the anatomical contextual information show to significantly improve whole tumor region segmentation. No statistically significant improvements could be seen when investigating HGG and LGG cases separately, nor when considering model training time, domain generalization, and compensation for fewer MR modalities. Overall, context-aware approaches implemented for brain tumor segmentation in the recent literature show only minor or no improvements. This suggests that effective integration of context awareness in deep learning models for glioma segmentation has yet to be explored.

Author Contributions

Conceptualization, A.E.; methodology, I.E.T., N.H.-H. and A.E.; validation, I.E.T.; formal analysis, I.E.T.; resources, N.H.-H. and A.E.; writing—original draft preparation, I.E.T.; writing—review and editing, I.E.T., N.H.-H. and A.E.; supervision, N.H.-H. and A.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Swedish Research Council (2018-05250), IMPACT (2018-02230), AIDA (2017-02447), Åke Wiberg (M20-0031), Linköping University Cancer organization grant 2020 and 2021.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://www.med.upenn.edu/cbica/brats2020/registration.html.

Conflicts of Interest

Anders Eklund has previously received hardware from NVIDIA; otherwise, there are no conflicts of interest regarding this paper.

References

  1. Davis, M.E. Glioblastoma: Overview of disease and treatment. Clin. J. Oncol. Nurs. 2016, 20, S2. [Google Scholar] [CrossRef] [Green Version]
  2. D’Amico, R.S.; Englander, Z.K.; Canoll, P.; Bruce, J.N. Extent of resection in glioma–a review of the cutting edge. World Neurosurg. 2017, 103, 538–549. [Google Scholar] [CrossRef] [PubMed]
  3. Juratli, T.A.; Tummala, S.S.; Riedl, A.; Daubner, D.; Hennig, S.; Penson, T.; Zolal, A.; Thiede, C.; Schackert, G.; Krex, D.; et al. Radiographic assessment of contrast enhancement and T2/FLAIR mismatch sign in lower grade gliomas: Correlation with molecular groups. J. Neurooncol. 2019, 141, 327–335. [Google Scholar] [CrossRef] [PubMed]
  4. Visser, M.; Müller, D.; van Duijn, R.; Smits, M.; Verburg, N.; Hendriks, E.; Nabuurs, R.; Bot, J.; Eijgelaar, R.; Witte, M.; et al. Inter-rater agreement in glioma segmentations on longitudinal MRI. NeuroImage Clin. 2019, 22, 101727. [Google Scholar] [CrossRef] [PubMed]
  5. Kubben, P.L.; Postma, A.A.; Kessels, A.G.; van Overbeeke, J.J.; van Santbrink, H. Intraobserver and interobserver agreement in volumetric assessment of glioblastoma multiforme resection. Neurosurgery 2010, 67, 1329–1334. [Google Scholar] [CrossRef]
  6. Tiwari, A.; Srivastava, S.; Pant, M. Brain tumor segmentation and classification from magnetic resonance images: Review of selected methods from 2014 to 2019. Pattern Recognit. Lett. 2020, 131, 244–260. [Google Scholar] [CrossRef]
  7. Bakas, S.; Reyes, M.; Jakab, A.; Bauer, S.; Rempfler, M.; Crimi, A.; Shinohara, R.T.; Berger, C.; Ha, S.M.; Rozycki, M.; et al. Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the BRATS challenge. arXiv 2018, arXiv:1811.02629. [Google Scholar]
  8. Işın, A.; Direkoğlu, C.; Şah, M. Review of MRI-based brain tumor image segmentation using deep learning methods. Procedia Comput. Sci. 2016, 102, 317–324. [Google Scholar] [CrossRef] [Green Version]
  9. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Munich, Germany, 2015; pp. 234–241. [Google Scholar]
  10. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin, Germany, 2018; pp. 3–11. [Google Scholar]
  11. Noori, M.; Bahri, A.; Mohammadi, K. Attention-Guided Version of 2D UNet for Automatic Brain Tumor Segmentation. In Proceedings of the 2019 9th International Conference on Computer and Knowledge Engineering (ICCKE), Mashhad, Iran, 24–25 October 2019; pp. 269–275. [Google Scholar]
  12. Isensee, F.; Kickingereder, P.; Wick, W.; Bendszus, M.; Maier-Hein, K.H. No new-net. In International MICCAI Brainlesion Workshop; Springer: Berlin, Germany, 2018; pp. 234–244. [Google Scholar]
  13. Vercauteren, T.; Unberath, M.; Padoy, N.; Navab, N. Cai4cai: The rise of contextual artificial intelligence in computer-assisted interventions. Proc. IEEE 2019, 108, 198–214. [Google Scholar] [CrossRef] [Green Version]
  14. Tu, Z.; Bai, X. Auto-context and its application to high-level vision tasks and 3d brain image segmentation. IEEE PAMI 2009, 32, 1744–1757. [Google Scholar]
  15. Liu, J.; Liu, H.; Tang, Z.; Gui, W.; Ma, T.; Gong, S.; Gao, Q.; Xie, Y.; Niyoyita, J.P. IOUC-3DSFCNN: Segmentation of brain tumors via IOU constraint 3D symmetric full convolution network with multimodal auto-context. Sci. Rep. 2020, 10, 1–15. [Google Scholar] [CrossRef] [PubMed]
  16. Ahmad, P.; Qamar, S.; Shen, L.; Saeed, A. Context Aware 3D UNet for Brain Tumor Segmentation. arXiv 2020, arXiv:2010.13082. [Google Scholar]
  17. Chandra, S.; Vakalopoulou, M.; Fidon, L.; Battistella, E.; Estienne, T.; Sun, R.; Robert, C.; Deutsch, E.; Paragios, N. Context aware 3D CNNs for brain tumor segmentation. In International MICCAI Brainlesion Workshop; Springer: Berlin, Germany, 2018; pp. 299–310. [Google Scholar]
  18. Liu, Z.; Tong, L.; Chen, L.; Zhou, F.; Jiang, Z.; Zhang, Q.; Wang, Y.; Shan, C.; Li, L.; Zhou, H. CANet: Context Aware Network for 3D Brain Tumor Segmentation. arXiv 2020, arXiv:2007.07788. [Google Scholar]
  19. Pei, L.; Vidyaratne, L.; Rahman, M.M.; Iftekharuddin, K.M. Context aware deep learning for brain tumor segmentation, subtype classification, and survival prediction using radiology images. Sci. Rep. 2020, 10, 1–11. [Google Scholar] [CrossRef] [PubMed]
  20. Le, N.; Yamazaki, K.; Truong, D.; Quach, K.G.; Savvides, M. A Multi-task Contextual Atrous Residual Network for Brain Tumor Detection & Segmentation. arXiv 2020, arXiv:2012.02073. [Google Scholar]
  21. Wachinger, C.; Reuter, M.; Klein, T. DeepNAT: Deep convolutional neural network for segmenting neuroanatomy. NeuroImage 2018, 170, 434–445. [Google Scholar] [CrossRef]
  22. Shen, H.; Wang, R.; Zhang, J.; McKenna, S.J. Boundary-aware fully convolutional network for brain tumor segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 10–14 September 2017; pp. 433–441. [Google Scholar]
  23. Shen, H.; Zhang, J.; Zheng, W. Efficient symmetry-driven fully convolutional network for multimodal brain tumor segmentation. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3864–3868. [Google Scholar]
  24. Kao, P.Y.; Ngo, T.; Zhang, A.; Chen, J.W.; Manjunath, B. Brain tumor segmentation and tractographic feature extraction from structural MR images for overall survival prediction. In Proceedings of the International MICCAI Brainlesion Workshop, Granada, Spain, 16 September 2018; pp. 128–141. [Google Scholar]
  25. Desikan, R.S.; Ségonne, F.; Fischl, B.; Quinn, B.T.; Dickerson, B.C.; Blacker, D.; Buckner, R.L.; Dale, A.M.; Maguire, R.P.; Hyman, B.T.; et al. An automated labeling system for subdividing the human cerebral cortex on MRI scans into gyral based regions of interest. Neuroimage 2006, 31, 968–980. [Google Scholar] [CrossRef] [PubMed]
  26. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 2014, 34, 1993–2024. [Google Scholar] [CrossRef]
  27. Bakas, S.; Akbari, H.; Sotiras, A.; Bilello, M.; Rozycki, M.; Kirby, J.S.; Freymann, J.B.; Farahani, K.; Davatzikos, C. Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Sci. Data 2017, 4, 170117. [Google Scholar] [CrossRef] [Green Version]
  28. Zhang, Y.; Brady, M.; Smith, S. Segmentation of brain MR images through a hidden Markov random field model and the expectation-maximization algorithm. IEEE Trans. Med. Imaging 2001, 20, 45–57. [Google Scholar] [CrossRef]
  29. Tudorascu, D.L.; Karim, H.T.; Maronge, J.M.; Alhilali, L.; Fakhran, S.; Aizenstein, H.J.; Muschelli, J.; Crainiceanu, C.M. Reproducibility and bias in healthy brain segmentation: Comparison of two popular neuroimaging platforms. Front. Neurosci. 2016, 10, 503. [Google Scholar] [CrossRef]
  30. Ashburner, J.; Barnes, G.; Chen, C.C.; Daunizeau, J.; Flandin, G.; Friston, K.; Kiebel, S.; Kilner, J.; Litvak, V.; Moran, R.; et al. SPM12 Manual; Wellcome Trust Centre for Neuroimaging: London, UK, 2014; Volume 2464. [Google Scholar]
  31. Bruce, F. FreeSurfer. NeuroImage 2012, 62, 774–781. [Google Scholar]
  32. Isensee, F.; Jaeger, P.F.; Kohl, S.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef] [PubMed]
  33. Isensee, F.; Jaeger, P.F.; Full, P.M.; Vollmuth, P.; Maier-Hein, K.H. nnU-Net for Brain Tumor Segmentation. arXiv 2020, arXiv:2011.00848. [Google Scholar]
  34. Tofts, P. Quantitative MRI of the Brain: Measuring Changes Caused by Disease; John Wiley & Sons: Hoboken, NJ, USA, 2005. [Google Scholar]
  35. Hollingsworth, K.G. Reducing acquisition time in clinical MRI by data undersampling and compressed sensing reconstruction. Phys. Med. Biol. 2015, 60, R297. [Google Scholar] [CrossRef] [PubMed]
  36. Zhou, K.; Liu, Z.; Qiao, Y.; Xiang, T.; Loy, C.C. Domain generalization: A survey. arXiv 2021, arXiv:2103.02503. [Google Scholar]
Figure 1. Schematic of the proposed method showing the segmentation for one subject when using all four conventional MR modalities (T1w = T1-weighted, T1Gd = T1-weighted with post-contrast gadolinium enhancement, T2w = T2-weighted, and FLAIR = T2 fluid attenuated inversion recovery) and the 3D U-Net-based deep learning model. By applying FMRIB’s automated segmentation tool (FAST) to the intensity normalized and zero-centered T1w volumes, the contextual information is obtained as white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF) masks. The red patch in the contextual masks shows where the tumor is located. In this region, the contextual masks are distorted. The final segmentation, obtained as an ensemble of three cross-validation (CV) folds and provides regions of tumor core (TC), enhancing tumor (ET), and edema (ED), shown here in green, red, and blue, respectively. BLM is the baseline model, and CIM and CIP are the contextual information models using binary masks and probability maps, respectively.
Figure 1. Schematic of the proposed method showing the segmentation for one subject when using all four conventional MR modalities (T1w = T1-weighted, T1Gd = T1-weighted with post-contrast gadolinium enhancement, T2w = T2-weighted, and FLAIR = T2 fluid attenuated inversion recovery) and the 3D U-Net-based deep learning model. By applying FMRIB’s automated segmentation tool (FAST) to the intensity normalized and zero-centered T1w volumes, the contextual information is obtained as white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF) masks. The red patch in the contextual masks shows where the tumor is located. In this region, the contextual masks are distorted. The final segmentation, obtained as an ensemble of three cross-validation (CV) folds and provides regions of tumor core (TC), enhancing tumor (ET), and edema (ED), shown here in green, red, and blue, respectively. BLM is the baseline model, and CIM and CIP are the contextual information models using binary masks and probability maps, respectively.
Diagnostics 11 01159 g001
Figure 2. Examples of contextual information obtained using FMRIB’s automated segmentation tool are shown for two cases in an axial slice with the tumor annotation overlay in red. From left to right, cerebrospinal fluid (CSF), gray matter (GM), and white matter (WM) masks. (a) A case where the soft tissue masks are highly distorted; (b) a case where masks are only distorted in the tumor region.
Figure 2. Examples of contextual information obtained using FMRIB’s automated segmentation tool are shown for two cases in an axial slice with the tumor annotation overlay in red. From left to right, cerebrospinal fluid (CSF), gray matter (GM), and white matter (WM) masks. (a) A case where the soft tissue masks are highly distorted; (b) a case where masks are only distorted in the tumor region.
Diagnostics 11 01159 g002
Figure 3. Segmentation results obtained for one of the independent test subjects and for the three models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) are shown as colored labels overlaid on a T1-weighted axial slice. Tumor core (TC), enhancing tumor (ET), and edema (ED) are shown in green, red, and blue, respectively. Dice scores for BLM, CIM, and CIP for the presented test case are 92.54, 92.53, and 92.33, respectively.
Figure 3. Segmentation results obtained for one of the independent test subjects and for the three models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) are shown as colored labels overlaid on a T1-weighted axial slice. Tumor core (TC), enhancing tumor (ET), and edema (ED) are shown in green, red, and blue, respectively. Dice scores for BLM, CIM, and CIP for the presented test case are 92.54, 92.53, and 92.33, respectively.
Diagnostics 11 01159 g003
Figure 4. Boxplot of Dice scores and 95% Hausdorff distance (HD) of different models trained on multimodal MR images computed for the official BraTS2020 validation dataset (125 subjects). Boxplots show median and range (with box showing 25% and 75% quantiles) for each model (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) and segmentation target (ET = enhancing tumor, TC = tumor core, and WT = whole tumor). The statistical analysis showed no significant difference (p > 0.05) when comparing CIM and CIP with BLM.
Figure 4. Boxplot of Dice scores and 95% Hausdorff distance (HD) of different models trained on multimodal MR images computed for the official BraTS2020 validation dataset (125 subjects). Boxplots show median and range (with box showing 25% and 75% quantiles) for each model (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) and segmentation target (ET = enhancing tumor, TC = tumor core, and WT = whole tumor). The statistical analysis showed no significant difference (p > 0.05) when comparing CIM and CIP with BLM.
Diagnostics 11 01159 g004
Figure 5. Boxplot of Dice scores and 95% Hausdorff distance (HD) of different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) trained using only T1-weighted with post-gadolinium enhancement (T1Gd) as single MR modality computed for the official BraTS2020 validation dataset (125 subjects). For comparison, the performance of BLM trained on data from all institutes is presented (blue box). Boxplots show median and range (with box showing 25% and 75% quantiles) for each model and segmentation target (ET = enhancing tumor, TC = tumor core, and WT = whole tumor). By comparing CIM and CIP with the BLM, Dice scores of the WT region show a statistically significant difference (p < 0.05).
Figure 5. Boxplot of Dice scores and 95% Hausdorff distance (HD) of different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) trained using only T1-weighted with post-gadolinium enhancement (T1Gd) as single MR modality computed for the official BraTS2020 validation dataset (125 subjects). For comparison, the performance of BLM trained on data from all institutes is presented (blue box). Boxplots show median and range (with box showing 25% and 75% quantiles) for each model and segmentation target (ET = enhancing tumor, TC = tumor core, and WT = whole tumor). By comparing CIM and CIP with the BLM, Dice scores of the WT region show a statistically significant difference (p < 0.05).
Diagnostics 11 01159 g005
Figure 6. Boxplot of Dice scores and 95% Hausdorff distance (HD) for the different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) trained on single center data computed for BraTS2020 validation dataset. For comparison, the performance of BLM trained on data from all institutes is presented (blue box). Boxplots show median and range (with box showing 25% and 75% quantiles) for each model and segmentation target (ET = enhancing tumor, TC = tumor core, and WT = whole tumor). The statistical difference was insignificant (p > 0.05) when comparing CIM and CIP with BLM.
Figure 6. Boxplot of Dice scores and 95% Hausdorff distance (HD) for the different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) trained on single center data computed for BraTS2020 validation dataset. For comparison, the performance of BLM trained on data from all institutes is presented (blue box). Boxplots show median and range (with box showing 25% and 75% quantiles) for each model and segmentation target (ET = enhancing tumor, TC = tumor core, and WT = whole tumor). The statistical difference was insignificant (p > 0.05) when comparing CIM and CIP with BLM.
Diagnostics 11 01159 g006
Table 1. Median Dice scores and 95% Hausdorff distance (HD) across target regions (ET = enhancing tumor, TC = tumor core, and WT = whole tumor) for the different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) trained on multimodal MR images. Results are shown for low grade glioma (LGG) and high grade glioma (HGG) cases independently computed on the independent dataset (36 subjects). There was no significant difference (p > 0.05) when comparing CIM and CIP to BLM for neither of LGG and HGG cases. The results in bold correspond to the best performing model in each tissue region.
Table 1. Median Dice scores and 95% Hausdorff distance (HD) across target regions (ET = enhancing tumor, TC = tumor core, and WT = whole tumor) for the different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) trained on multimodal MR images. Results are shown for low grade glioma (LGG) and high grade glioma (HGG) cases independently computed on the independent dataset (36 subjects). There was no significant difference (p > 0.05) when comparing CIM and CIP to BLM for neither of LGG and HGG cases. The results in bold correspond to the best performing model in each tissue region.
GradeModelMedian Dice ScoreMedian 95% HD
[Min, Max][Min, Max]
TCETWTTCETWT
LGG
(18 cases)
BLM79.9636.7293.287.2139.663.00
[0.00, 98.39][0.00, 100.00][0.00, 97.09][1.00, 373.13][0.00, 373.13][1.00, 373.13]
CIM81.9544.8793.516.9825.183.00
[0.00, 98.38][0.00, 100.00][0.00, 96.98][1.00, 373.13][0.00, 373.13][1.41, 373.13]
CIP82.2751.2793.227.6423.173.08
[0.00, 80.00][0.00, 100.00][10.08, 97.18][1.00, 373.13][0.00, 373.13][1.00, 18.49]
HGG
(18 cases)
BLM94.8389.2292.791.411.412.00
[65.76, 97.56][75.54, 96.57][86.64, 97.19][1.00, 21.47][1.00, 3.00][1.00, 7.87]
CIM94.8089.1692.941.571.411.87
[64.22, 97.37][74.74, 96.17][86.90, 97.17][1.00, 22.03][1.00, 3.00][1.00, 7.87]
CIP94.5489.1292.741.571.411.87
[67.68, 97.48][74.21, 96.55][86.96, 97.27][1.00, 14.87][1.00, 3.00][1.00, 7.55]
Table 2. Mean Dice scores reported by similar studies using context-aware methods. * identifies the method ranking first in BraTS2020 (not using context-awareness mechanisms). 1 obtained via official BraTS test dataset, 2 obtained via official BraTS validation datasets, 3 obtained from a randomly selected subset of BraTS training data. ET = enhancing tumor, TC = tumor core, and WT = whole tumor, and BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps. The results in bold correspond to the best performing model in each tissue region.
Table 2. Mean Dice scores reported by similar studies using context-aware methods. * identifies the method ranking first in BraTS2020 (not using context-awareness mechanisms). 1 obtained via official BraTS test dataset, 2 obtained via official BraTS validation datasets, 3 obtained from a randomly selected subset of BraTS training data. ET = enhancing tumor, TC = tumor core, and WT = whole tumor, and BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps. The results in bold correspond to the best performing model in each tissue region.
ModelDatasetMean Dice ScoreClaimed Improvement
Using Context Awareness
TCETWT
Isensee et al. [33] *BraTS2020 185.9582.0388.95no context-aware mechanism used
Liu et al. [15]BraTS2017 384.0078.0089.00comparison between
models with and without
context-aware mechanism
not available
Liu et al. [18]BraTS2019 285.1075.9088.50
Ahmad et al. [16]BraTS2020 184.6779.1089.12
Chandra et al. [17]BraTS2018 173.3361.8282.99
Pei et al. [19]BraTS2019/20 383.5082.1089.50
Shen et al. [23]BraTS2013 371.8072.5088.702–3% (no p-value)
Shen et al. [22]BraTS2015 182.0075.0087.001.3% (p-value < 0.01)
Kao et al. [24]BraTS2018 179.3074.9087.501–2% (no p-value)
Le et al. [20]BraTS2018 288.8881.4190.952% (no p-value)
BLMBraTS2020 3
(36 cases)
81.8067.2090.80none
CIM81.9077.0090.10
CIP81.8070.4090.50
BLMBraTS2020 2
(125 cases)
81.6073.4190.54none
CIM81.6173.4390.58
CIP81.4474.0590.69
Table 3. Average training epochs and time across folds for the different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps).
Table 3. Average training epochs and time across folds for the different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps).
ModelAverage Training Time [hh:mm:ss]Epochs [250 Mini-Batches Each]
BLM8:15:1179
CIM20:02:39140
CIP12:58:57103
Table 4. Median Dice scores and 95% Hausdorff distance (HD) across target regions (ET = enhancing tumor, TC = tumor core, and WT = whole tumor) for the different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) trained using only T1-weighted with post-gadolinium enhancement as MR modality. Values are shown for low grade glioma (LGG) and high grade glioma (HGG) cases from the independent test dataset. There was no significant difference (p > 0.05) for neither of LGG and HGG cases when comparing CIM and CIP with BLM. The results in bold correspond to the best performing model in each tissue region.
Table 4. Median Dice scores and 95% Hausdorff distance (HD) across target regions (ET = enhancing tumor, TC = tumor core, and WT = whole tumor) for the different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) trained using only T1-weighted with post-gadolinium enhancement as MR modality. Values are shown for low grade glioma (LGG) and high grade glioma (HGG) cases from the independent test dataset. There was no significant difference (p > 0.05) for neither of LGG and HGG cases when comparing CIM and CIP with BLM. The results in bold correspond to the best performing model in each tissue region.
GradeModelMedian Dice ScoreMedian 95% HD
[Min, Max][Min, Max]
TCETWTTCETWT
LGG
(18 cases)
BLM79.9961.0886.659.25176.6786.01
[0.00, 98.06][0.00, 100.00][0.00, 95,34][1.00, 373.13][0.00, 373.13][0.00, 373.13]
CIM78.2137.9383.856.4237.859.26
[0.00, 98.04][0.00, 100.00][0.00, 95.16][1.00, 373.13][0.00, 373.13][2.00, 373.13]
CIP79.5852.2283.277.8723.4310.37
[0.00, 98.21][0.00, 100.00][0.00, 94.87][1.00, 373.13][0.00, 373.13][2.24, 373.13]
HGG
(18 cases)
BLM94.4189.7090.351.871.413.86
[87.55, 97.10][69.71, 96.29][71.90, 94.51][1.00, 4.90][1.00, 2.83][1.73, 15.17]
CIM93.9089.8689.581.871.413.93
[66.94, 97.06][70.93, 96.41][77.27, 94.49][1.00, 13.49][1.00, 3.00][2.00, 13.00]
CIP94.1089.7589.961.871.413.86
[63.08, 97.34][72.14, 96.45][79.54, 94.61][1.00, 22.38][1.00, 3.00][1.73, 10.72]
Table 5. Median Dice scores and 95% Hausdorff distance (HD) across target regions (ET = enhancing tumor, TC = tumor core, and WT = whole tumor) for the different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) trained on single center data. Values are shown for low grade glioma (LGG) and high grade glioma (HGG) cases from the independent test dataset (36 subjects). The statistical difference was insignificant (p > 0.05) for either LGG or HGG cases when comparing CIM and CIP with BLM. In bold, the results of the better performing model for each segmentation target.
Table 5. Median Dice scores and 95% Hausdorff distance (HD) across target regions (ET = enhancing tumor, TC = tumor core, and WT = whole tumor) for the different models (BLM = baseline model, CIM = contextual information model using binary masks, and CIP using probability maps) trained on single center data. Values are shown for low grade glioma (LGG) and high grade glioma (HGG) cases from the independent test dataset (36 subjects). The statistical difference was insignificant (p > 0.05) for either LGG or HGG cases when comparing CIM and CIP with BLM. In bold, the results of the better performing model for each segmentation target.
GradeModelMedian Dice ScoreMedian 95% HD
[Min, Max][Min, Max]
TCETWTTCETWT
LGG
(18 cases)
BLM78.4760.8192.776.407.642.73
[0.00, 97.31][0.00, 100.00][76.48, 97.11][0.00, 97.31][1.00, 373.13][0.00, 373.13]
CIM75.9956.7193.026.7113.422.53
[0.00, 96.85][0.00, 100.00][84.68, 97.21][1.00, 373.13][0.00, 373.13][1.41, 12.69]
CIP73.5360.1993.296.5413.982.45
[0.00, 97.05][0.00, 100.00][14.81, 96.94][1.00, 373.13][0.00, 373.13][1.00, 14.78]
HGG
(18 cases)
BLM91.6885.6689.302.242.003.23
[63.67, 96.15][53.57, 94.62][76.57, 96,62][1.00, 12.37][1.00, 6.63][1.00, 75.21]
CIM90.9786.3789.132.342.003.67
[67.18, 96.62][53.79, 94.29][80.83, 96.84][1.14, 10.05][1.00, 6.48][1.00, 13.64]
CIP92.1686.3389.412.271.233.38
[63.55, 96.36][54.59, 94.25][81.16, 96.63][1.41, 13.19][1.00, 6.40][1.00, 10.20]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tampu, I.E.; Haj-Hosseini, N.; Eklund, A. Does Anatomical Contextual Information Improve 3D U-Net-Based Brain Tumor Segmentation? Diagnostics 2021, 11, 1159. https://doi.org/10.3390/diagnostics11071159

AMA Style

Tampu IE, Haj-Hosseini N, Eklund A. Does Anatomical Contextual Information Improve 3D U-Net-Based Brain Tumor Segmentation? Diagnostics. 2021; 11(7):1159. https://doi.org/10.3390/diagnostics11071159

Chicago/Turabian Style

Tampu, Iulian Emil, Neda Haj-Hosseini, and Anders Eklund. 2021. "Does Anatomical Contextual Information Improve 3D U-Net-Based Brain Tumor Segmentation?" Diagnostics 11, no. 7: 1159. https://doi.org/10.3390/diagnostics11071159

APA Style

Tampu, I. E., Haj-Hosseini, N., & Eklund, A. (2021). Does Anatomical Contextual Information Improve 3D U-Net-Based Brain Tumor Segmentation? Diagnostics, 11(7), 1159. https://doi.org/10.3390/diagnostics11071159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop