Next Article in Journal
Research on Low-Light Environment Object Detection Algorithm Based on YOLO_GD
Previous Article in Journal / Special Issue
Development of a Seafloor Litter Database and Application of Image Preprocessing Techniques for UAV-Based Detection of Seafloor Objects
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multidimensional Framework Incorporating 2D U-Net and 3D Attention U-Net for the Segmentation of Organs from 3D Fluorodeoxyglucose-Positron Emission Tomography Images

Biomedical Engineering Laboratory, School of Electrical & Computer Engineering, National Technical University of Athens, 15773 Athens, Greece
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(17), 3526; https://doi.org/10.3390/electronics13173526
Submission received: 24 July 2024 / Revised: 2 September 2024 / Accepted: 3 September 2024 / Published: 5 September 2024
(This article belongs to the Special Issue Artificial Intelligence in Image Processing and Computer Vision)

Abstract

:
Accurate analysis of Fluorodeoxyglucose (FDG)-Positron Emission Tomography (PET) images is crucial for the diagnosis, treatment assessment, and monitoring of patients suffering from various cancer types. FDG-PET images provide valuable insights by revealing regions where FDG, a glucose analog, accumulates within the body. While regions of high FDG uptake include suspicious tumor lesions, FDG also accumulates in non-tumor-specific regions and organs. Identifying these regions is crucial for excluding them from certain measurements, or calculating useful parameters, for example, the mean standardized uptake value (SUV) to assess the metabolic activity of the liver. Manual organ delineation from FDG-PET by clinicians demands significant effort and time, which is often not feasible in real clinical workflows with high patient loads. For this reason, this study focuses on automatically identifying key organs with high FDG uptake, namely the brain, left cardiac ventricle, kidneys, liver, and bladder. To this end, an ensemble approach is adopted, where a three-dimensional Attention U-Net (3D AU-Net) is employed for robust three-dimensional analysis, while a two-dimensional U-Net (2D U-Net) is utilized for analysis in the coronal plane. The 3D AU-Net demonstrates highly detailed organ segmentations, but also includes many false positive regions. In contrast, 2D U-Net achieves higher reliability with minimal false positive regions, but lacks the 3D details. Experiments conducted on a subset of the public AutoPET dataset with 60 PET scans demonstrate that the proposed ensemble model achieves high accuracy in segmenting the required organs, surpassing current state-of-the-art techniques, and supporting the potential utilization of the proposed methodology in accelerating and enhancing the clinical workflow of cancer patients.

1. Introduction

Medical imaging plays a vital role in the diagnosis, treatment planning, and monitoring of various diseases. Among the different imaging modalities available, Fluorodeoxyglucose-Positron Emission Tomography (FDG-PET) has emerged as a powerful tool in oncology and neurology due to its ability to provide functional information about metabolic activity within tissues [1]. The accurate and automated segmentation of organs (allowing for the precise delineation and extraction of anatomical structures or regions of interest) from FDG-PET images is crucial for precise disease assessment, treatment response evaluation, and surgical planning [2]. However, in clinical practice, it is imperative that medical professionals have confidence in the tools they use, accurately measuring the size, volume, and growth rate of organs, tumors, and lesions [3]. Traditional manual segmentation methods are not only time consuming and labor intensive but also prone to variability between and within observers [4]. Consequently, there is an increasing demand for advanced computer-assisted techniques that can automate this process while maintaining high levels of accuracy and reproducibility. In this regard, machine learning has already made significant contributions to medicine and will continue to assist with various medical tasks, allowing for the earlier and more accurate detection of medical conditions [5,6]. Furthermore, advanced segmentation algorithms facilitate precise delineation of anatomical structures and pathological regions, reducing the likelihood of misdiagnosis and helping to ensure that subtle abnormalities are not overlooked [7]. In surgical planning, clear organ delineation helps surgeons plan procedures with greater confidence, reducing operative time, minimizing the risk of complications, and increasing the success rate of surgeries. When segmentation guides therapy, such as in radiotherapy or surgery, more precise targeting reduces the risk of damaging healthy tissue, leading to fewer side effects and better overall patient quality of life [8].
Semantic segmentation (the task of assigning a predefined class to each pixel) is central to these efforts, demanding a profound visual understanding to ensure precision and reliability in clinical outcomes. On this premise, semantic segmentation networks have evolved, with some of the most impactful architectures contributing to advancements in organ segmentation. For instance, SegNet, introduced by Vijay Badrinarayanan et al. [9], is an encoder–decoder network, which uses convolutional and pooling layers for down-sampling in the encoder and up-sampling in the decoder, laid the groundwork for subsequent developments in the field. In addition, architectures like U-Net, DeepLab, and SegResNet have emerged, offering enhanced performance for organ segmentation tasks [10,11,12]. Among these, U-Net has become particularly popular, featuring a characteristic “U” shape formed by its encoding contracting pathway and expansive decoding pathway [10]. The contracting pathway down-samples the input image through convolutional and pooling operations, while the expansive pathway upscales the feature maps using transposed convolutional layers. Skip connections between corresponding layers in the encoding and decoding pathways allow U-Net to recover spatial details effectively, leading to more accurate segmentation outcomes. Expanding upon the U-Net framework, nnU-Net can self-configure and adapt to any dataset, delivering high performance, enabling it to excel across various medical segmentation tasks [13]. Moreover, DeepLab has made notable contributions to semantic segmentation, particularly through the use of atrous (or dilated) convolutions [11]. This allows DeepLab models, including DeepLabv2, DeepLabv3, and DeepLabv3+, to capture multi-scale information effectively without significantly increasing computational costs [11,14,15]. Recently, the transformers can capture global context and long-range dependencies, a feature that traditional convolutional neural networks (CNNs) lack [16]. UNETR, a hybrid architecture combining U-Net and transformers, leverages this capability to address the limitations of conventional CNNs, offering enhanced segmentation accuracy [17].
Despite significant recent advancements in machine learning for medical imaging, segmenting organs from 3D FDG-PET images remains a challenge due to several inherent factors of diagnostic visualization. FDG-PET images are characterized by low spatial resolution, limited contrast, and substantial noise, all of which can impede the precise delineation of organ boundaries [18]. Additionally, variations in image intensity caused by differences in patient physiology, imaging protocols, and equipment settings further complicate the segmentation process. In our previous study focusing on the identification of Metastatic Melanoma tumors from FDG-PET/CT images, organs with high uptake in them produce many false positives when a model tries to locate tumor lesions throughout the body [19].
To overcome these obstacles, in this study, we propose a machine learning framework for organ segmentation in PET data through an ensemble model that effectively integrates 2D and 3D data perspectives. Our rationale stems from the fact that 2D and 3D networks (i.e., Attention U-Net (3D AU-Net) [20,21] and two-dimensional U-Net [10] (2D U-Net)) can enhance flexibility and improve performance (especially when 3D data is unreliable or computational resources are limited), harnessing slice-wise processing (2D U-Net) and the ability to capture spatial dependencies across slices (3D AU-Net’s). From this standpoint, the combined model can increase adaptability and may lead to better generalization across varied datasets or conditions. As such we aim to carry out the following:
  • Address a significant gap in the field by providing a dedicated solution for segmenting PET images;
  • Leverage the power of combined machine learning models to enhance automatic segmentation performance;
  • Propose a new, ensemble approach, that is particularly suited for achieving high validation performance when training on limited samples.
In this regard, utilizing an annotated PET dataset of 60 scans, we combine the 3D AU-Net volumetric understanding with the 2D U-Net slice-wise processing capabilities to enhance specific clinical applications in PET image segmentation which are often overlooked in favor of other imaging modalities like CT or MRI. The proposed method results in a high segmentation performance regarding the brain, left cardiac ventricle, kidneys, liver, and bladder. Rigorous testing displays significant improvement in segmentation accuracy, outperforming benchmark models and the individual 2D U-Net and 3D AU-Net counterparts.

2. Materials and Methods

2.1. Dataset

In the proposed methodology, a subset of the public AutoPET [22] dataset with 60 PET scans was used, due to these volumes being the only ones to contain annotations for all target organs. The AutoPet dataset contains a total of 1,014 PET scans. Of these scans, 501 studies were from patients with malignant lymphoma, melanoma, and non-small cell lung cancer (NSCLC), and 513 studies without PET-positive malignant lesions (negative controls). All examinations were conducted at the University Hospital Tübingen using a single PET/CT scanner (Siemens Biograph mCT, Siemens Healthineers, Knoxville, TN, USA). Both CT and PET scans followed standardized protocols to ensure uniformity across all studies. Specifically, CT scans were performed with a consistent set of scan parameters, including slice thickness, tube voltage, and current, while PET scans adhered to a uniform acquisition time and reconstruction algorithm. The only variable was the dose of the F-FDG injection, which was adjusted based on the patient’s weight to optimize image quality. For this subset, experts created detailed masks for the following organs: bladder, brain, liver, left ventricle of the heart, and kidneys. The dataset offers 3D volumes of both CT and FDG PET data, presented as stacks of axial slices, and all the data covers comprehensive whole-body examinations. The usual scan range encompasses from the skull base to the mid-thigh level, with the potential for extension to include the entire body, head to legs, based on clinical necessity. The chosen scans consist of whole body FDG PET/CT images. The PET data were reconstructed iteratively with Gaussian post-reconstruction smoothing. Then, standardization was applied by converting image units from activity counts to standardized uptake values (SUV).

2.2. Ensemble Model Architecture

The neural network approach proposed in this study consists of an ensemble of two U-Net architectures that complement each other.

2.2.1. 3D AU-Net

The first U-Net utilized is a three-dimensional Attention U-Net for robust three-dimensional analysis. What sets this architecture apart is the use of Attention Gates, a mechanism that improves performance in tasks requiring precise localization by learning to selectively focus on particular regions of interest [20]. It takes feature maps from the encoder and decoder, computes attention maps to highlight important regions, and uses these maps to weight the feature maps from the encoder before merging them with the decoder’s feature maps. The most significant advantage of using a 3D architecture is its ability to leverage volumetric information. It considers spatial relationships in all three dimensions (X, Y, and Z), capturing contextual information from adjacent slices. This allows the model to better understand the 3D structure of objects and organs within the volume. The architecture of our three-dimensional Attention U-Net comprises four convolutional blocks in each of the encoding and decoding pathways, as depicted in Figure 1. The original volumes are of shape N × 400 × 400 × 1, where N is the depth dimension and varies between individuals. During training, each volume is divided into overlapping (along the depth dimension) patches of size 80 × 400 × 400 × 1, which are individually fed to the network and are gradually reduced to size 5 × 25 × 25 × 128, where 128 is the number of feature maps produced. In the decoding pathway, the patch volumes are gradually upscaled to 80 × 400 × 400 × 8 voxels. Subsequently, a 1 × 1 convolutional operation is applied, which is often referred to as a pointwise convolution. This operation serves to merge information from the different channels and generate the final output. Deep supervision [23] was added to the model for the computation of loss, a technique pivotal in enhancing the training process of neural networks. Deep supervision involves the incorporation of additional supervision signals at intermediate layers of the network architecture, thereby facilitating more effective gradient flow and addressing the issue of vanishing gradients. This approach fosters better convergence during training, enabling the network to learn more efficiently and yielding improved performance outcomes.

2.2.2. 2D U-Net

The second U-Net used in this study is a two-dimensional U-Net (2D U-Net). To facilitate this transition, the three-dimensional data were converted into two-dimensional coronal view images, as it was heuristically observed that the model exhibited superior recall performance in this orientation compared to sagittal and axial views. The input to this network is an image with dimensions 400 × 400 × 1 pixels. The architecture, like in the 3D variant, includes four convolutional blocks in each of the encoding and decoding pathways, as depicted in Figure 2. The encoding pathway progressively reduces the spatial dimensions to 25 × 25 × 1024, before being up-sampled back to its original size by the decoding pathway.

2.2.3. Ensemble

The final neural network model employed in this study is an ensemble of 2D U-Net and 3D AU-Net. Despite their architectural similarities, each model yields different end results and thus their combination enhances the overall final accuracy. The two models are trained independently. Upon completion of training for both networks, a straightforward post-processing procedure is applied.
Initially, the segmentation masks computed by 2D U-Net are combined into a unified 3D volume. Since 2D U-Net operates independently on individual slices, any false positive regions are likely to be isolated to specific slices. When the 2D segmentation masks are combined into a 3D volume, these isolated false positives do not align or connect with one another across slices. Since false positives tend to be random and scattered, their lack of spatial continuity means they remain small and isolated when viewed in the context of the entire 3D volume. In contrast, the true organ regions are more likely to be spatially continuous across multiple adjacent slices. When these consistent regions are combined into a 3D volume, they form a large, connected component, representing the actual structure of the organ. The largest connected component in the 3D mask is therefore likely to correspond to the true organ, while the false positives, being smaller and more fragmented, do not combine into significant structures. To isolate this true positive component, we apply a simple criterion of selecting the largest connected component within the mask.
Subsequently, the segmentation results of the patches from the 3D AU-Net are aggregated to generate a second segmentation mask for the entire volume. Since the 3D AU-Net processes patches of the 3D volume, and not the entire scan at once, each patch is segmented independently. For overlapping patches, their union is computed. However, because of the complexity and sensitivity of the model, this also introduces numerous false positives. To improve the accuracy of the segmentation, the outputs of both 2D U-Net and 3D AU-Net are combined. 2D U-Net provides a coarser but more reliable segmentation, giving a general idea of where the target organ is located. The 3D AU-Net, while more detailed, includes many false positives due to its finer resolution. Therefore, the final step involves selectively merging the outputs from 2D U-Net and 3D AU-Net. By focusing on the regions where the segmentations from both networks overlap, the method capitalizes on the strengths of each model: 2D U-Net’s robustness in avoiding false positives and 3D AU-Net’s detailed segmentation. The rationale is that the true organ region is likely to be consistently identified by both networks, resulting in an overlapping area that represents the most reliable segmentation. Figure 3 graphically depicts this process, visualizing the outputs masks of 2D U-Net, 3D AU-Net, and the ensemble method.

2.3. Training Setup

To ensure data consistency and improve training effectiveness through a uniform voxel spacing, all volumes were ensured to have the same uniform voxel spacing values of (2.036, 2.036, 3.0) prior to training, as computed by the median value. The voxel spacing values indicate the distance between the slices in the Z plane, and the distance between neighboring pixels in the X and Y plane. Voxel spacing normalization is crucial for medical imaging, as scans from different patients or scanners might have varying resolutions or slice thicknesses. It has previously been shown that voxel normalization is a crucial pre-processing step as it directly impacts performance in texture-related studies [24]. In addition, the uniformity achieved across all samples ensures that the models do not learn scanner- or patient-specific biases corresponding to different resolutions and scales [25].
In addition, intensity normalization was applied, centering the data around their mean and scaling them by their standard deviation, resulting in voxels with a mean intensity of 0 and a standard deviation of 1. This pre-processing step helps in stabilizing the gradients during backpropagation, which contributes to solving problems like vanishing or exploding gradients. Moreover, intensity normalization makes sure that all input features are on a similar scale, resulting in faster model convergence during training. Otherwise, the model needs to adjust its weights to accommodate the large differences in input scales, slowing down the training [26].
Regularization techniques are critical in neural networks to prevent overfitting and improve generalization to unseen data. Overfitting occurs when a model performs well on the training data but fails to generalize to new data. One widely used regularization method is Batch Normalization, which normalizes the inputs to each layer by adjusting and scaling activations. By stabilizing the learning process, Batch Normalization inherently acts as a form of regularization, potentially reducing the need for other techniques like Dropout [27]. We implemented this technique in all our experiments. Another regularization method used in our experiments is data augmentation. Data augmentation is a technique used to artificially expand the size of a training dataset by applying various transformations to the existing data. By introducing variations such as rotations, translations, flips, scaling, and adjustments in brightness or contrast, it enhances the model’s ability to generalize by exposing it to a wider variety of scenarios during training, and thereby reducing the risk of overfitting. In our training procedures, we applied random rotations of up to 20 degrees, horizontal and vertical flips with a 50% probability, scaling within a range of 0.8 to 1.2, and brightness adjustments varying by ±10%.
Each network was individually trained multiple times using varying amounts of data (15, 20, 30, 40, and 50 samples), with consistent validation across all experiments using a fixed set of 10 data points. This approach ensured robustness and reliability in our assessment, as it allowed us to observe the performance of each network under different training conditions and data quantities. The 3D AU-Net network was trained for a total of 50 epochs, and 2D U-Net for a total of 25 epochs, as further training was not found to further decrease training loss. To calculate the loss, the combined sum of the Dice score [28] and cross-entropy [29] loss was utilized. Dice score is a popular metric employed in numerous image segmentation tasks, including biomedical segmentation. The Dice loss focuses on measuring the spatial agreement between the predicted segmentation mask and the ground truth, making it particularly effective for dealing with imbalanced data. Mathematically, the Dice loss is calculated as follows:
Dice Loss = 1 2 · Intersection Union + Intersection .
In this formula, “Intersection” refers to the number of pixels correctly classified in both the predicted and ground truth masks, while “Union” represents the total number of pixels belonging to either the predicted or ground truth masks. By utilizing the Dice loss during training, the models can better capture fine details and boundaries, leading to improved segmentation performance and robustness in various applications, especially in the context of biomedical imaging and analysis.
Binary cross-entropy is a fundamental loss function in binary classification tasks, extensively used in training neural networks. It quantifies the disparity between predicted probabilities and true binary labels, computing the average number of bits required to encode the true labels given the predicted probabilities. Mathematically, it is expressed as a summation of the logarithmic differences between the true and predicted probabilities for each sample. Binary cross-entropy effectively penalizes deviations between predicted and actual labels, facilitating accurate classification. This loss function is commonly employed in neural network training for its effectiveness in binary classification scenarios.
The binary cross-entropy loss function is expressed as follows:
H ( y , y ^ ) = i N y i log ( y ^ i ) ,
where H ( y , y ^ ) is the binary cross-entropy, N is the total number of samples, y i is the true binary label for sample i (0 or 1), and y ^ i is the predicted probability that sample i belongs to class 1 (the positive class).
The Adam optimizer was used for both networks, with different learning rates for each. The values were found heuristically, and were set to 0.0003 and 0.003 for 2D U-Net and 3D U-Net respectively.
The choice of batch size and learning rate can significantly influence model performance. These parameters must be carefully selected, as they directly affect the training process. Additionally, computational resources play a critical role in determining these values. Poor choices can lead to exceeding resource limits or causing the training process to slow down, potentially hindering the model’s overall performance and development. This has been demonstrated in the past; improper selection of batch size and learning rate can result in suboptimal convergence and may even cause models to diverge during training [30]. The batch sizes used were 1 for 3D U-Net and 5 for 2D U-Net. While this choice leads to significantly slower training, it can also result in better model generalization, due to the noise introduced by the small batch sizes acting as a form of regularization.
For a more comprehensive comparison and analysis of the ensemble method’s performance, two additional models were trained: DeepLabV3 [14] and SegResNet [12]. Both models were trained using the same parameters applied in the training of 2D U-Net and 3D Attention U-Net, ensuring consistency across all models.

3. Results

The evaluation of our proposed method was based on the Dice score, mean Intersection over Union (IoU), precision, and recall. Additionally, the standard deviation of those metrics was also calculated. These metrics provided a comprehensive assessment of the segmentation performance and enabled a thorough comparison with existing approaches.
  • Dice Score (also known as Dice Coefficient or F1 Score), similarly to the Dice Loss, is defined as the ratio of the intersection of the predicted and ground truth regions to the average size of the regions. The Dice score ranges from 0 to 1, where a value of 1 indicates a perfect overlap between the segmentation and ground truth.
  • Intersection over Union (IoU): IoU is another popular metric for evaluating segmentation performance. It measures the average overlap between the predicted and ground truth regions for all classes of interest. The IoU for each class is calculated as the ratio of the intersection to the union of the predicted and ground truth regions. The IoU provides a comprehensive evaluation of the model’s performance across all classes and is also known as the Jaccard Index.
  • Precision and Recall [31]: Precision and recall are metrics commonly used in binary or multi-class segmentation tasks. Precision (also known as positive predictive value) measures the proportion of true positive predictions (correctly identified pixels) out of all pixels predicted as positive (both true positive and false positive). Recall (also known as sensitivity) measures the proportion of true positive predictions out of all actual positive pixels (both true positive and false negative). Precision and recall are complementary metrics, and a trade-off between them is often encountered when fine-tuning segmentation models.
  • The standard deviation (std) of the scores of the metrics measures the variability or consistency of the model’s performance across different samples. A lower standard deviation indicates more stable and reliable segmentation results, while a higher standard deviation suggests greater variability in performance.
Table 1, Table 2, Table 3, Table 4 and Table 5 demonstrate the efficacy of the devised ensemble methodology in achieving a superior performance across the employed metrics, regardless of the quantity of training data utilized. Notably, incremental increases in training data did not yield substantial improvements in scores. Intriguingly, certain scenarios indicated that training with a reduced dataset yielded higher scores. A comparative analysis between 2D U-Net and 3D Attention U-Net reveals a nuanced performance landscape with no superiority identified between the two models. It is noteworthy, however, that, in the context of the segmentation of the brain, 2D U-Net exhibited superior results.
Furthermore, a paired t-test was employed to compare the statistical relationships of the obtained results. This test is particularly useful for determining whether there is a significant difference between the means of two related groups. It operates by calculating the difference between each pair of observations and then analyzing these differences, with the assumption that the differences are normally distributed. By focusing on these paired differences rather than the absolute values, the paired t-test effectively controls for variability between subjects, providing a more precise measure of the effect or change. If the p-value is less than or equal to 0.05, the difference is typically considered statistically significant, meaning there is strong evidence to reject the null hypothesis. If the p-value is greater than 0.05, the difference is generally not considered statistically significant [32]. The p-values were obtained by comparing the Dice scores on the validation dataset of each model with those of the final ensemble model. The p-values are presented in Table 6.

4. Discussion

Analysis of the results presented in Section 4 reveals that, when combining the two models, namely 2D U-Net and 3D Attention U-Net, their individual strengths and weaknesses are effectively balanced. Even when working with a limited dataset, a significantly high segmentation performance was achieved.
Notably, our best-performing models, particularly in brain and bladder segmentation tasks, demonstrated a remarkable accuracy, reaching a Dice score of 97%. This surpasses the performance of the individual models by a significant margin. In the case of bladder segmentation, neither of the individual models achieved a Dice score higher than 50%; however, our ensemble method consistently delivered strong results, attaining a Dice score close to 97%. In the case of brain segmentation, the 2D model exhibited satisfactory results, achieving up to 86% Dice score. On the other hand, the performance of the 3D Attention U-Net was lacking, yielding an overall Dice score of less than 50%. However, when these two models were combined, their ensemble yielded a significantly higher score, reaching up to 97%.
As demonstrated by our results, the significant increase in the Dice scores can be primarily attributed to the increased precision scores. Specifically, the models’ capability of accurately identifying organs is apparent from the consistently high recall values across all segmentation tasks. Nonetheless, precision scores suffer due to a high number of false positive results. Notably, our ensemble method substantially improves precision metrics, resulting in a considerable increase in the overall Dice score.
The kidneys were a notable exception as the performance of the ensemble model was lower than that of the 3D AU-Net. This behavior can be attributed to the extraordinarily low performance of the 2D U-Net model, which struggled to correctly identify both kidneys, often identifying one kidney and a false positive mask instead.
Overall, the model performance increased with the number of samples, in all but one task. The segmentation of the left ventricle of the heart is a challenging task due to the fact that it is the smallest of all the target organs. Therefore, the networks struggled to accurately identify this particular organ, consistently demonstrating low precision values and inconsistent results, indicative of a high level of variance.
When comparing SegResNet with the other models, it is evident that it outperforms all of them except for the ensemble method. For the bladder, brain, left ventricle, and liver, SegResNet achieves the second-best results, just behind the ensemble model. In the case of the kidneys, SegResNet surpasses the ensemble model when trained on 50 samples but does not outperform the 3D Attention U-Net. On the other hand, DeepLab generally performs less effectively than SegResNet but closely matches the results of 2D U-Net and 3D Attention U-Net, and in many cases providing better results. Notably, DeepLab shows performance variability, with results clearly enhancing as the training data sample size increases.
Figure 4 and Figure 5 exemplify the performance of the final ensemble model, providing a visual representation of its effectiveness in organ segmentation. Those figures demonstrate the complementarity between the two models, the post-processed 2D U-Net and 3D Attention U-Net. This is particularly evident in Figure 4, showcasing the coronal view of outputs, where 3D Attention U-Net segments more effectively all the organs but presents several false positives. False positives become apparent due to the overlapping colors. In contrast, the 2D model exhibits almost zero false positive organ regions and correctly segments the organs, though it misses a few details. However, the ensemble model’s final output achieves better segmentation without any false positives, showcasing the synergistic benefits of combining these approaches. Figure 5 showcases the sagittal view of the outputs. Interestingly, the same consistent pattern as the previous figures emerges. The 3D Attention U-Net model tends to produce numerous false positives in its identification. In this case, 2D U-Net also produced many false positive regions. The ensemble method’s output is almost identical to the ground truth, except in Figure 5, where it fails to segment the kidney. This result is consistent with the Dice scores reported.
Finally, the results presented in Table 6 show the p-values obtained from the paired t-test. When comparing the two U-Net models (2D U-Net and 3D Attention U-Net), most p-values are below 0.01, with the highest values being 0.005 for the heart and 0.01 for the liver, reflecting highly significant differences between the compared groups. However, for the kidneys, where the Dice scores did not favor the ensemble model, when the models were trained on 20 and 30 samples, the p-values were higher in comparison to 3D Attention U-Net. This suggests that the difference between groups for the kidneys was not statistically significant. Moreover, because 3D Attention U-Net outperforms the ensemble model when trained with 40 and 50 data, the p-value remains smaller than 0.005.
When comparing the p-values of DeepLab and SegResNet against the ensemble method, their results are generally higher than those of the two U-Net models. Despite these improved results, their p-values in most cases are below 0.05, indicating significant differences compared to the ensemble method. For the segmentation of the brain, SegResNet shows p-values of 0.1 and 0.4 when trained on 15 and 20 samples, respectively, indicating no significant difference from the ensemble model. Similarly, for the kidneys, DeepLab yields p-values of 0.19 and 0.8 when trained on 30 and 50 samples, respectively, while SegResNet shows a p-value of 0.2 with 30 samples, again suggesting no significant difference.

Limitations and Future Recommendations

Some considerations are warranted when interpreting the results of this study. Firstly, while the results demonstrate that the proposed ensemble approach can achieve considerably better results than the individual models when applied on a limited dataset of full body FDG-PET scans, broader validation across multiple, diverse datasets is necessary to establish the generalizability of the proposed approach. Therefore, future work should include experiments on more varied datasets, as well as ablation studies to better understand the contribution of each component in the hybrid model.
Secondly, it is important to emphasize that the proposed approach is particularly suited to limited datasets. As the dataset sample size increases, the advantages of the ensemble approach is likely to diminish due to a convergence in the performance of the individual 2D and 3D models. In such cases, the added parameters and computational overhead introduced by combining two separate deep neural networks may outweigh its benefits, rendering it less efficient compared to other established techniques. While further validation on larger datasets is necessary to assess the scalability of the proposed approach, it is crucial to recognize that the primary contribution of this work lies in addressing the challenges associated with data scarcity, as is often the case in medical imaging datasets, and demonstrate that two models with sub-optimal performance due to training on scarce data can provide exceptional results when combined.
Finally, the study concentrated on the performance of algorithmic segmentation, without addressing potential limitations related to their application in clinical practice, such as computational resource demands, training duration, and integration with existing medical imaging systems. Consequently, caution is advised before deploying the deep learning model in real-world settings, and further refinements or enhancements may be required to facilitate the adoption of automated deep learning segmentation systems.

5. Conclusions

In this study, we introduced a novel approach for semantic organ segmentation in FDG-PET images, harnessing the strengths of a three-dimensional Attention U-Net (3D AU-Net) and a two-dimensional U-Net (2D U-Net) in an ensemble framework. Our combined model exhibited substantial performance improvements compared to the individual 2D and 3D models. Notably, while the individual models’ performance was limited when training on few data samples, our ensemble model consistently achieved higher performance. Thus, our results demonstrate that, where the individual 2D and 3D U-Net models may fail to converge to an optimal solution due to data scarcity, their combination in an ensemble model can still provide excellent results.
By combining the outputs of both models we leveraged their complementary capabilities achieving a much more effective and robust segmentation model. The 3D model’s ability to capture finer details and provide highly detailed organ segmentation is valuable for a more holistic understanding of the organ’s overall structure. It excels in handling complex anatomical features and capturing spatial relationships across different slices. On the other hand, the 2D model’s proficiency in achieving more detailed categorization is beneficial in capturing local features within individual slices. Its capacity to accurately determine the spatial location of the organ within the volume helps in precisely localizing the boundaries, thus reducing false positive predictions and enhancing segmentation accuracy in the final model.
In conclusion, the fusion of 3D and 2D models generated a synergistic effect, surpassing the capabilities of individual models. By continuing to improve upon segmentation techniques, machine learning models have the potential to empower clinicians with more accurate insights, reducing diagnostic uncertainties, and enabling timely, targeted treatments that will ultimately lead to better patient outcomes.

Author Contributions

Conceptualization, A.V.; methodology, A.V. and T.P.V.; software, A.V. and I.V.; validation, A.V., I.V. and I.K.; formal analysis, A.V.; investigation, A.V.; resources, I.V., T.P.V. and G.K.M.; data curation, A.V. and T.P.V.; writing—original draft preparation, A.V.; writing—review and editing, A.V., I.V. and T.P.V.; visualization, A.V.; supervision, G.K.M.; project administration, G.K.M.; funding acquisition, G.K.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original data presented in the study are openly available at https://www.cancerimagingarchive.net/collection/fdg-pet-ct-lesions/ (accessed on 15 July 2022).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Phelps, M.E. Positron emission tomography provides molecular imaging of biological processes. Proc. Natl. Acad. Sci. USA 2000, 97, 9226–9233. [Google Scholar] [CrossRef] [PubMed]
  2. Wahl, R.L.; Jacene, H.; Kasamon, Y.; Lodge, M.A. From RECIST to PERCIST: Evolving Considerations for PET Response Criteria in Solid Tumors. J. Nucl. Med. 2009, 50, 122S–150S. [Google Scholar] [CrossRef] [PubMed]
  3. Javaid, M.; Haleem, A.; Pratap Singh, R.; Suman, R.; Rab, S. Significance of machine learning in healthcare: Features, pillars and applications. Int. J. Intell. Networks 2022, 3, 58–73. [Google Scholar] [CrossRef]
  4. Montagne, S.; Hamzaoui, D.; Allera, A.; Ezziane, M.; Luzurier, A.; Quint, R.; Kalai, M.; Ayache, N.; Delingette, H.; Renard-Penna, R. Challenge of prostate MRI segmentation on T2-weighted images: Inter-observer variability and impact of prostate morphology. Insights Imaging 2021, 12, 71. [Google Scholar] [CrossRef]
  5. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  6. Iliadou, V.; Kakkos, I.; Karaiskos, P.; Kouloulias, V.; Platoni, K.; Zygogianni, A.; Matsopoulos, G.K. Early Prediction of Planning Adaptation Requirement Indication Due to Volumetric Alterations in Head and Neck Cancer Radiotherapy: A Machine Learning Approach. Cancers 2022, 14, 3573. [Google Scholar] [CrossRef]
  7. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A Survey on Deep Learning in Medical Image Analysis. Med Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef]
  8. Kakkos, I.; Vagenas, T.P.; Zygogianni, A.; Matsopoulos, G.K. Towards Automation in Radiotherapy Planning: A Deep Learning Approach for the Delineation of Parotid Glands in Head and Neck Cancer. Bioengineering 2024, 11, 214. [Google Scholar] [CrossRef]
  9. Badrinarayanan, V.; Handa, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Robust Semantic Pixel-Wise Labelling. arXiv 2015, arXiv:1505.07293. [Google Scholar] [CrossRef]
  10. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  11. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. arXiv 2017, arXiv:1606.00915. [Google Scholar] [CrossRef]
  12. Myronenko, A. 3D MRI brain tumor segmentation using autoencoder regularization. arXiv 2018, arXiv:1810.11654. [Google Scholar]
  13. Isensee, F.; Petersen, J.; Klein, A.; Zimmerer, D.; Jaeger, P.F.; Kohl, S.; Wasserthal, J.; Koehler, G.; Norajitra, T.; Wirkert, S.; et al. nnU-Net: Self-adapting Framework for U-Net-Based Medical Image Segmentation. arXiv 2018, arXiv:1809.10486. [Google Scholar] [CrossRef]
  14. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. arXiv 2018, arXiv:1802.02611. [Google Scholar]
  15. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  16. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  17. Hatamizadeh, A.; Tang, Y.; Nath, V.; Yang, D.; Myronenko, A.; Landman, B.; Roth, H.; Xu, D. UNETR: Transformers for 3D Medical Image Segmentation. arXiv 2021, arXiv:2103.10504. [Google Scholar]
  18. Eiber, M.; Weirich, G.; Holzapfel, K.; Souvatzoglou, M.; Haller, B.; Rauscher, I.; Beer, A.J.; Wester, H.J.; Gschwend, J.; Schwaiger, M.; et al. Simultaneous 68Ga-PSMA HBED-CC PET/MRI Improves the Localization of Primary Prostate Cancer. Eur. Urol. 2016, 70, 829–836. [Google Scholar] [CrossRef]
  19. Vagenas, T.P.; Economopoulos, T.L.; Sachpekidis, C.; Dimitrakopoulou-Strauss, A.; Pan, L.; Provata, A.; Matsopoulos, G.K. A Decision Support System for the Identification of Metastases of Metastatic Melanoma Using Whole-Body FDG PET/CT Images. IEEE J. Biomed. Health Inform. 2023, 27, 1397–1408. [Google Scholar] [CrossRef]
  20. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-Net: Learning Where to Look for the Pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
  21. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2023, arXiv:1706.03762. [Google Scholar] [CrossRef]
  22. Gatidis, S.; Hepp, T.; Früh, M.; Fougère, C.L.; Nikolaou, K.; Pfannenberg, C.; Schölkopf, B.; Küstner, T.; Cyran, C.; Rubin, D. A whole-body FDG-PET/CT Dataset with manually annotated Tumor Lesions. Sci. Data 2022, 9, 601. [Google Scholar] [CrossRef]
  23. Lee, C.Y.; Xie, S.; Gallagher, P.W.; Zhang, Z.; Tu, Z. Deeply-Supervised Nets. In Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics, San Diego, CA, USA, 9–12 May 2015. [Google Scholar]
  24. Wahid, K.A.; He, R.; McDonald, B.A.; Anderson, B.M.; Salzillo, T.; Mulder, S.; Wang, J.; Sharafi, C.S.; McCoy, L.A.; Naser, M.A.; et al. Intensity standardization methods in magnetic resonance imaging of head and neck cancer. Phys. Imaging Radiat. Oncol. 2021, 20, 88–93. [Google Scholar] [CrossRef] [PubMed]
  25. Hsiao, C.C.; Peng, C.H.; Wu, F.Z.; Cheng, D.C. Impact of Voxel Normalization on a Machine Learning-Based Method: A Study on Pulmonary Nodule Malignancy Diagnosis Using Low-Dose Computed Tomography (LDCT). Diagnostics 2023, 13, 3690. [Google Scholar] [CrossRef] [PubMed]
  26. Ghazvanchahi, A.; Maralani, P.J.; Moody, A.R.; Khademi, A. Effect of Intensity Standardization on Deep Learning for WML Segmentation in Multi-Centre FLAIR MRI. Proc. Mach. Learn. Res. 2023, 227, 1923–1940. [Google Scholar]
  27. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar] [CrossRef]
  28. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Jorge Cardoso, M. Generalised Dice Overlap as a Deep Learning Loss Function for Highly Unbalanced Segmentations. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Cardoso, M.J., Arbel, T., Carneiro, G., Syeda-Mahmood, T., Tavares, J.M.R., Moradi, M., Bradley, A., Greenspan, H., Papa, J.P., Madabhushi, A., et al., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2017; pp. 240–248. [Google Scholar] [CrossRef]
  29. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  30. Smith, L.N. Cyclical Learning Rates for Training Neural Networks. arXiv 2017, arXiv:1506.01186. [Google Scholar]
  31. Powers, D.M.W. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv 2020, arXiv:2010.16061. [Google Scholar] [CrossRef]
  32. Ruxton, G.D. The unequal variance t-test is an underused alternative to Student’s t-test and the Mann–Whitney U test. Behav. Ecol. 2006, 17, 688–690. [Google Scholar] [CrossRef]
Figure 1. Architecture of 3D Attention U-Net. Each block in the diagram is colored according to its operation type. The output size of the feature maps from each block is specified below it. The skip connections between the encoder and the decoder parts of the network, indicated with the blue arrows, also contain an Attention Gate, denoted with the circled A.
Figure 1. Architecture of 3D Attention U-Net. Each block in the diagram is colored according to its operation type. The output size of the feature maps from each block is specified below it. The skip connections between the encoder and the decoder parts of the network, indicated with the blue arrows, also contain an Attention Gate, denoted with the circled A.
Electronics 13 03526 g001
Figure 2. Architecture of 2D U-Net. Differences between its 3D counterpart in Figure 1 include the fewer dimensions as indicated below each block, as well as the absence of the Attention Gate in the skip connections indicated by the blue arrows.
Figure 2. Architecture of 2D U-Net. Differences between its 3D counterpart in Figure 1 include the fewer dimensions as indicated below each block, as well as the absence of the Attention Gate in the skip connections indicated by the blue arrows.
Electronics 13 03526 g002
Figure 3. Example of the proposed approach process for the case of brain segmentation. Green color indicates the ground truth, while blue indicates network predictions. (a) depicts the ground truth segmentation. (b) depicts the 2D U-Net output, which, while successful in localizing the brain region, fails to accurately capture the brain’s geometry. (c) The 3D Attention U-Net output provides a more complete segmentation of the full brain geometry, but additionally introduces several false positive segmentation masks. (d) The final output merges the results of the 2D and 3D U-Net counterparts, preserving only the correct brain segmentation.
Figure 3. Example of the proposed approach process for the case of brain segmentation. Green color indicates the ground truth, while blue indicates network predictions. (a) depicts the ground truth segmentation. (b) depicts the 2D U-Net output, which, while successful in localizing the brain region, fails to accurately capture the brain’s geometry. (c) The 3D Attention U-Net output provides a more complete segmentation of the full brain geometry, but additionally introduces several false positive segmentation masks. (d) The final output merges the results of the 2D and 3D U-Net counterparts, preserving only the correct brain segmentation.
Electronics 13 03526 g003
Figure 4. Comparison of the ensemble method with the individual models and ground truth, displayed in the coronal plane. (a) shows the output of 2D U-Net, (b) shows the output of 3D Attention U-Net, (c) displays the output of the ensemble method, and (d) presents the ground truth. The masks are color coded as follows: red for the left ventricle of the heart, blue for the liver and brain, yellow for the bladder, and green for the kidneys. Overlapping colors, particularly in (b), indicate false positive regions. The final ensemble method output, shown in (c), eliminates false positives closely resembling the ground truth.
Figure 4. Comparison of the ensemble method with the individual models and ground truth, displayed in the coronal plane. (a) shows the output of 2D U-Net, (b) shows the output of 3D Attention U-Net, (c) displays the output of the ensemble method, and (d) presents the ground truth. The masks are color coded as follows: red for the left ventricle of the heart, blue for the liver and brain, yellow for the bladder, and green for the kidneys. Overlapping colors, particularly in (b), indicate false positive regions. The final ensemble method output, shown in (c), eliminates false positives closely resembling the ground truth.
Electronics 13 03526 g004
Figure 5. Comparison of the ensemble method with individual models and the ground truth, displayed in the sagittal view. (a) shows the output of 2D U-Net, (b) shows the output of 3D Attention U-Net, (c) displays the output of the ensemble method, and (d) presents the ground truth. The masks are color coded as follows: red for the left ventricle, blue for the liver and brain, yellow for the bladder, and green for the kidneys. The bladder and liver are not present in this specific slice, yet their color-coded masks are visible, indicating some falsely segmented regions. False positive regions are evident in (a,b). The final ensemble output, shown in (c), reduces the false positives, and closely resembles the ground truth but fails to include the kidney mask.
Figure 5. Comparison of the ensemble method with individual models and the ground truth, displayed in the sagittal view. (a) shows the output of 2D U-Net, (b) shows the output of 3D Attention U-Net, (c) displays the output of the ensemble method, and (d) presents the ground truth. The masks are color coded as follows: red for the left ventricle, blue for the liver and brain, yellow for the bladder, and green for the kidneys. The bladder and liver are not present in this specific slice, yet their color-coded masks are visible, indicating some falsely segmented regions. False positive regions are evident in (a,b). The final ensemble output, shown in (c), reduces the false positives, and closely resembles the ground truth but fails to include the kidney mask.
Electronics 13 03526 g005
Table 1. Results of the segmentation metrics that the models scored for the bladder. Each score is followed by its corresponding standard deviation (±std).
Table 1. Results of the segmentation metrics that the models scored for the bladder. Each score is followed by its corresponding standard deviation (±std).
OrganModelsMetricsTrained on 15 Data20 Data30 Data40 Data50 Data
BladderEnsembleDice Score92.49 ± 3.1296 ± 4.9095.15 ± 2.4586.29 ± 2.2396.89 ± 2.21
Precision88.07 ± 5.110.9684 ± 6.500.9422 ± 4.920.7808 ± 4.250.9509 ± 4.01
Recall0.9812 ± 2.150.9535 ± 4.360.9659 ± 5.550.9958 ± 5.450.9887 ± 1.22
IoU0.8646 ± 5.180.9241 ± 3.360.9100 ± 3.670.7769 ± 6.340.9399 ± 3.01
2DDice Score50.26 ± 12.0246.43 ± 10.1048.28 ± 6.3531.35 ± 5.0230.76 ± 5.09
Precision35.63 ± 15.0132.19 ± 16.7833.77 ± 10.5019.31 ± 20.1118.61 ± 18.47
Recall94.48 ± 2.1994.18 ± 4.3489.81 ± 2.2599.04 ± 2.2596.99 ± 4.12
IoU34.97 ± 10.5431.60 ± 12.1032.41 ± 10.6419.26 ± 14.9218.49 ± 14.48
3DDice Score26.16 ± 5.1049.80 ± 17.5458.80 ± 18.0228.19 ± 13.0532.90 ± 17.06
Precision15.57 ± 5.5135.84 ± 15.9045.38 ± 10.3116.83 ± 5.3420.05 ± 8.90
Recall96.58 ± 4.8088.60 ± 7.8291.86 ± 5.6697.79 ± 4.3697.41 ± 5.97
IoU15.50 ± 7.8234.31 ± 10.1044.03 ± 15.3916.78 ± 9.2519.95 ± 9.50
DeeplabDice Score32.78 ± 10.2517.13 ± 6.0146.45 ± 17.5654.77 ± 13.7849.30 ± 16.34
Precision20.61 ± 5.409.61 ± 4.1634.87 ± 6.8538.88 ± 6.3235.98 ± 8.40
Recall91.41 ± 7.8189.08 ± 3.2081.55 ± 12.5841.14 ± 6.9897.41 ± 10.95
IoU20.09 ± 8.919.49 ± 4.2132.05 ± 7.9287.07 ± 7.5089.21 ± 11.35
SegResNetDice Score69.28 ± 14.9073.39 ± 15.3063.11 ± 18.0581.30 ± 10.2280.50 ± 13.13
Precision60.09 ± 10.5565.77 ± 11.1250.40 ± 15.2576.71 ± 10.2086.68 ± 12.40
Recall86.59 ± 8.4087.30 ± 10.9294.02 ± 12.0887.97 ± 7.7177.61 ± 4.55
IoU54.90 ± 12.6660.27 ± 12.6048.89 ± 15.0569.68 ± 11.9569.32 ± 9.38
Table 2. Results of the segmentation metrics that the models scored for the brain. Each score is followed by its corresponding standard deviation (±std).
Table 2. Results of the segmentation metrics that the models scored for the brain. Each score is followed by its corresponding standard deviation (±std).
OrganModelsMetricsTrained on 15 Data20 Data30 Data40 Data50 Data
BrainEnsembleDice Score93.70 ± 2.1294.70 ± 3.1596.53 ± 3.3497.09 ± 2.9097.01 ± 3.03
Precision89.09 ± 3.3293.65 ± 4.3895.36 ± 4.8595.83 ± 3.1196.55 ± 3.85
Recall99.08 ± 3.0596.50 ± 2.9597.88 ± 3.0998.52 ± 2.4797.67 ± 2.67
IoU88.31 ± 4.0190.22 ± 3.5493.33 ± 4.1694.40 ± 3.1694.27 ± 3.21
2DDice Score78.79 ± 5.1059.02 ± 2.8570.13 ± 7.2353.41 ± 3.6786.14 ± 7.31
Precision70.27 ± 7.9543.29 ± 1.8856.09 ± 3.6837.20 ± 1.1578.95 ± 4.44
Recall91.72 ± 5.1694.73 ± 7.3996.42 ± 5.2296.72 ± 6.1095.78 ± 3.08
IoU65.57 ± 9.1242.32 ± 8.2254.66 ± 8.2336.72 ± 4.5476.24 ± 4.45
3DDice Score34.29 ± 4.1137.59 ± 2.8738.16 ± 9.0148.68 ± 8.0348.99 ± 7.44
Precision21.57 ± 3.1224.39 ± 1.4624.76 ± 7.8034.01 ± 5.634.31 ± 7.05
Recall98.16 ± 9.7992.76 ± 9.5997.25 ± 10.0598.27 ± 11.6196.93 ± 9.90
IoU21.51 ± 3.6524.21 ± 2.7624.62 ± 7.4133.86 ± 7.9934.02 ± 8.31
DeeplabDice Score26.16 ± 5.4049.80 ± 10.1258.80 ± 7.0977.39 ± 25.8881.19 ± 8.67
Precision15.57 ± 4.3035.84 ± 8.9245.38 ± 6.4489.88 ± 23.0185.32 ± 12.08
Recall96.58 ± 8.3388.60 ± 12.3191.86 ± 10.0572.87 ± 19.3079.30 ± 5.02
IoU15.50 ± 4.0534.31 ± 7.3544.03 ± 5.5365.74 ± 15.7169.21 ± 7.10
SegResNetDice Score91.12 ± 10.2293.62 ± 4.1490.90 ± 4.0693.83 ± 3.1793.38 ± 4.81
Precision96.78 ± 15.3594.02 ± 7.8189.94 ± 7.6692.16 ± 8.3196.23 ± 8.10
Recall87.96 ± 8.5194.05 ± 3.2593.08 ± 2.1296.15 ± 3.2897.41 ± 3.54
IoU85.07 ± 10.5088.28 ± 6.0283.57 ± 7.0588.55 ± 5.9191.50 ± 5.19
Table 3. Results of the segmentation metrics that the models scored for the kidneys. Each score is followed by its corresponding standard deviation (±std).
Table 3. Results of the segmentation metrics that the models scored for the kidneys. Each score is followed by its corresponding standard deviation (±std).
OrganModelsMetricsTrained on 15 Data20 Data30 Data40 Data50 Data
KidneysEnsembleDice Score53.02 ± 2.2850.93 ± 1.4767.27 ± 2.9455.13 ± 3.5236.51 ± 3.92
Precision48.08 ± 3.7643.90 ± 1.8963.68 ± 3.4545.10 ± 3.1128.05 ± 4.02
Recall62.09 ± 4.9865.67 ± 3.1474.93 ± 4.1073.28 ± 4.8957.66 ± 3.09
IoU37.70 ± 2.7835.87 ± 2.8753.73 ± 5.2139.74 ± 3.2523.82 ± 3.07
2DDice Score35.24 ± 2.5639.64 ± 4.7535.23 ± 5.1936.16 ± 5.9516.81 ± 3.47
Precision24.49 ± 5.0129.96 ± 5.1023.98 ± 6.2423.12 ± 7.0209.38 ± 2.11
Recall67.77 ± 2.4063.67 ± 4.5173.69 ± 4.7287.54 ± 4.0385.89 ± 3.69
IoU21.74 ± 3.5825.18 ± 4.9321.70 ± 4.7122.17 ± 6.1209.22 ± 1.19
3DDice Score37.61 ± 9.7145.33 ± 13.2833.25 ± 5.3673.36 ± 7.8073.64 ± 5.41
Precision26.91 ± 8.6132.90 ± 7.6222.91 ± 3.2172.85 ± 8.5367.50 ± 7.02
Recall69.14 ± 2.2276.64 ± 13.4967.27 ± 7.8576.71 ± 7.7581.76 ± 4.91
IoU24.78 ± 4.4830.87 ± 7.4020.84 ± 3.8259.30 ± 12.0159.71 ± 5.06
DeeplabDice Score39.75 ± 15.9931.70 ± 17.3547.41 ± 20.1440.89 ± 15.7746.89 ± 20.00
Precision39.04 ± 14.6122.81 ± 11.1254.96 ± 19.6835.38 ± 10.3443.31 ± 18.60
Recall26.86 ± 10.1054.60 ± 18.3246.06 ± 20.1252.93 ± 17.9153.40 ± 21.42
IoU15.50 ± 9.9220.10 ± 10.0434.06 ± 18.3527.10 ± 9.0533.12 ± 17.12
SegResNetDice Score37.27 ± 12.6022.30 ± 17.1661.54 ± 10.0033.51 ± 7.5153.08 ± 5.10
Precision30.91 ± 11.9114.97 ± 15.0164.38 ± 12.9221.51 ± 5.9838.57 ± 5.11
Recall48.90 ± 13.7563.38 ± 19.3360.40 ± 10.5678.20 ± 10.7786.03 ± 7.80
IoU23.56 ± 10.0413.61 ± 9.6545.16 ± 9.7220.36 ± 6.1236.29 ± 4.11
Table 4. Results of the segmentation metrics that the models scored for the left ventricle of the heart. Each score is followed by its corresponding standard deviation (±std).
Table 4. Results of the segmentation metrics that the models scored for the left ventricle of the heart. Each score is followed by its corresponding standard deviation (±std).
OrganModelsMetricsTrained on 15 Data20 Data30 Data40 Data50 Data
Left Ventricle
of the Heart
EnsembleDice Score76.77  ± 13.5371.63 ± 18.6582.33 ± 20.0767.43 ± 14.5472.05 ± 13.19
Precision79.12 ± 14.1269.78 ± 15.3184.18 ± 22.3054.86 ± 15.5166.57 ± 13.01
Recall81.65 ± 10.1081.75 ± 19.4584.33 ± 18.4059.55 ± 13.6686.31 ± 15.21
IoU63.09 ± 14.5457.04 ± 14.0970.99 ± 19.3690.17 ± 16.7159.82 ± 12.53
2DDice Score22.53 ± 10.4325.03 ± 10.1236.96 ± 2.8128.98 ± 14.8134.88 ± 9.64
Precision15.24 ± 4.1116.58 ± 5.5027.12 ± 4.0119.05 ± 10.1224.64 ± 7.31
Recall45.61 ± 14.3370.01 ± 15.9971.64 ± 2.0283.90 ± 15.1181.88 ± 12.30
IoU12.76 ± 4.7214.50 ± 5.0222.96 ± 3.3017.83 ± 5.3022.30 ± 5.11
3DDice Score35.26 ± 13.2622.36 ± 10.6529.02 ± 4.1824.85 ± 11.2329.47 ± 13.42
Precision23.17 ± 12.6214.11 ± 8.3418.19 ± 3.8515.57 ± 9.6820.71 ± 11.64
Recall78.84 ± 14.8562.32 ± 12.4578.26 ± 7.8476.06 ± 13.2466.80 ± 14.67
IoU21.53 ± 11.4613.02 ± 7.4317.08 ± 4.4815.03 ± 8.5418.31 ± 10.62
DeeplabDice Score21.89 ± 12.3229.03 ± 11.3234.66 ± 9.8918.33 ± 10.8444.68 ± 15.25
Precision19.15 ± 9.4723.06 ± 10.4335.66 ± 8.5712.40 ± 8.4659.95 ± 17.21
Recall38.92 ± 13.0644.74 ± 14.0538.57 ± 10.3016.16 ± 11.0137.72 ± 12.12
IoU14.09 ± 9.4819.22 ± 10.3024.20 ± 9.5437.27 ± 9.3032.72 ± 11.81
SegResNetDice Score61.50 ± 25.6030.94 ± 13.3577.11 ± 2.4656.96 ± 19.4068.40 ± 15.84
Precision61.72 ± 22.6119.58 ± 10.3678.82 ± 5.1044.37 ± 15.5755.62 ± 10.11
Recall64.89 ± 26.0681.41 ± 15.0377.06 ± 1.9295.16 ± 20.8495.10 ± 18.12
IoU49.23 ± 20.2419.07 ± 9.8464.35 ± 2.1142.43 ± 19.8553.81 ± 11.02
Table 5. Results of the segmentation metrics that the models scored for the liver. Each score is followed by its corresponding standard deviation (±std).
Table 5. Results of the segmentation metrics that the models scored for the liver. Each score is followed by its corresponding standard deviation (±std).
OrganModelsMetricsTrained on 15 Data20 Data30 Data40 Data50 Data
LiverEnsembleDice Score83.10 ±  2.5485.27  ± 3.0888.11  ±  3.6888.42  ±  3.8688.46 ± 3.46
Precision76.83 ± 4.5483.41 ± 4.0485.83 ± 3.5686.15 ± 2.8788.76 ± 3.78
Recall92.29 ± 2.6088.91 ± 3.0091.23 ± 3.5491.34 ± 4.0588.98 ± 3.41
IoU71.64 ± 3.5075 ± 3.0478.93 ± 3.5479.41 ± 2.0879.55 ± 2.01
2DDice Score67.58 ±  5.4053.49 ±  5.5481.18 ± 3.4570.81 ±  3.5468.88 ± 4.85
Precision55.61 ± 4.5141.56 ± 4.6176.52 ± 2.8860.37 ± 2.8162.80 ± 4.16
Recall89.73 ± 6.7380.57 ± 7.1187.35 ± 4.7189.59 ± 6.0682.23 ± 6.54
IoU51.17 ± 4.0237.55 ± 4.5868.61 ± 3.9255.88 ± 4.2054.31 ± 3.93
3DDice Score57.54 ±  7.4257.15 ± 5.1071.99 ± 5.0380.89 ±  3.9174.02 ± 30.01
Precision46.59 ± 5.1244.47 ± 5.0364.06 ± 4.8582.44 ± 3.0567.27 ± 25.01
Recall79.92 ± 9.0184.37 ± 7.0684.47 ± 7.1280.07 ± 4.2685.87 ± 30.62
IoU41.09 ± 5.8240.65 ± 4.4056.72 ± 5.0168.31 ± 3.1959.50 ± 28.30
DeeplabDice Score41.57 ± 20.0148.86 ± 15.5750.52 ± 14.1952.85 ± 20.5855.94 ± 15.95
Precision31.23 ± 15.0047.78 ± 14.0451.18 ± 14.2249.23 ± 19.5855.73 ± 15.21
Recall74.20 ± 22.1855.10 ± 16.9353.10 ± 15.0361.04 ± 22.4158.03 ± 15.99
IoU28.71 ± 15.3736.08 ± 14.0237.53 ± 13.8740.57 ± 18.6244.08 ± 13.83
SegResNetDice Score75.36 ± 15.7881.91 ± 4.9187.23 ± 2.8986.43 ± 2.1766.04 ± 10.02
Precision75.43 ± 14.3990.66 ± 7.9088.03 ± 1.9081.98 ± 3.4752.29 ± 8.30
Recall78.20 ± 16.0975.13 ± 10.1586.55 ± 2.0591.63 ± 4.3293.06 ± 12.02
IoU63.19 ± 13.4569.56 ± 5.3177.47 ± 5.1276.21 ± 5.9550.15 ± 9.76
Table 6. The table presents the results of the paired t-test comparing the Dice scores of each different model against the equivalent ensemble model. The p-values are listed below the columns that indicate the number of training data used. Statistically significant p-values (less than 0.05) are highlighted in bold.
Table 6. The table presents the results of the paired t-test comparing the Dice scores of each different model against the equivalent ensemble model. The p-values are listed below the columns that indicate the number of training data used. Statistically significant p-values (less than 0.05) are highlighted in bold.
OrganModelTrained On 15 Data20 Data30 Data40 Data50 Data
Liver2D1.3 × 10−51.0 × 10−68 × 10−43.5 × 10−54.5 × 10−5
3D1.4 × 10−52.1 × 10−72.3 × 10−60.014.9 × 10−7
Deeplab1.5 × 10−40.0010.00080.0021.02 × 10−5
SegResNet0.050.00070.020.00055.4 × 10−5
Bladder2D1.7 × 10−58.4 × 10−73.8 × 10−81.1 × 10−101.2 × 10−9
3D1.4 × 10−90.00053.6 × 10−56 × 10−60.0001
Deeplab4.1 × 10−86.5 × 10−92.7 × 10−52.8 × 10−52.8 × 10−5
SegResNet0.0010.0040.00090.0060.01
Left Ventricle
of the Heart
2D2.3 × 10−81.4 × 10−56.7 × 10−85 × 10−74.5 × 10−7
3D3.5 × 10−70.0058.8 × 10−93.1 × 10−81.5 × 10−8
Deeplab4.3 × 10−50.0050.0021 × 10−50.008
SegResNet0.070.00070.400.00090.27
Brain2D5.4 × 10−93.8 × 10−103.3 × 10−81.4 × 10−93.6 × 10−6
3D4 × 10−97.9 × 10−121.8 × 10−86.8 × 10−91.4 × 10−9
Deeplab5.3 × 10−11 × 10−81.6 × 10−100.010.0002
SegResNet0.10.40.0030.0010.08
Kidneys2D1.2 × 10−64.3 × 10−54.6 × 10−54.2 × 10−91.2 × 10−7
3D0.0010.080.21.2 × 10−76.4 × 10−8
Deeplab0.010.0010.190.030.8
SegResNet0.0010.00010.27.6 × 10−50.001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Vezakis, A.; Vezakis, I.; Vagenas, T.P.; Kakkos, I.; Matsopoulos, G.K. A Multidimensional Framework Incorporating 2D U-Net and 3D Attention U-Net for the Segmentation of Organs from 3D Fluorodeoxyglucose-Positron Emission Tomography Images. Electronics 2024, 13, 3526. https://doi.org/10.3390/electronics13173526

AMA Style

Vezakis A, Vezakis I, Vagenas TP, Kakkos I, Matsopoulos GK. A Multidimensional Framework Incorporating 2D U-Net and 3D Attention U-Net for the Segmentation of Organs from 3D Fluorodeoxyglucose-Positron Emission Tomography Images. Electronics. 2024; 13(17):3526. https://doi.org/10.3390/electronics13173526

Chicago/Turabian Style

Vezakis, Andreas, Ioannis Vezakis, Theodoros P. Vagenas, Ioannis Kakkos, and George K. Matsopoulos. 2024. "A Multidimensional Framework Incorporating 2D U-Net and 3D Attention U-Net for the Segmentation of Organs from 3D Fluorodeoxyglucose-Positron Emission Tomography Images" Electronics 13, no. 17: 3526. https://doi.org/10.3390/electronics13173526

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop