Next Article in Journal
A 5G NR FR2 Beamforming System with Integrated Transceiver Module
Next Article in Special Issue
Investigation of Appropriate Scaling of Networks and Images for Convolutional Neural Network-Based Nerve Detection in Ultrasound-Guided Nerve Blocks
Previous Article in Journal
Damage Detection in Glass Fibre Composites Using Cointegrated Hyperspectral Images
Previous Article in Special Issue
Intraoperative Angle Measurement of Anatomical Structures: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Analysis of Loss Functions for Heavily Imbalanced Lesion Segmentation

1
Brain and Mind Centre, The University of Sydney, Camperdown, NSW 2050, Australia
2
Faculty of Science, Yamagata University, Yamagata 990-8560, Japan
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(6), 1981; https://doi.org/10.3390/s24061981
Submission received: 11 January 2024 / Revised: 13 March 2024 / Accepted: 15 March 2024 / Published: 20 March 2024
(This article belongs to the Collection Biomedical Imaging and Sensing)

Abstract

:
Heavily imbalanced datasets are common in lesion segmentation. Specifically, the lesions usually comprise less than 5% of the whole image volume when dealing with brain MRI. A common solution when training with a limited dataset is the use of specific loss functions that rebalance the effect of background and foreground voxels. These approaches are usually evaluated running a single cross-validation split without taking into account other possible random aspects that might affect the true improvement of the final metric (i.e., random weight initialisation or random shuffling). Furthermore, the evolution of the effect of the loss on the heavily imbalanced class is usually not analysed during the training phase. In this work, we present an analysis of different common loss metrics during training on public datasets dealing with brain lesion segmentation in heavy imbalanced datasets. In order to limit the effect of hyperparameter tuning and architecture, we chose a 3D Unet architecture due to its ability to provide good performance on different segmentation applications. We evaluated this framework on two public datasets and we observed that weighted losses have a similar performance on average, even though heavily weighting the gradient of the foreground class gives better performance in terms of true positive segmentation.

1. Introduction

Magnetic resonance imaging (MRI) is one of the most common techniques using sensors to obtain imaging information from the brain. Furthermore, lesion segmentation is the most common image analysis task applied to brain MRI for different neurological pathologies, such as multiple sclerosis [1,2], Alzheimer’s disease [3] or vascular cognitive impairment [4]. Brain lesions, usually identified as white matter hyperintensities on T2-weighted images, represent a small percentage of the whole brain volume, can have different shapes and volumes and their spatial distribution can greatly vary between patients. Due to these issues, the accurate detection and segmentation of white matter hyperintensities is a challenging segmentation task.
Recently, deep learning approaches have rapidly become the state of the art for medical image analysis [5,6], including binary image segmentation [7]. However, a severe data imbalance can bias the training of the models and produce unexpected results (i.e., reduce the accuracy of the lesion segmentation) [8,9]. The most common solutions are the use of sampling techniques to increase the number of positive samples [10] and loss functions that give different weights to the foreground and background classes, either implicitly [11,12] or explicitly [13]. While different losses have been presented for medical imaging [14], they are rarely analysed for datasets with heavy imbalances. Specifically, their effect on the underrepresented class is not properly understood, how the loss evolves through time per epoch is rarely studied, and the effect of randomness is usually excluded from the results, even though weight optimisation is inherently a stochastic process. To that end, multiple studies have been published on understanding prediction uncertainty through probabilistic frameworks and deep ensembles [15,16,17,18], suggesting the importance of reporting results with multiple runs and random seeds. Furthermore, losses are usually defined based on their value, but they are rarely designed taking into account their derivative and their effect on optimisation.
In this paper, we present an analysis of the most common losses used for imbalanced datasets, and we study their evolution during training focusing on the effect on the underrepresented class. Furthermore, we propose a new weighted loss function based on the gradient of the cross-entropy loss (as opposed to focusing only on the loss function alone). The rest of the paper is structured as follows: in Section 2, we present the framework and public datasets we used, followed by the results in Section 3, a short discussion in Section 4 and conclusions and future works in Section 5.

2. Materials and Methods

2.1. Public Datasets

For this analysis, we used two public datasets with manual annotations of brain lesions and a heavy imbalance. The statistics of the number of foreground and background voxels are summarised in Table 1 and examples of images are given in Figure 1. While the ratio between foreground and background voxels is bigger inside the training patches, they still represent less than 5% of the whole patch on average.

2.1.1. White Matter Hyperintensities (WMH) Challenge 2017 [4]

The WMH Segmentation Challenge was held during the MICCAI 2017 conference, and it provided a standardized assessment of automatic methods for the segmentation of WMH. The task for the challenge was defined as the segmentation of WMH of presumed vascular origin on brain MR images. A total of 60 labelled images were acquired from three different scanners with different sensors from three different vendors, in three different institutes. For each subject, a 3D T1-weighted and a 2D multi-slice FLAIR image were provided. Details on the acquisition and preprocessing steps for these images are summarised in the original challenge paper [4]. In brief, all images were manually labelled for white matter hyperintensities by an expert, and these annotations were afterwards reviewed by a second expert who created the final mask. Other pathological regions were also labelled. However, these regions were not evaluated during the challenge and, thus, we ignored this label for our analysis. Finally, T1-weighted images were co-registered to the FLAIR space during pre-processing, and we used these pre-processed images for our experiments.
The parameters of each acquisition are summarised as follows:
  • UMC Utrecht, 3 T Philips Achieva: 3D T1-weighted sequence (192 slices, slice size: 256 × 256, voxel size: 1.00 × 1.00 × 1.00 mm3, repetition time (TR)/echo time (TE): 7.9/4.5 ms), 2D FLAIR sequence (48 slices, slice size: 240 × 240, voxel size: 0.96 × 0.95 × 3.00 mm3, TR/TE/inversion time (TI): 11,000/125/2800 ms).
  • NUHS Singapore, 3 T Siemens TrioTim: 3D T1-weighted sequence (192 slices, slice size: 256 × 256, voxel size: 1.00 × 1.00 × 1.00 mm3, TR/TE/TI: 2, 300/1.9/900 ms), 2D FLAIR sequence (48 slices, slice size: 256 × 256, voxel size: 1.00 × 1.00 × 3.00 mm3, TR/TE/TI: 9000/82/2500 ms).
  • VU Amsterdam, 3 T GE Signa HDxt: 3D T1-weighted sequence (176 slices, slice size: 256 × 256, voxel size: 0.94 × 0.94 × 1.00   mm 3 , TR/TE: 7.8/3.0 ms), 3D FLAIR sequence (132 slices, slice size: 83 × 256, voxel size: 0.98 × 0.98 × 1.20   mm 3 , TR/TE/TI: 8000/126/2340 ms).

2.1.2. Ljubljana Longitudinal Multiple Sclerosis Lesion Dataset [1]

This database contains baseline and follow-up MR images of 20 multiple sclerosis (MS) patients. The images were acquired on a 1.5 T Philips MRI machine at the University Medical Centre Ljubljana (UMCL), and the data were anonymized [1]. Each patient’s MR acquisition contained a 2D T1-weighted (spin echo sequence, repetition time (TR) = 600 ms, echo time (TE) = 15 ms, flip angle (FA) = 90°, sampling of 0.9 × 0.9 × 3 mm with no inter-slice gap resulting in a 256 × 256 × 45 lattice), a 2D T2-weighted (spin echo sequence, TR = 4500 ms, TE = 100 ms, FA = 90°, sampling of 0.45 × 0.45 × 3 mm with no inter-slice gap, resulting in a 512 × 512 × 45 lattice), and a 2D FLAIR image (TR = 11,000, TE = 140, TI = 2800, FA = 90, sampling of 0.9 × 0.9 × 3 mm with no inter-slice gap, resulting in a 256 × 256 × 49 lattice). The median time between the baseline and follow-up studies was 311 days, ranging from 81 to 723 days with the interquartile range (IQR) of 223 days. For our analysis, we only focused on the FLAIR images that were skull stripped and bias corrected using N4, followed by coregistration to the follow-up space using rigid registration. To create a ground truth mask, an image neuroanalyst expert at our group labelled all positive activity (new and enlarging lesions) using the coregistered FLAIR images and their subtraction. Furthermore, 3 of the subjects were excluded due to a lack of positive activity after the labelling. Therefore, only 17 patients from the original dataset were used for the experiments.

2.1.3. Data Preparation

For both datasets we extracted patches of size 32 × 32 × 32 using a sliding window (with an overlap of 16 voxels) on the brain bounding box to minimise the number of voxels that did not belong to the brain. These patches were then sampled to decrease the imbalance between foreground and background voxels. Only patches containing lesion voxels were used for training in groups of 16.

2.2. Network Architecture

In the last few years, the 3D Unet architecture [19] has become the state of the art in volumetric medical image segmentation for different tasks and images [7]. In order to train all the losses with a robust architecture, we decided to use a simple 3D Unet of 2,906,913 parameters using basic convolutional blocks (with 32, 64, 128 and 256 kernels of size 3 × 3 × 3 ), followed by ReLU activation, group norm (with groups of 8 channels) and max pooling of size 2 (see Figure 2). These hyperparameters were chosen empirically based on the best performance of our previous works on other private datasets for lesion segmentation [20,21]. To guarantee a fair comparison, these hyperparameters were also shared for all the experiments (different seeds and loss functions). The weights were randomly initialised using the same seed for all the losses in a given experiment to reduce the effect of random initialisation. This seed was also used to ensure that the training batches per epoch were shuffled the same way for all the losses.

2.3. Loss Functions

2.3.1. Cross-Entropy

The cross-entropy loss (xent) is one of the most commonly used loss functions for deep neural networks. If we consider the distribution of the real labels ( y i ) and the predicted ones ( y i ^ ), the cross-entropy expresses the similarity of these two distributions in the following form:
L C E ( y ^ i , y i ) = log y ^ i , if foreground log ( 1 y ^ i ) , if background
In our case, this equation represents the loss for a single voxel, and the final value is computed as the mean of the loss for all the voxels. By definition, all voxels have an equal weight; therefore, during training, the optimiser will tend to favour the majority class.

2.3.2. Focal Loss

The focal loss function was presented by Tsung-Yi Lin et al. [13] for dense object detection in the presence of an imbalance between positive and negative samples. Through the use of a weighting variable ( α ) and a modulating factor ( γ ), they extended the cross-entropy loss to increase the importance of positive samples and reduce the effect of samples correctly classified with a high confidence. Following the notation for the cross-entropy loss, the focal loss for a prediction y i ^ can be defined as:
L F L ( y ^ i , y i ) = α ( 1 y i ^ ) γ log y i ^ , if foreground ( 1 α ) y i ^ γ log ( 1 y i ^ ) , if background
For our experiments, we used γ = 2 as suggested on the paper, and two different α values: α 1 = 0.25 (focal1) and α 2 = 0.75 (focal2). On the original paper, they suggest to decrease α as γ is increased, and they noted that α = 0.25 gave the best results with γ = 2 . However, we also wanted to test the loss with a higher α value to give a higher weight to positive samples.

2.3.3. Generalised Dice Loss

The generalised Dice loss was presented by Sudre et al. [12] to balance the effect of positive and negative samples, based on a generalisation of the Dice similarity metric (DSC) [22]. This loss function can be defined as:
L g D S C ( Y ^ , Y ) = 1 2 · ( w f g + w b g ) · i N y ^ i · y i + w b g · ( N i N ( y ^ i + y i ) ( w f g w b g ) · i N ( y ^ i + y i ) + 2 · N w b g ,
where Y ^ and Y represent the predicted and true segmentation for the N image voxels ( y ^ i and y i ), and w b g and w f g are weights for the background and foreground class, respectively.
In the special case where w b g = 0 , the generalised Dice loss becomes the commonly used binary Dice loss originally presented by Milletari et al. [11]:
L D S C ( Y ^ , Y ) = 1 2 · i N y ^ i · y i + ϵ i N ( y ^ i + y i ) + ϵ ,
where ϵ is a small smoothing constant to avoid numerical issues when dividing by 0. In our experiments, we used the generalised loss (gdsc) as defined on the original paper ( w b g = 1 / N b g 2 and w f g = 1 / N f g 2 , where N b g and N f g refer to the number of background and foreground voxels), the binary Dice loss (dsc) ( w b g = 0 and w f g = 1 ) and a combination of the binary Dice loss and cross-entropy with equal weights (mixed).

2.3.4. Weighted Gradient Loss

Similarly to the focal loss, we propose to introduce a weighting factor and a modular factor to the cross-entropy loss. However, instead of applying that factor on the loss function itself, we propose to apply it to the gradient of the cross-entropy when combined with the gradient of the sigmoid function as follows:
L w C E ( y ^ i , y i ) f i = α ( 1 y ^ i ) γ , if foreground ( 1 α ) y ^ i γ , if background
where f i represents the output of the last layer before applying the final sigmoid activation. In our experiments, we set γ = 2 and α = N b g / N (new).

2.4. Experimental Design

In order to counter the effect on random initialisation and shuffling, we developed a framework to train the same network with the same initialisation and batches by setting the same random seed for each loss for a given experiment. With that setup, we run 5 different 5-fold cross-validation experiments with the same training/testing splits. We then evaluate the models at the end of each epoch for a maximum of 25 epochs by analysing the Dice similarity coefficient ( D S C = 2 T P 2 T P + F N + F P ); sensitivity, also known as true positive fraction or T P f for short ( T P f = T P T P + F N ); and precision ( P = T P T P + F P ). We compute all three metrics for the training patches, and the training and testing image segmentations reconstructed from patches. In brief, patches of 32 × 32 × 32 are uniformly sampled for each image with a sliding window of 16 × 16 × 16, and the final prediction for each voxel is averaged between all the overlapping results. For the binary metrics, we first binarise the predicted output of the network (with a threshold of 0.5). The goal of this setup is to analyse whether the difference in different executions are actually due to the training variance caused by randomisation or the losses themselves.

2.5. Implementation Details

The framework was implemented using Python 3.6.9 and pytorch version 1.5.0 and the code is publicly available at https://github.com/marianocabezas/rethinking_dsc (accessed on 1 January 2024). The whole analysis was run on a NVIDIA DGX-1 server with 80 CPU cores, 504 GB of RAM and two Tesla V100-SXM2.

3. Results

To analyse the performance of the different trained models for each epoch and seed, we used the T P f and precision to illustrate the balance between over- and undersegmentation and the D S C to highlight the overall segmentation. Ideally, we would like to maximise all metrics, but in practise, increasing the T P f also increases the false positives, affecting the precision and D S C values. Table 2 summarises the best performance when the best epoch and random seed are known, while Figure 3 illustrates the curves for all folds and seeds for each dataset in terms of each of the three different metrics we analysed (plots for each fold are detailed in Figure 4 and Figure 5). The curves are separated by the type of data being evaluated (input patches and whole image segmentations reconstructed from patch segmentations) and represent a summarised version of all the metrics for each different training seed setup. The mean metric value is represented with a solid coloured line, while error bands around that line are bounded by the minimum and maximum value per epoch for all the seeds. By using upper and lower bounds, we can represent stochastic variability that could lead to different conclusions if different seeds are used for each method. Furthermore, the error bands also illustrate how robust the models can be to the problems represented by each dataset.
Observing the table and curves, there is a clear discrepancy between the performance on the training patches and the overall segmentation for the whole image. This difference is emphasised on the Dice-based losses. Furthermore, it is also clear that weighting the foreground class, either explicitly (focal and weighted loss) or implicitly (Dice based), improves the performance in terms of the final D S C metric. However, the larger the weight, the better the results. Furthermore, a high weight (higher than 0.9 for the weighted loss) has a direct impact on the T P f measure throughout the whole training process.
A closer analysis of the curves highlights the trade-off between the p r e c i s i o n and T P f segmentation measures (also illustrated in the qualitative examples from Figure 4 and Figure 5). As the training progresses, the T P f metric decreases slowly, while the p r e c i s i o n improves (leading to an overall slightly better D S C value). Furthermore, if we observe the bands for the losses, there is a clear overlap between all of them for the D S C metric. This suggests that there is a large variability in the model performance depending on the chosen random seed and fold. In general, the observed training pattern is similar for the weighted losses (Dice based, focal and weighted) with an initial higher peak in terms of T P f and a slower decrease for the losses with a higher weight for the foreground class. While the behaviour is similar for both datasets, there is a bigger difference between losses for the longitudinal dataset, which also presents a higher imbalance between classes. In fact, the variability in the LIT dataset is also larger (wider error bands), leading to a larger overlap and an increased difficulty to separate between methods. Curves for the WMH dataset, on the other hand, have thinner error bands, and there is a clear separation between our proposed loss function and the rest of methods for the T P f and p r e c i s i o n metrics on the first few epochs, even though the D S C values remain similar for the training and testing images. Nonetheless, the curves converge to similar results as the number of epochs trained increases. Finally, the curves on Figure 4 and Figure 5 show the same trends observed on the general dataset curves, albeit with reduced error bands (fewer samples are analysed per fold). Analysing folds carefully (each row on the figures), the LIT shows larger differences between folds. Specifically, the first row shows overall higher T P f and D S C values for all methods when compared to the other folds.
To complement the analysis of the curves, qualitative results for epoch 25 (the last epoch) are provided in Figure 6 and Figure 7 for the LIT and WMH dataset, respectively. While there seems to be a general agreement in terms of true positives (green colour), most losses and examples show an increase in false negatives (blue colour) when compared to the new proposal (as also evidenced by the numerical results). In particular, small lesions tend to be missed by losses that do not heavily weight positive predictions, with cross-entropy having the lowest detection. This is contrasted by a decrease in the number of false positives (red colour), also evidenced by the p r e c i s i o n metric on the numerical results. From the Dice-based losses, mixed has the overall best results, with the dsc loss having a large number of false positives (especially on the LIT dataset). Comparing the two focal losses, the results are fairly similar with small differences in detection depending on the case, suggesting that the value of α might not be as important as expected.
Finally, we summarise the D S C , T P f and p r e c i s i o n metrics on the training and testing images using a violin plot in Figure 8 to analyse the distribution of these metrics for each loss according to the random seeds. Once again, we can observe that the best performance (highest bound) is obtained with the heavily weighted gradient loss (new), even though the plots present a similar shape for D S C in both datasets for most of the losses and a large variability in general. However, we can also observe some important differences on the T P f distribution for both datasets. On the longitudinal dataset, the focal losses and the cross-entropy present a larger concentration of lower T P f values when compared to the other losses, while only the lightly weighted focal loss (focal1) and cross-entropy present a lower performance on the WMH dataset. These two losses also present the worst overall performance on the longitudinal dataset in terms of D S C (as illustrated by a longer and spread plot, concentrated on the lower bound).

4. Discussion

4.1. The Effect of Confident Errors on the Loss Function

The Dice loss is one of the most commonly used techniques to address class imbalance due to its application as a metric for segmentation [23]. Furthermore, it is usually believed that it is one of the best choices for heavily imbalanced datasets. However, the original metric was not developed for optimisation and was designed to address a set theory problems (specifically, to measure the overlap between two sets). As a consequence, there are certain unexpected side effects when pairing it with common activations for segmentation (sigmoid and softmax). One of these unexpected side effects is the effect of confidently wrong predictions (for a mathematical analysis through derivation, we refer the reader to Appendix A). This effect is evidenced by the metric curves for the Dice loss where we found the largest discrepancy between patch-based metrics (the ones during training) and image-based metrics (the final desired output). On the first step, the Dice loss favours the prediction of positive samples, leading to a high true T P f at the cost of a large number of false positive predictions (as evidenced by the low DSC value). Some of these erroneous predictions maintain a high confidence score throughout the whole process, leading to false positive predictions that are never corrected even with a large number of epochs. These findings align with another recent study where the effect of different loss functions on the logits of the network were analysed [9] and extends to the other two related losses (mixed and gdsc).
If we zoom out and analyse the other losses we explored in this study, we can observe that they can be easily derived from the cross-entropy loss. This loss in particular is ubiquitous in classification studies (including segmentation defined as pixel classification) due to its simplicity and its relationship to logistic regression. As opposed to the Dice loss, this function was derived to interact with sigmoid (and softmax in problems with multiple labels), leading to a linear gradient. In that sense, if we observe the curves for these losses (xent, focal1, focal2 and new), we can observe a smoother transition and increasing tendency when it comes to sensitivity over number of epochs and a higher mean precision. This suggests that false confident predictions are improved over time, as the weight updates are mostly linear with respect to the error on the prediction. Furthermore, the weights on the target class (lesions) are the most important trade-off between sensitivity and precision as observed in the differences between the focal1 and focal2 losses. Finally, even though cross-entropy is the most basic function, it is also highly reliable and can obtain similar results, even though it might have a higher number of epochs.
Overall, the positive detection of lesions voxels remains high after 25 epochs as evidenced by the qualitative examples in Figure 6 and Figure 7. However, small differences in false positives and negatives are observed depending on the loss, with a common trade-off between them.

4.2. The Effect of Randomness during Training

To understand the effect of randomness during training, we use five random sets that affect the initial value of the layer’s weights and the shuffling of the patches used for training at each epoch. To illustrate this effect, we plot the results of each epoch with a set of bands with the minimum, maximum and mean values over all seeds for each method in Section 3. One of the first conclusions that can be observed is the large overlap between all methods and metrics for image-wise results. As expected, this effect is made stronger on the testing images that were not seen during the training process. In that sense, while the mean curves can be easily distinguished between methods, it is also obvious that changing the random seed could lead to different conclusions if the lowest bound for the “baseline” methods is selected and the highest bound “seed” is used for a method we would like to highlight over the others. Furthermore, datasets with a low percentage of foreground voxels are more sensitive to that issue as exemplified by the curves on the LIT dataset shown in Figure 4.

4.3. The Discrepancy between Patch-Based Results and Image-Based Results

The most common way to analyse how the training process evolves is to monitor the training and validation losses. In the scenario where a model is trained on patches (due to memory constraints), these two curves represent the ability of the model to segment patches. Even though such measures are generally considered a good proxy for the real end goal (segmenting lesions on the whole brain), our experiments suggest that these values are not only over-optimistic but also highly misleading (especially when comparing methods). If we look at the patch-based results for both datasets, all methods have high D S C values over 0.7 (considered a good value for lesions) [24], with some losses reaching a value close to 1 (a perfect score). While this is not surprising because the model is trained with these same patches, the same metric drops drastically for the LIT dataset (the one with the lowest percentage of lesion voxels) and less so for the WMH dataset. Furthermore, our proposed loss function that obtains the lowest D S C on the training patches by a large margin obtains comparable results when evaluated on the image segmentation results reconstructed from patches as shown in Figure 3.
On a related note, while it might seem counterintuitive that the sensitivity is also affected when comparing results between patches and images, this phenomenon might be caused by the reconstruction of the final results through averaging. In particular, boundary lesion voxels might have a decreased score when considering different predictions for the same voxel in different relative positions within the patch. In fact, previous studies on CNNs have proven that while convolutions are inherently shift-invariant, pooling and padding can lead to encoding positional information during training [25,26]. In that sense, both reconstructing image segmentations from patches or performing inference directly on the whole image would lead to different results when comparing metrics on images and patches. Moreover, if overlapping patches are used during training, the loss metric might be unrealistic image-wise, as some voxels would be counted as independent occurrences more than once.
To conclude, these results suggest a few possible explanations that have also been corroborated by previous studies focusing on the D S C metric and its role as a measure of overlap that provides a balanced overview in terms of true and false predictions. On one hand, it is well known that the D S C metric is sensitive to the size of the object being analysed [23,27]. In our results, this is reflected by the difference between patch-wise and image-wise results but also by the fact that LIT results are a decimal point lower than those obtained with the WMH dataset. On the other hand, patch-wise results focus exclusively on patches with both foreground and background voxels (as background-only patches would not contribute to improve lesion segmentation and could be detrimental). In that sense, our analysis further emphasises the importance of not relying on loss values alone when evaluating model performance and to validate with metrics that are as close as possible to the desired end goal.

5. Conclusions

In this paper, we presented an evaluation of a set of common losses used for binary lesion segmentation on heavily imbalanced scenarios. We observed that training with sampled patches from images gives an unrealistic measure of the final model when applied to whole images. In other words, while some losses might provide a better performance on the training patches, that improvement is not reflected on the training images as a whole, due to the bias introduced during sampling. Furthermore, we have observed that random initialisation and shuffling can cause a large variance on the performance metrics, giving a range of different conclusions depending on which random seed and stopping epoch we take to evaluate the losses. However, most of the analysed losses have similarly optimal performance results (if the best stopping epoch and seed are known), even though they show clearly different trends throughout the training process depending on what the loss prioritises (e.g., an increase in positive predictions for our proposal). To conclude, heavily weighting the gradient of the foreground class gives the best results in terms of true positives, while still being able to obtain a competitive DSC metric on average, with an acceptable number of false positives.

Author Contributions

Conceptualisation, M.C. and Y.D.; methodology, M.C. and Y.D.; software, M.C. and Y.D.; validation, M.C. and Y.D.; formal analysis, M.C. and Y.D.; investigation, M.C. and Y.D.; resources, M.C. and Y.D.; data curation, M.C. and Y.D.; writing—original draft preparation, M.C. and Y.D.; writing—review and editing, M.C. and Y.D.; visualisation, M.C. and Y.D.; supervision, M.C. and Y.D.; project administration, M.C. and Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to using anonymised publicly available data.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data for this study are publicly available on the websites of the challenge organisers. The code used for this study is also publicly available at https://github.com/marianocabezas/rethinking_dsc (accessed on 1 January of 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Mathematical Analysis of the Dice Loss

Appendix A.1. The Dice Similarity Coefficient for Binary Masks

The Dice similarity coefficient was presented by Dice in 1945 to quantify the association between ecological species [22]. Generally speaking, this metric evaluates the overlap between two binary sets. Formally, given a set A { 0 , 1 } N and a set B { 0 , 1 } N , this coefficient is computed with the following equation:
D S C = 2 · | A B | | A | + | B | ,
where N is the number of elements in both sets. By definition, this metric is bounded between 0 and 1, with 1 representing perfect overlap and 0 no overlap between the sets. Furthermore, this metric is undefined when | A | = | B | = 0 .
This metric has become the de facto standard in medical imaging to evaluate the automatic segmentation of a tissue, lesion or structure given a previous manual or semiautomatic annotation considering the ground truth. In that sense, the binary masks from the automatic and manual segmentation can be interpreted as two sets over the voxels of the image, where 1 represents the presence of a tissue and 0 the lack of it.

Appendix A.2. The Probabilistic Dice Function

Given the previous definition of the DSC metric and the domain of A and B, Equation (A1) is equivalent to the following one:
p D S C = 2 · i N m i n ( a i , b i ) i N ( a i + b i ) ,
where A = { a 0 , , a n } and B = { b 0 , , b n } and a i , b i { 0 , 1 } . This new interpretation of the DSC metric can be used as an extension to fuzzy sets where now a i , b i 0 , 1 . In our case, only A is a fuzzy domain, and we can consider its values as probabilities that represent the confidence of the binary segmentation. Thus, we can use the following equation which is equivalent to (A2):
p D S C = 2 · i N a i · b i i N ( a i + b i ) .
This metric is still bounded between 0 and 1, with 1 still representing perfect overlap and 0 representing no overlap. We can easily generalise that metric to a multi-class generalised coefficient (gDSC) following Crum et al.’s work [28] as follows:
g D S C = 2 · k K w k · i N a i · b i k K w k · i N ( a i + b i ) ,
where K is the total number of classes and w k is the weight for class k. Furthermore, if we consider binary segmentation to be a two-class segmentation problem, we can simplify that equation to:
g D S C = 2 · ( w 1 + w 0 ) · i N a i · b i + w 0 · ( N i N ( a i + b i ) ) ( w 1 w 0 ) · i N ( a i + b i ) + 2 · w 0 N ,
where w 0 and w 1 are the weights for the background and foreground class, respectively. The pDSC (Equation (A3)) is a special case of the gDSC, where w 0 = 0 .

Appendix A.3. The Dice Loss

Deep learning methods are currently the state of the art on medical image segmentation. Letting X = { x 0 , , x N } represent the input data, each voxel x i has D features (i.e., different contrasts), Y = { y 0 , , y N } represents the labels of that patch and y i { 0 , 1 } is the real binary value of a given voxel, representing whether it belongs to the foreground or background class. Finally, we can define a deep learning model as:
Y ^ = σ ( f i ( X , Θ ) ) ,
where σ is the sigmoid function that provides the final estimate of whether the voxel belongs to the foreground and f i ( X , Θ ) is the output i of the network according to its weights Θ and input X.
These models rely on gradient descent techniques to update their weights according to a loss function that we want to minimise. A common loss for imbalanced datasets (i.e., small lesions inside the brain) is the so-called DSC loss. Given Equation (A5) for the gDSC, the generalised DSC loss presented by Sudre et al. [12] can be defined as:
L D S C ( Y ^ , Y ) = 1 2 · ( w 1 + w 0 ) · i N y ^ i · y i + w 0 · ( N i N ( y ^ i + y i ) ( w 1 w 0 ) · i N ( y ^ i + y i ) + 2 · N w 0 ,
where Y ^ is the estimated segmentation from Equation (A6) and Y is the real segmentation. By definition, the gDSC is maximal when there is a perfect overlap; thus, we need to negate the original equation for a minimisation approach.
Following the chain rule, the derivative of Equation (A7) with respect to the network parameters Θ is given by:
L D S C ( Y ^ , Y ) Θ = L D S C ( Y ^ , Y ) Y ^ Y ^ Θ = i N L D S C ( Y ^ , Y ) y ^ i y ^ i Θ .
Since we are only interested in how the loss affects the gradient of any network, we will only analyse formally its derivative with respect to y ^ i :
F ( Y ^ ) = ( w 1 + w 0 ) · i N y ^ i · y i + w 0 · ( N i N ( y ^ i + y i )
G ( Y ^ ) = ( w 1 w 0 ) · i N ( y ^ i + y i ) + 2 · N w 0
L D S C ( Y ^ , Y ) = 1 2 · F ( Y ^ ) / G ( Y ^ )
L D S C ( Y ^ , Y ) y ^ i = 2 · F ( Y ^ ) y ^ i G ( Y ^ ) G ( Y ^ ) y ^ i F ( Y ^ ) G ( Y ^ ) 2
F ( Y ^ ) y ^ i G ( Y ^ ) = ( ( w 1 + w 0 ) y i w 0 ) ( ( w 1 w 0 ) · j N ( y ^ j + y j ) + 2 · N w 0 ) = ( w 1 2 w 0 2 ) · y i · j N ( y ^ j + y j ) + 2 · N w 0 · ( w 1 + w 0 ) · y i w 0 · ( w 1 w 0 ) · j N ( y ^ j + y j ) 2 · N w 0 2
G ( Y ^ ) y ^ i F ( Y ^ ) = ( w 1 w 0 ) ( ( w 1 + w 0 ) · j N y ^ j y j + w 0 · ( N j N ( y ^ j + y j ) ) ) = ( w 1 2 w 0 2 ) · j N y ^ j y j + w 0 · ( w 1 w 0 ) · j N ( y ^ j + y j ) N w 0 · ( w 1 w 0 )
L D S C ( Y ^ , Y ) y ^ i = 2 · ( w 1 + w 0 ) ( w 1 w 0 ) y i · j N ( y ^ j + y j ) j N ( y ^ j y j ) + N w 0 · ( 2 y i 1 ) ( w 1 w 0 ) · j N ( y ^ j + y j ) + 2 · N w 0 2 .
This equation is independent of the prediction y ^ i . Moreover, the denominator is the same for all possible network outputs. Therefore, from now on, we will refer to it as the normalising factor N ( Y ^ , Y ) = 2 · ( w 1 + w 0 ) ( w 1 w 0 ) · j N ( y ^ j + y j ) + 2 · N w 0 2 to simplify the analysis.
If we go back to Equation (A15), we can observe three different terms in the numerator: a global one G D S C y ^ i = ( w 1 w 0 ) j N ( y ^ j y j ) that is applied to the gradient of all voxels, a term which only affects the foreground voxels F D S C y ^ i = ( w 1 w 0 ) j N ( y ^ j + y j ) and a mixed term that depends on the label M D S C y ^ i = N w 0 · ( 2 y i 1 ) . We will analyse each of them independently with respect to a given weight Θ , and afterwards we will group them together.
Given a true binary segmentation Y, we can define Y 0 = { i : y i = 0 } and Y 1 = { i : y i = 1 } as the sets of voxels from the background and foreground class, respectively. By definition, | Y 0 | + | Y 1 | = N . Now, starting with the global term, we can define the gradient with respect to the output y ^ i as follows:
G D S C Θ = ( w 1 w 0 ) | Y 1 | E i Y 1 y ^ i | Y 0 | E i Y 0 y ^ i Θ + | Y 1 | E i Y 1 y ^ i Θ ,
where E i Y 0 represents the expectation of the background voxels and E i Y 1 the expectation of the foreground voxels. Continuing with the foreground term:
F D S C Θ = ( w 1 w 0 ) | Y 1 | + | Y 0 | E i Y 0 y ^ i + | Y 1 | E i Y 1 y ^ i | Y 1 | E i Y 1 y ^ i Θ .
Finally, we can express the third term as:
M D S C Θ = N w 0 | Y 1 | E i Y 1 y ^ i Θ | Y 0 | E i Y 0 y ^ i Θ .
If we now group all three terms together again, the final gradient with respect to Θ is given by:
L D S C Θ = N ( Y ^ , Y ) · G D S C Θ + F D S C Θ + M D S C Θ .
If we ignore the normalising factor to focus on the relationship between background and foreground voxels, we can simplify that equation as:
L D S C Θ ( G D S C Θ + F D S C Θ + M D S C Θ ) = | Y 1 | ( w 1 w 0 ) | Y 0 | E i Y 1 y ^ i E i Y 0 y ^ i Θ ( | Y 1 | + | Y 0 | E i Y 0 y ^ i ) E i Y 1 y ^ i Θ + N w 0 | Y 0 | E i Y 0 y ^ i Θ | Y 1 | E i Y 1 y ^ i Θ .
Since y ^ i = σ ( f i ( X , Θ ) ) as defined by Equation (A6), its derivative with respect to a given weight is given by:
y ^ i Θ = y ^ i ( 1 y ^ i ) f i ( X , Θ ) Θ ,
and so:
E y ^ i Θ = E y ^ i ( 1 y ^ i ) f i ( X , Θ ) Θ .
Analysing that expectation, it becomes 0 when all the predictions have a value of either 0 or 1, independently of the real label. Furthermore, the maximum gradient for a voxel is given when its probability is 0.5. This is usually the case after initialisation. However, as the weights are updated and that probability moves away from 0.5, that gradient will decrease whether the voxel is well-classified or not.
Taking that into account, the gradient from Equation (A19) becomes 0 when all voxels (including background and foreground) have a prediction of 0 or 1. That means that extreme false positives and false negatives will not have an effect on updating the network parameters using gradient descent. Furthermore, in the extreme case where all the foreground voxels have been misclassified as background, they will not be taken into account while updating the network weights. In an extremely unbalanced dataset, this risk is important since the network might not be able to recover from that and miss all positive examples.
Now, if we analyse Equation (A20) with respect to the expectations of y ^ i , we can see how E i Y 1 y ^ i represents the true positive rate for foreground voxels, while E i Y 0 y ^ i can be regarded as a the rate of misclassified background voxels. In that sense, if we analyse the first term, the gradient of the background voxels increases as the true positives increase, while the gradient of the foreground voxels decreases as the false positive predictions decrease. On the other hand, the second term only depends on the number of positive and negative voxels. In an extremely unbalanced case, the gradient of the background voxels would dominate that second term since | Y 0 | | Y 1 | . Furthermore, in order for the weight of the first term ( | Y 1 | ( w 1 w 0 ) ) to be larger than the weight of the second term ( N w 0 ), w 1 > N w 0 | Y 1 | + w 0 .
Finally, in the special case where w 0 = 0 , the most common use of the Dice loss originally presented by Milletari et al. [11], the second term becomes 0 and the normalising factor is simplified to N ( Y ^ , Y ) = 2 j N ( y ^ j + y j ) 2 . In that case, the derivative (and loss function) are undefined when all predictions are set to 0 and all the voxels belong to the background. Furthermore, the issues from the first term and the derivative of the sigmoid still apply to that function.

References

  1. Lesjak, Ž.; Pernuš, F.; Likar, B.; Špiclin, Ž. Validation of White-Matter Lesion Change Detection Methods on a Novel Publicly Available MRI Image Database. Neuroinformatics 2016, 1, 403–420. [Google Scholar] [CrossRef] [PubMed]
  2. Commowick, O.; Istace, A.; Kain, M.; Laurent, B.; Leray, F.; Simon, M.; Pop, S.C.; Girard, P.; Améli, R.; Ferré, J.C.; et al. Objective evaluation of multiple sclerosis lesion segmentation using a data management and processing infrastructure. Nat. Sci. Rep. 2018, 8, 1–17. [Google Scholar] [CrossRef]
  3. Vanderbecq, Q.; Xu, E.; Ströer, S.; Couvy-Duchesne, B.; Diaz Melo, M.; Dormont, D.; Colliot, O. Comparison and validation of seven white matter hyperintensities segmentation software in elderly patients. Neuroimage Clin. 2020, 27, 102357. [Google Scholar] [CrossRef] [PubMed]
  4. Kuijf, H.; Biesbroek, J.; Bresser, J.; Heinen, R.; Andermatt, S.; Bento, M.; Berseth, M.; Belyaev, M.; Cardoso, M.; Casamitjana, A.; et al. Standardized Assessment of Automatic Segmentation of White Matter Hyperintensities and Results of the WMH Segmentation Challenge. IEEE Trans. Med. Imag. 2019, 38, 2556–2568. [Google Scholar] [CrossRef] [PubMed]
  5. Zhou, S.K.; Greenspan, H.; Davatzikos, C.; Duncan, J.S.; Van Ginneken, B.; Madabhushi, A.; Prince, J.L.; Rueckert, D.; Summers, R.M. A Review of Deep Learning in Medical Imaging: Imaging Traits, Technology Trends, Case Studies with Progress Highlights, and Future Promises. Proc. IEEE 2021, 109, 820–838. [Google Scholar] [CrossRef] [PubMed]
  6. Bernal, J.; Kushibar, K.; Asfaw, D.S.; Valverde, S.; Oliver, A.; Martí, R.; Lladó, X. Deep convolutional neural networks for brain image analysis on magnetic resonance imaging: A review. Artif. Intell. Med. 2019, 95, 64–81. [Google Scholar] [CrossRef]
  7. Isensee, F.; Jaeger, P.F.; Kohl, S.A.A.; Maier-Hein, K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef] [PubMed]
  8. Valverde, S.; Cabezas, M.; Roura, E.; González-Villà, S.; Pareto, D.; Vilanova, J.C.; Ramió-Torrentà, L.; Rovira, A.; Oliver, A.; Lladó, X. Improving automated multiple sclerosis lesion segmentation with a cascaded 3D convolutional neural network approach. Neuroimage 2017, 155, 159–168. [Google Scholar] [CrossRef] [PubMed]
  9. Li, Z.; Kamnitsas, K.; Glocker, B. Analyzing Overfitting under Class Imbalance in Neural Networks for Image Segmentation. IEEE Trans. Med. Imag. 2021, 40, 1065–1077. [Google Scholar] [CrossRef]
  10. Valverde, S.; Salem, M.; Cabezas, M.; Pareto, D.; Vilanova, J.C.; Ramió-Torrentà, L.; Rovira, À.; Salvi, J.; Oliver, A.; Lladó, X. One-shot domain adaptation in multiple sclerosis lesion segmentation using convolutional neural networks. Neuroimage Clin. 2019, 21, 101638. [Google Scholar] [CrossRef]
  11. Milletari, F.; Navab, N.; Ahmadi, S. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  12. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Jorge Cardoso, M. Generalised Dice Overlap as a Deep Learning Loss Function for Highly Unbalanced Segmentations. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support Workshop—MICCAI, Québec City, QC, Canada, 14 September 2017; pp. 240–248. [Google Scholar]
  13. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  14. Ma, J.; Chen, J.; Ng, M.; Huang, R.; Li, Y.; Li, C.; Yang, X.; Martel, A.L. Loss odyssey in medical image segmentation. Med. Image Anal. 2021, 71, 102035. [Google Scholar] [CrossRef] [PubMed]
  15. Kononenko, I. Bayesian neural networks. Biol. Cybern. 1989, 61, 361–370. [Google Scholar] [CrossRef]
  16. Hernández-Lobato, J.M.; Adams, R. Probabilistic backpropagation for scalable learning of bayesian neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015; pp. 1861–1869. [Google Scholar]
  17. Lakshminarayanan, B.; Pritzel, A.; Blundell, C. Simple and scalable predictive uncertainty estimation using deep ensembles. arXiv 2017, arXiv:1612.01474v3. [Google Scholar]
  18. Izmailov, P.; Vikram, S.; Hoffman, M.D.; Wilson, A.G.G. What Are Bayesian Neural Network Posteriors Really Like? In Proceedings of the 38th International Conference on Machine Learning, Online, 18–24 July 2021; pp. 4629–4640. [Google Scholar]
  19. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 424–432. [Google Scholar]
  20. Cabezas, M.; Luo, Y.; Kyle, K.; Ly, L.; Wang, C.; Barnett, M. Estimating lesion activity through feature similarity: A dual path Unet approach for the MSSEG2 MICCAI challenge. In Proceedings of the MSSEG-2 Challenge Proceedings—MICCAI 2021, Strasbourg, France, 2 September 2021; pp. 107–110. [Google Scholar]
  21. Barnett, M.; Wang, D.; Beadnall, H.; Bischof, A.; Brunacci, D.; Butzkueven, H.; Brown, J.W.L.; Cabezas, M.; Das, T.; Dugal, T.; et al. A real-world clinical validation for AI-based MRI monitoring in multiple sclerosis. NPJ Digit. Med. 2023, 6, 196. [Google Scholar] [CrossRef] [PubMed]
  22. Dice, L.R. Measures of the amount of ecologic association between species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  23. Ostmeier, S.; Axelrod, B.; Isensee, F.; Bertels, J.; Mlynash, M.; Christensen, S.; Lansberg, M.G.; Albers, G.W.; Sheth, R.; Verhaaren, B.F.; et al. USE-Evaluator: Performance metrics for medical image segmentation models supervised by uncertain, small or empty reference annotations in neuroimaging. Med. Image Anal. 2023, 90, 102927. [Google Scholar] [CrossRef] [PubMed]
  24. Ma, Y.; Zhang, C.; Cabezas, M.; Song, Y.; Tang, Z.; Liu, D.; Cai, W.; Barnett, M.; Wang, C. Multiple Sclerosis Lesion Analysis in Brain Magnetic Resonance Images: Techniques and Clinical Applications. IEEE J. Biomed. Health Inform. 2022, 26, 2680–2692. [Google Scholar] [CrossRef] [PubMed]
  25. Kayhan, O.S.; Gemert, J.C.v. On translation invariance in cnns: Convolutional layers can exploit absolute spatial location. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 14274–14285. [Google Scholar]
  26. Islam, M.A.; Kowal, M.; Jia, S.; Derpanis, K.G.; Bruce, N.D. Global pooling, more than meets the eye: Position information is encoded channel-wise in CNNs. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 793–801. [Google Scholar]
  27. Carass, A.; Roy, S.; Gherman, A.; Reinhold, J.C.; Jesson, A.; Arbel, T.; Maier, O.; Handels, H.; Ghafoorian, M.; Platel, B.; et al. Evaluating White Matter Lesion Segmentations with Refined Sørensen-Dice Analysis. Sci. Rep. 2020, 10, 8242. [Google Scholar] [CrossRef] [PubMed]
  28. Crum, W.; Camara, O.; Hill, D. Generalized Overlap Measures for Evaluation and Validation in Medical Image Analysis. IEEE Trans. Med. Imaging 2006, 25, 1451–1461. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Example of slices of a randomly selected subject for each dataset. The first row depicts an example of a T1 slice, a FLAIR slice and the boundary of all the lesions for a WMH2017 subject, while the second row depicts an example of a baseline, follow-up image and the boundary of all the new lesions (not appearing on the baseline image) for a LIT longitudinal subject.
Figure 1. Example of slices of a randomly selected subject for each dataset. The first row depicts an example of a T1 slice, a FLAIR slice and the boundary of all the lesions for a WMH2017 subject, while the second row depicts an example of a baseline, follow-up image and the boundary of all the new lesions (not appearing on the baseline image) for a LIT longitudinal subject.
Sensors 24 01981 g001
Figure 2. Scheme of the simple Unet architecture used for this analysis.
Figure 2. Scheme of the simple Unet architecture used for this analysis.
Sensors 24 01981 g002
Figure 3. Band plots of the D S C , s e n s i t i v i t y and p r e c i s i o n metric for the weights of each epoch for all the folds. The upper and lower bands represent the minimum and maximum values, while the middle line represents the mean for all random seeds and folds. The first three rows summarise the metrics for the LIT dataset, while the last three summarise the results for the WMH challenge dataset.
Figure 3. Band plots of the D S C , s e n s i t i v i t y and p r e c i s i o n metric for the weights of each epoch for all the folds. The upper and lower bands represent the minimum and maximum values, while the middle line represents the mean for all random seeds and folds. The first three rows summarise the metrics for the LIT dataset, while the last three summarise the results for the WMH challenge dataset.
Sensors 24 01981 g003
Figure 4. Band plots of the D S C and s e n s i t i v i t y metric for the weights of each epoch for each fold of the LIT dataset. The upper and lower bands represent the minimum and maximum values, while the middle line represents the mean for all the random seeds.
Figure 4. Band plots of the D S C and s e n s i t i v i t y metric for the weights of each epoch for each fold of the LIT dataset. The upper and lower bands represent the minimum and maximum values, while the middle line represents the mean for all the random seeds.
Sensors 24 01981 g004
Figure 5. Band plots of the D S C and s e n s i t i v i t y metric for the weights of each epoch for each fold of the WMH dataset. The upper and lower bands represent the minimum and maximum values, while the middle line represents the mean for all the random seeds.
Figure 5. Band plots of the D S C and s e n s i t i v i t y metric for the weights of each epoch for each fold of the WMH dataset. The upper and lower bands represent the minimum and maximum values, while the middle line represents the mean for all the random seeds.
Sensors 24 01981 g005
Figure 6. Qualitative examples of three subject from the LIT dataset where new lesions are segmented (those that appear only on the follow-up image on the right). Classified voxel are colour coded to represent true positives (green), false positives (red) and false negatives (blue).
Figure 6. Qualitative examples of three subject from the LIT dataset where new lesions are segmented (those that appear only on the follow-up image on the right). Classified voxel are colour coded to represent true positives (green), false positives (red) and false negatives (blue).
Sensors 24 01981 g006
Figure 7. Qualitative examples of three subject from the WMH dataset, where white matter hyperintensities are segmented. Classified voxel are colour coded to represent true positives (green), false positives (red) and false negatives (blue).
Figure 7. Qualitative examples of three subject from the WMH dataset, where white matter hyperintensities are segmented. Classified voxel are colour coded to represent true positives (green), false positives (red) and false negatives (blue).
Sensors 24 01981 g007
Figure 8. Violin plots for the D S C , s e n s i t i v i t y and p r e c i s i o n metrics for all the folds and training seeds. The first row represents the results on the LIT dataset, while the second row represents the results on the WMH dataset.
Figure 8. Violin plots for the D S C , s e n s i t i v i t y and p r e c i s i o n metrics for all the folds and training seeds. The first row represents the results on the LIT dataset, while the second row represents the results on the WMH dataset.
Sensors 24 01981 g008
Table 1. Statistics on the number of voxels for the foreground and background class for the analysed datasets. % (brain) refers to the lesion voxel percentage mean ± standard deviation for the whole brain, while % (batch) refers to the lesion voxel percentage for all the batches. Numbers in brackets represent the global percentage for the whole dataset and batches.
Table 1. Statistics on the number of voxels for the foreground and background class for the analysed datasets. % (brain) refers to the lesion voxel percentage mean ± standard deviation for the whole brain, while % (batch) refers to the lesion voxel percentage for all the batches. Numbers in brackets represent the global percentage for the whole dataset and batches.
Dataset% (Brain)% (Batch)Mean Lesion SizeLesion Number
WMH20171.54 ± 1.48 [1.57]1.53 ± 2.40 [1.73]111.17503679
LIT longitudinal0.14 ± 0.18 [0.15]0.48 ± 0.84 [0.51]116.13156
Table 2. Summary of the measures for the average of the best results for each fold. Patch and train refer to the patch-wise and image-wise measures for the training set, while test stands for the image-wise measures for the testing set.
Table 2. Summary of the measures for the average of the best results for each fold. Patch and train refer to the patch-wise and image-wise measures for the training set, while test stands for the image-wise measures for the testing set.
Loss DSC Sensitivity ( TP f )Precision (P)
PatchTrainTestPatchTrainTestPatchTrainTest
LIT longitudinal dataset
xent0.640.510.460.620.550.490.540.480.46
gdsc0.930.540.530.960.810.770.820.200.29
dsc0.960.550.520.990.890.840.540.290.24
mixed0.930.550.540.980.860.820.740.180.24
focal10.590.480.460.520.490.490.550.530.49
focal20.890.600.560.970.700.630.780.480.38
new0.790.580.561.000.940.900.360.140.23
WMH challenge 2017
xent0.960.680.420.950.600.550.970.800.47
gdsc0.860.710.690.850.640.630.880.790.76
dsc0.880.560.540.870.550.650.890.600.60
mixed0.890.560.440.890.580.630.900.620.47
focal10.940.710.680.910.640.590.970.800.81
focal20.960.720.720.980.720.690.940.730.75
new0.790.700.711.000.850.810.650.610.63
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cabezas, M.; Diez, Y. An Analysis of Loss Functions for Heavily Imbalanced Lesion Segmentation. Sensors 2024, 24, 1981. https://doi.org/10.3390/s24061981

AMA Style

Cabezas M, Diez Y. An Analysis of Loss Functions for Heavily Imbalanced Lesion Segmentation. Sensors. 2024; 24(6):1981. https://doi.org/10.3390/s24061981

Chicago/Turabian Style

Cabezas, Mariano, and Yago Diez. 2024. "An Analysis of Loss Functions for Heavily Imbalanced Lesion Segmentation" Sensors 24, no. 6: 1981. https://doi.org/10.3390/s24061981

APA Style

Cabezas, M., & Diez, Y. (2024). An Analysis of Loss Functions for Heavily Imbalanced Lesion Segmentation. Sensors, 24(6), 1981. https://doi.org/10.3390/s24061981

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop