Next Article in Journal
Clinical Outcomes of Diabetic Ketoacidosis in Type 2 Diabetes Patients with and without SGLT2 Inhibitor Treatment: A Retrospective Study
Next Article in Special Issue
Alveolar Type 2 Epithelial Cell Organoids: Focus on Culture Methods
Previous Article in Journal
Regulation of Protein-Induced Apoptosis and Autophagy in Human Hepatocytes Treated with Metformin and Paclitaxel In Silico and In Vitro
Previous Article in Special Issue
The Role of Patient-Derived Organoids in Triple-Negative Breast Cancer Drug Screening
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mu-Net a Light Architecture for Small Dataset Segmentation of Brain Organoid Bright-Field Images

by
Clara Brémond Martin
1,2,
Camille Simon Chane
1,*,
Cédric Clouchoux
2 and
Aymeric Histace
1
1
ETIS Laboratory UMR 8051 (CY Cergy Paris Université, ENSEA, CNRS), 6 Avenue du Ponceau, 95000 Cergy, France
2
Witsee, 33 Ave. des Champs-Élysées, 75008 Paris, France
*
Author to whom correspondence should be addressed.
Biomedicines 2023, 11(10), 2687; https://doi.org/10.3390/biomedicines11102687
Submission received: 29 May 2023 / Revised: 30 July 2023 / Accepted: 4 August 2023 / Published: 30 September 2023
(This article belongs to the Special Issue Organoids in Biomedical Research)

Abstract

:
To characterize the growth of brain organoids (BOs), cultures that replicate some early physiological or pathological developments of the human brain are usually manually extracted. Due to their novelty, only small datasets of these images are available, but segmenting the organoid shape automatically with deep learning (DL) tools requires a larger number of images. Light U-Net segmentation architectures, which reduce the training time while increasing the sensitivity under small input datasets, have recently emerged. We further reduce the U-Net architecture and compare the proposed architecture (MU-Net) with U-Net and UNet-Mini on bright-field images of BOs using several data augmentation strategies. In each case, we perform leave-one-out cross-validation on 40 original and 40 synthesized images with an optimized adversarial autoencoder (AAE) or on 40 transformed images. The best results are achieved with U-Net segmentation trained on optimized augmentation. However, our novel method, MU-Net, is more robust: it achieves nearly as accurate segmentation results regardless of the dataset used for training (various AAEs or a transformation augmentation). In this study, we confirm that small datasets of BOs can be segmented with a light U-Net method almost as accurately as with the original method.

Graphical Abstract

1. Introduction

Brain organoids (BOs) are 3D in vitro brain-like cultures that are paving the way for promising alternatives to in vivo models for analyzing brain structure and function and performing personal diagnosis for drug screening. Indeed, these cultures replicate cell type complexity (for instance, communication), the self-organization of tissues, and brain organization: they present various distinct cerebral regions, such as ventricular zones [1]. However, BOs suffer from batch syndrome: BOs in the same culture environment do not necessarily develop the same morphology [2]. To individually follow the development of cultures that model physiological or pathological development, neurobiologists use image analysis tools that are often not automated and adapted to the nonspherical shape of the cultures [3].
In 2020, a team developed software to analyze the fluorescence volumes of intact BOs [4]. This software includes a proposed image segmentation tool that extracts the object shape from its background using the well-known U-Net architecture to characterize the morphology of ventricular zones with transformation data augmentation strategies (flip-flop, rotation) [4]. U-Net is an encoder–decoder network created by [5] to segment annotated images [5]. The contracting path captures the context of images and enables cell localization. In a previous work, we proposed using U-Net with various data augmentation optimization strategies in an adversarial autoencoder (AAE) (which synthesizes natural and variable images from the original ones) during the training step of the segmentation process to extract the shape of BOs in bright-field images [6]. However, this architecture does not seem suited for BO extraction due to some issues: the prolonged training time, the image number requirements (and classic data augmentation strategies are insufficient for reaching a high enough accuracy), and poor neuroepithelial zone segmentation accuracy depending on the optimization used (essential for determining whether the brain culture is growing well after 13 days of culture).
Since its first implementation, U-Net has become a reference for biomedical object segmentation, and recent developments could help progress brain organoid segmentation [7]. For instance, HookNet includes a multibranch encoder–decoder approach, with one branch used for contextual information and the other for target information [8], and BB-Unet, a reduction layer and a bounding box prior filtering approach [9]. Ref. [10] reduces only convolutional layers and uses dropout regularization [10]. In U-Net++, the structure is lightened by employing densely nested strip pathways [11]. The stripped-down U-Net (SD-UNet) is composed of depthwise separable convolutions with a weighted standardization to accelerate performance and decrease computing time. This model contains fewer parameters to learn, a smaller architecture and a shorter computation time than U-Net [12]. In contrast to SD-Unet, SmaAT-UNet adds an attention module alongside use of depthwise separable convolutions [13]. AFF-UNet uses skip connections with an attention channel and a spatial attention module [14]. DMF-AU includes multiscale constraints on the backbone to reduce information loss and lighten the architecture for small datasets [15]. This network employs a differential match filtering layer to identify structures and anisotropic attention to reinforce information on the most representative features. Others use cascaded networks to reduce the number of parameters to learn [16]. In the CNL-UNet structure, a pretrained encoder is proposed to allow the segmentation of small datasets. By using the transfer learning properties of the encoder, the modified skip connections for reducing gaps between mirrored layers and the classifier and localizer module CNL, they efficiently learn the small amount of data [17]. Another encoder architecture update consists of using transformers after convolutional layers to lighten the architecture, such as in LeViT-UNet [18].
However, these procedures modify a simple aspect of the U-Net architecture, the encoder and decoder structure. A recent U-Net variation is μ -Net, which reduces the vanishing gradient phenomenon and accelerates the model convergence with weight matrices in the decoder, which are the transposes of those in the encoder. However, this strategy must be improved for complex boundary segmentation on variable biological structures due to the architecture depth [19]. Recently, a UNet-Mini architecture was created to segment small datasets. This architecture drastically reduces the number of parameters by updating only three elements [20], using element-wise skip connections (addition instead of concatenation) and reducing the number of layers and the kernel size.
In this paper, we propose Mu-Net, a reduced U-Net architecture dedicated to the segmentation of images from small bright-field BO datasets. To validate our results, we compare our architectural proposition with the current U-Net and UNet-Mini architectures using shallow methods, such as active contour or k-means, and an ablation study.
The procedure consists of training networks with images generated from an AAE using various loss and noise optimization strategies. We compare them with classic data augmentation strategies from [6,21], to highlight a tool for BO segmentations.

2. Materials and Methods

2.1. Resources

To our knowledge, only one of the six open-access organoid image datasets available in the literature concerns BOs with bright-field acquisition, which allows us to study the global development of these cultures [3]. Other imaging methods that require coloration and slicing, for instance, focus mainly on cell composition in BO developmental structures or organoid regions. The original dataset, consisting of 40 images, is from [22]. Twenty pathological and physiological cultures are captured with a bright-field microscope over 3 days. The grayscale images are 1088 × 1388 pixels. However, to execute our script in a reasonable time, the input images are cropped and resized to 250 × 250 pixels, maintaining the original proportions. Transform groups are composed by original images transformed by randomly chosen linear data augmentation strategies (flip flop, rotation and whitening). We created 40 transformed images to compare the AAE with a classic approach. The generated groups consist of 240 images in 6 loss groups optimized through an AAE [6] and 160 images in 4 noise groups optimized through an AAE from [21]. Each of the 6 loss groups is composed of 40 images generated with a binary cross entropy (BCE), a BCE with a normalization (BCE + L1), a least square (LS), a Poisson (POISSON), a Wasserstein loss function (Wass.), and a perceptual Wasserstein loss function (P.Wass.). Bright-field image acquisitions are characterized by shot noise [23], which cannot be generated with GAN loss optimizations [21]. Thus, we also test 4 noise groups composed of 40 images generated with Gaussian noise (the usual noise used in GAN, characterizing IRM acquisition [24]), speckle noise, salt and pepper noise (these two noises, which we consider interesting to test, are used to characterize other acquisitions in the biomedical field, such as laser, radar or microscopic transmission system [25]) and a shot noise injection [23]. For more details on the optimized generative methods used to synthesize images, please refer to [6] for loss update and normalization and to [21] for noise injection. Used datasets and amount of images are summarized in Table 1. We develop all scripts in Python 3.6 with an Anaconda framework containing Keras 2.3.1 and TensorFlow 2.1 and run the scripts on an Intel Core i7-9850HCPU with 2.60 GHz and an NVIDIA Quadro RTX 3000s GPU device.

2.2. Segmentation Architectures

We build an architecture based on the U-Net architecture because we want to reduce execution time while conserving precision. From this perspective, we compare U-Net with UNet-Mini and our Mu-Net architecture. The differences between the three models are depicted in Table 2.

2.2.1. U-Net

Originally created in 2015 by [5], U-Net creates detailed segmentation maps that are particularly effective in the biomedical domain due to its two-part (encoder and decoder) U-shaped architecture, which spreads the contextual information of images inside the network. The encoder part employs a CNN architecture: two 3 × 3 3 convolutions followed by a ReLU activation unit and a max-pooling layer repeated several times. U-Net differs from previous CNNs in its second part, called the decoder part, where at each stage, the feature map is upsampled using 2 × 2 upconvolution. In the encoder part, the feature map from the corresponding layer is cropped and concatenated onto the upsampling map, which is a long skip connection. The final stage consists of a 1 × 1 convolution to reduce the feature map and produce the segmented image.

2.2.2. UNet-Mini

Small dataset segmentation is a challenge in the domain. In 2020, ref. [20] proposed a new architecture based on U-Net, with a few changes: they updated the skip connections with the addition of an elementwise product instead of concatenations; they reduced the layer number to four to reduce overfitting situations; and they reduced the number of kernels used in each convolution (ks = 3 × 3 and stride s = 1. They used a soft-max activation task for non-binary classification.) [20]. This last update reduces the execution time and the number of hyper-parameters to learn (from 128 K instead of 17 M with the original implementation).

2.2.3. Mu-Net

Our reduced U-Net architecture contains 4 similar UNet-Mini encoder–decoder layers with 16, 32, 64 and 64 kernels, as our objective is to reduce the execution time see Figure 1. However, we remove the original skip connections and instead use concatenation for feature reusability and compaction. We also update the size strides (the number of steps between filter applications to the input image) when we downsample in convolution layers, whereas the stride stays at 1 regardless of the convolution layer in UNet-Mini. We use an Adam optimizer with a binary cross entropy loss and a sigmoid activation (for binary segmentation). The model is summarized in Table 3.

2.3. Training

To segment the original images, we train the chosen architectures in a leave-one-out strategy summarized in Figure 2: all original and augmentation groups are used during the training stage except one original image, which is segmented during the test phase using the generated model. We process 40 leave-one-out training models × 6 loss groups + 40 leave-one-out training models × 4 noise groups + 40 leave-one-out training models using a classic augmentation dataset constituted by flip-flopping, rotating, or whitening of original images, resulting in 440 training launches for each model. We retrieve the training time of the models. To end the training and avoid overfitting, we retrieve the training time before the loss value plateaus. Original, transformed and generated images are manually segmented with ITK-SNAP software to obtain ground truth images [26].
To compare the training models, we calculate the execution time of each launch, of each leave-one-out loop, the GPU memory required to run the process, the memory required for model weights, the total number of model parameters and the number of FLOPs. To compare the number of FLOPs across models, we also calculate the FLOPs per parameter ratio.

2.4. Classical Segmentation Methods

We next compare the segmentation results of Mu-Net with those of classic segmentation methods, including an Otsu thresholding strategy from [27], a region growing algorithm [28], a Kmean clustering with a particle swarm optimization from [29], an active contour from [30], a level set from [31], and a watershed method from [32]. The threshold is set to 185 pixels. For the region growing algorithm, we set the regional threshold to 0.2 and the size of growth to 1.

2.5. Comparison of Segmentations

To compare ground truth cerebral organoid content segmentation ( G T ) and derivatives of U-Net (u) under various conditions, mean D i c e scores are calculated as
D i c e ( G T , u ) = 2 | G T u | | G T | + | u |
Using the true positive T P (a pixel identified as BO content which is BO content in the ground truth G T ), false positive F P (a pixel identified as BO content, which is actually background content in the GT), true negative T N (correctly identified as background content) and false negative F N (incorrectly identified as BO content) values, we calculate the accuracy, the specificity, the sensitivity, and the F1-score of the methods. The accuracy is the ratio of true to positive labels:
A c c u r a c y = T P + T N T P + F P + T N + F N
The sensitivity is the ratio between how many pixels were correctly identified as positives and all positive pixels:
S e n s i t i v i t y = T P T P + F N
The specificity is the ratio between how many pixels were correctly identified as negative and all negative pixels:
S p e c i f i c i t y = T N T N + F P
The precision is the ratio between how many pixels were correctly identified as positives and all pixels labeled as positives:
P r e c i s i o n = T P T P + F P
The F1-score allows us to summarize the precision and recall (sensitivity) through a unique metric:
F 1 S c o r e = 2 P r e c i s i o n S e n s i t i v i t y P r e c i s i o n + S e n s i t i v i t y

2.6. Visualisation

To highlight real/false positive/negative segmentation, we create a superimposed image composed of the ground truth and a sample of each segmentation resulting from the various trainings. We update the pixel values in light pink for the F P cerebral organoid segmentations and in light green for the F N .

2.7. Ablation study

To validate the benefits and inconveniences of our methodology, we perform an ablation study between UNet, UNet-Mini and our architecture in terms of scores. As a result, we tested N models with the following configurations:
  • Unet,
  • “Layer”,
  • “filter”,
  • “kernel”,
  • “Ewise”,
  • “layer + filter”,
  • “layer + kernel”,
  • “layer + ewise”,
  • “Filter + kernel”,
  • “Filter + ewise”,
  • “Kernel + ewise”,
  • “layer + filter + kernel”,
  • “layer + kernel + ewise”,
  • “layer + filter + ewise”,
  • “filter + kernel + ewise”,
  • “layer + filter + kernel + ewise”,
  • “layer + filter + kernel + stride” (Mu-Net),
  • “layer + filter + kernel + ewise + stride + activation” (UNet-Mini).
We determine a binary content and always use a sigmoid activation.

3. Results

We suggest completing segmentation tasks using a leave-one-out strategy (n = 79 for training and n = 1 for testing for each data augmentation strategy). We choose the classic U-Net architecture and its two lighter architectures: UNet-Mini and Mu-Net.

3.1. Qualitative

We assume that the data augmentation strategy used during each architecture training time could modify the segmentation results. Table 4 shows a comparison of segmented samples from U-Net, UNet-Mini and Mu-Net using the AAE loss optimizations against samples resulting from classic data augmentation strategies, while Table 5 presents segmented samples according to the AAE noise injection. As shown in Table 4 the U-Net architecture seems to be the most accurate, whereas the UNet-Mini and particularly Mu-Net segmentations exhibit high numbers of false-positive pixels under all strategies except for the Wasserstein loss optimization strategy. As shown in Table 5, U-Net and Mu-Net exhibit similar results, including a small number of false positives around the shape of the cerebral organoid. UNet-Mini results in a particular false positive around the contours of the images, and the salt and pepper noise results in a large number of false negatives.

3.2. Quantitative

As we cannot decipher the real precision of segmentations using human observation on a single image from a dataset, we calculate median scores for each case summarized for loss optimizations, presented in Table 6 and for noise injections, presented in Table 7. Dice scores are higher for all architectures under AAE loss optimization strategies than under classic data augmentation. The highest Dice score is achieved by the U-Net architecture when AAE Perceptual Wasserstein loss optimization is used during training. Use of the loss optimization in UNet-Mini training also results in a high Dice score, but this score is lower than that of the U-Net architecture. As shown by the median Dice scores in Table 7, Mu-Net seems to be the most accurate, followed by U-Net and UNet-Mini. Images from an AAE optimized with a Gaussian noise injection used during Mu-Net training exhibit the highest Dice scores. However, other noise injections in the generative process do not improve segmentation performance.
To summarize, U-Net exhibits the most accurate segmentation results for bright-field images of BOs. However, Mu-Net is a promising alternative that achieves almost the same accuracy as the classic data augmentation strategy, particularly in execution time. UNet-Mini would also be a solution if it achieved similar results in the noise augmentation cases. However, under the salt and pepper, the speckle or shot noise optimized synthetic images used during the training, the results of Mini-Unet are too far from the highest ranking results to consider the shape well segmented.
To verify the effect of each component of the architecture on the segmentation precision of bright-field images in a specific data augmentation strategy, we conduct an ablation study that consists of sequentially suppressing or adding an architecture component and calculating the median Dice scores for each intermediate architecture. These results are summarized in the lowest parts of Table 6 and Table 7. Initially, the layer reduction increases the Dice score regardless of the data augmentation strategy realized, whereas filtration does not improve the segmentation results. Kernel reduction results in the highest Dice score for the binary cross-entropy. Layer reduction of the U-Net architecture increases the segmentation accuracy, particularly when images generated with perceptual Wasserstein loss are used during the training of this model. Similar results are obtained with a layer, filter and kernel reduction combined with images generated with BCE loss used during the training. The best results are obtained when P.Wass loss is used for the generated images during the training time. The best noise injection results are obtained using Gaussian noise regardless of the reduction made.

3.3. Computational Comparisons

We verify our assumption by measuring the average execution time for training and testing a single image and for the full leave-one-out strategy (40 images) of each architecture, regardless of the data augmentation strategy used for training. These results are summarized in Table 8. As assumed, UNet-Mini and Mu-Net achieve faster executions than the original implementation: reduced architectures require 2 h for one launch and, respectively, 104 and 100 h for a leave-one-out loop; U-Net architecture requires 16.5 h for one loop and 660 h for the full leave-one-out procedure. Reduced architectures require less memory to be executed than U-Net. However, U-Net results in a higher number of FLOPs than the reduced architectures. This is possibly because the total number of processes is increased for this architecture. Indeed, if this number of FLOPs is compared to the number of parameters, Unet-Mini exhibits the highest number of FLOPs per parameter. The memory and FLOP results of Mu-Net approach the UNet-Mini efficiency. While Mu-Net achieves the best execution time, UNet-Mini also follows this trend. In summary, UNet-Mini and Mu-Net are similarly effective in terms of computational cost, and both perform better than U-Net.

3.4. Mu-Net and Machine Learning Comparisons

3.4.1. Qualitative Analysis

A sample of each segmentation is presented in Figure 3. The threshold, active contour, Kmean and Mu-Net seem to render the best qualitative results. The region growing algorithm and threshold render the worst watershed results, possibly due to their lower precision linked to the manual setting of some parameters.

3.4.2. Quantitative Analysis

The metric comparisons between the GT and each segmentation, presented in Table 9, are consistent with the previous observations in Figure 3. A comparison of the indices of the Mu-Net and classic machine learning (ML) methods shows that the best results are rendered with the Kmean algorithm. However, this algorithm does not outperform the proposed method in terms of the Dice score and specificity but rather in terms of the F1-score, sensitivity and accuracy. The other segmentation methods do not achieve high scores, with the threshold and region growing algorithms performing the worst.

4. Discussion

The main outcomes of this study are as follows: we present a new light segmentation methodology (Mu-Net), similar to architectures from the literature and employing various data augmentation strategies, to segment small bright-field image datasets of BOs, which have particular formations that are essential to extract. The presented framework, based on the U-Net network, runs the quickest for a single launch or for an entire leave-one-out process but has a computational efficiency similar to that of UNet-Mini. Mu-Net exhibits nearly the highest precision scores obtained with U-Net. It reaches these scores using classic data augmentation strategies during training, whereas similar scores obtained with U-Net use optimized AAE strategies. The ablation study confirms that another intermediate architecture, based upon a single layer reduction, could be used to obtain higher scores. UNet-Mini does not obtain the highest scores in some noise optimization cases, rendering this architecture less robust than Mu-Net or U-Net. Some ML segmentation algorithms seem ill-suited for extracting the shape of BOs without optimization through preprocessing, and these algorithms are only semiautomated. Mu-Net outperforms almost all the ML procedures except the Kmean clustering algorithm according to some scores.
To our knowledge, this is the first automated segmentation dedicated to a small dataset of two-dimensional bright-field images of BOs. Segmentation of small datasets (which are not always labeled) remains an issue in the biomedical image domain. Segmentation is crucial for extracting the morphological shapes and characterizing the physiological or pathological development of BOs to form a brain in vitro model that could fill the gaps of in vivo models [1]. Only U-Net has been used for automatically segmenting ventricular zones in the Scout software for clarified BO images and for extracting the shapes of two dimensional BO images to validate the accuracy of the data augmentation strategy [4,6]. However, the training time, the required number of labeled images and the approximate segmentation of neuro-epithelial regions (essential for determining whether a culture is well developed) convinced us to find an appropriate solution for these particular cultures.
Among light U-Net architectures, Mu-Net and UNet-Mini reduce the training time. This could be explained by the drastic reduction of learned hyperparameters and the layer reduction in these networks [16,20]. Usually, the architecture is specified to produce highly accurate segmentation. However, this involves higher processing and computation costs [13,14,15]. Unlike U-Net derivatives applied to other biomedical images (3D, attention, inception, residual, current neural network, and adversarial), the reduced architecture requires few updates [7] and could be quickly implemented on image segmentation platforms.
Mu-Net requires a small number of parameters, processing memory, and higher FLOPs/parameters than U-Net; however, it does not achieve better computational results than UNet- Mini. This reduced performance improves resource consumption with respect to the number of models to process, benefiting the small architecture field. The model performance needs to be tested in other domains. A Mu-Net optimization could be useful for achieving computation performances similar to those of UNet-Mini. However, Mu-Net exhibits higher segmentation quality than UNet-Mini. Additionally, UNet-Mini is less robust than Mu-Net for maintaining high Dice scores regardless of the noise optimization used in the training step. Thus, Mu-Net is, in our opinion, the best alternative to U-Net, which requires more resources than UNet-Mini. Future research could concern the best compromise with respect to computational performance, specifically the number of parameters, to perform qualitative segmentation that is satisfactory for biomedical experts. Indeed, reduced architectures run faster and require less consumption and resources while achieving segmentation levels that allow shape characterization. In our opinion, architecture reduction is beneficial only if it does not degrade the segmentation quality. Mu-Net may be limited from reaching similar high results on other datasets with higher resolutions. However, to overcome this potential issue, employing an attention module in our reduced architecture can improve the result [13,14].
Mu-Net is almost as accurate as U-Net depending on the data augmentation strategy used during training. In contrast, UNet-Mini never achieves higher Dice scores than 0.85, despite being implemented similarly to Mu-Net. This is explained by the transfer learning method used during training: while U-Net and our Mu-Net architecture use concatenations for long skip connections, UNet-Mini uses an additional long skip connection based upon an elementwise product [5,20]. A skip connection skips layers in the neural network and uses the output of one layer as the input to future layers. In addition, data captured in the initial layers are features corresponding to lower semantic information that are extracted from the input and shared to the output. In concatenation, low-level information is shared between the input and output, and this information is directly shared across the network. This method allows feature compaction and high standardization, which is not possible using addition.
Mu-Net achieves its highest Dice scores when a classic data augmentation strategy is used during its training. The ablation study confirmed that combining layer reduction and kernel size maintains high Dice scores in segmentation, whereas using only layer reduction does not. We propose combining the reduction layers and kernel size and hypothesize that concatenation allows transfer learning on essential features. The classic data augmentation chosen here is supported by simple transformations such as reflections, rotations of 90 degrees or filtered whitening of images. GAN solutions are based on generative and discriminative networks, which have to be trained and optimized. If this behavior is observed with another dataset of bright-field images, this architecture could become more attractive (for avoiding using data augmentation strategies based upon GAN architectures, which is time consuming).
A high rate of false positives is present in the samples chosen to illustrate the architecture results under classic data augmentation. A false positive is defined as an instance of an organoid region that is present in the predicted mask and has zero intersection with the ground truth mask region. The surroundings of BOs include spreading cells [2] that could be interpreted by these networks as organoid regions. The ground truth could impact this result due to the small dataset used for training and the more shallow architecture, which could cause surrounding details to be approximately learned. Observations of the segmentation results using all data augmentation strategies in the same training dataset could be interesting.
Other segmentation strategies do not achieve strong results. The main drawback observed in this study is the need to manually preset some parameters, rendering the segmentation process semiautomated. These unexpected scores are also linked with the fact that ML algorithms are inappropriate for datasets with strong variation [33]. However, in the early developmental stage of BOs, no neuroepithelial formation is exhibited, whereas after 14 days, the BO shape is transformed by this structure [22]. The algorithms are also influenced by batch syndrome: BOs grow differently in the same environment, resulting in one or more neuroepithelial zones [1]. Thus, instead of of furnishing a preset parameter for each image in a particular dataset, ML solutions do not consider these cultures.
The Kmean machine learning algorithm is an exception, achieving the best segmentation scores. By separating the number of parts to segment (here, the background and the BO), this algorithm does not require another setting that influences the score. The Kmean clustering algorithm uniquely compares the cluster containing a centroid with a pixel [29]. Particle swarm optimization allows us to obtain a more precise clustering efficiency and thus a better accuracy than using a single Kmean algorithm.
Mu-Net outperforms all ML strategies except Kmean. Our reduced model of segmentation for a tiny dataset of BO images is automated and does not require presetting parameters. The light architecture allows the model to perform image segmentation almost as quickly as using a Kmean algorithm, producing similar results. However, this architecture currently needs to be improved for segmenting tiny BO datasets. An improvement could be to use prior information during Mu-Net training (such as providing shape or topological information on the kind of image segmented) [34], but this requires a better understanding of the morphological characterization of these cultures.
The proposed Mu-Net architecture could also be compared with other U-Net variations, such as those described in [7]. We observe the fundamental role of concatenation in this architecture; however, other U-Net variations are characterized by dense connections [35]. Combining such an update with the lighter architecture to observe its contributions to the bright-field image segmentation of these cultures could be interesting. Additionally, we choose a supervised version (with manually labeled datasets as ground truth in this paper. In future comparisons, the identification of a zero-shot learning solution to decrease the time/labor of neurobiologists in manual segmentation could be interesting [36]. Concerning the metrics comparison, we could calculate other known scores (F1, accuracy, etc.) in future studies. Augmentation strategies could also be formed with nonlinear transformations or other GAN architectures [37], and other data augmentation strategies could be tested with semantically optimized segmentation using an optimized U-Net [38]. We mention that the noise injection optimization benefit during training must be thorough. We verify the contributions of the ablation study, which includes only a few configurations. In the future, conducting an ablation study on filtering, kernel reduction, or even other skip connections would be interesting.
The main future objective is to reproduce these procedures on another bright-field dataset of BO images or on another biomedical culture (miBrain, cortical, etc.) if no BO dataset is available. Concerning the dataset we use [22], a quantitative comparison of the segmentation quality at various stages (9, 14 and 15 days of development) and states (physiological and pathological) could be interesting. In this paper, we quantify the segmentation quality only for the entire dataset, proving that Mu-Net is robust for all states and stages. Once all the procedures are validated on other datasets, these reduced algorithms could also be used in biosensing, as described in [39]. Indeed, Mu-Net could be used for real-time, label-free detection of biological components in a cost-effective way.

5. Conclusions

In this article, we report the first results on Mu-Net, a light deep architecture for bright-field BO image segmentation in small datasets. The proposed segmentation architecture achieves state-of-the-art performance (U-Net) with a faster computation time than other architectures. Mu-Net also outperforms other computer vision segmentation algorithms and is particularly robust to various data augmentation strategies, unlike UNet-Mini. To consolidate these initial results beyond validating our methodology on other datasets, we also need to examine the particular contributions of combining a lightened architecture with other SOTA methodologies. We believe that optimization, including that of shape/topological information given during the data augmentation step to better train our model, could improve the segmentation.

Author Contributions

This article is an idea of C.B.M., C.C., C.S.C. and A.H. The literature search, segmentation algorithms and data analysis was performed by C.B.M. Graphics tables and figures were conceived by C.B.M. and C.C. The first draft was written by C.B.M. while C.C., C.S.C. and A.H. critically revised the work. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Association nationale de la recherche et de la technologie (ANRT) grant number 2019/0635.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available upon corresponding author request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
BOBrain organoid
AAEAdversarial autoencoder
GANGenerative adversarial network
MLmachine learning
DLdeep learning
GTground truth
BCEBinary cross entropy
BCE + L1Binary cross entropy with L1 normalization
LSLeast square
Wass.Wasserstein
P. WassPerceptual Wasserstein
Salt Pepp.Salt and pepper noise

References

  1. Lancaster, M.A.; Renner, M.; Martin, C.A.; Wenzel, D.; Bicknell, L.S.; Hurles, M.E.; Homfray, T.; Penninger, J.M.; Jackson, A.P.; Knoblich, J.A. Cerebral organoids model human brain development and microcephaly. Nature 2013, 501, 373–379. [Google Scholar] [CrossRef]
  2. Kelava, I.; Lancaster, M.A. Dishing out mini-brains: Current progress and future prospects in brain organoid research. Dev. Biol. 2016, 420, 199–209. [Google Scholar] [CrossRef]
  3. Brémond Martin, C.; Simon Chane, C.; Clouchoux, C.; Histace, A. Recent Trends and Perspectives in Cerebral Organoids Imaging and Analysis. Front. Neurosci. 2021, 15, 629067. [Google Scholar] [CrossRef]
  4. Albanese, A.; Swaney, J.M.; Yun, D.H.; Evans, N.B.; Antonucci, J.M.; Velasco, S.; Sohn, C.H.; Arlotta, P.; Gehrke, L.; Chung, K. Multiscale 3D phenotyping of human cerebral organoids. Sci. Rep. 2020, 10, 21487. [Google Scholar] [CrossRef]
  5. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Volume 18, pp. 234–241. [Google Scholar]
  6. Brémond Martin, C.; Simon Chane, C.; Clouchoux, C.; Histace, A. AAEGAN Loss Optimizations Supporting Data Augmentation on Cerebral Organoid Bright-Field Images. In Proceedings of the 17th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISAPPGRAPP), Virtual Event, 8–10 February 2021; p. 8. [Google Scholar]
  7. Siddique, N.; Sidike, P.; Elkin, C.; Devabhaktuni, V. U-Net and its variants for medical image segmentation: Theory and applications. IEEE Access 2021, 1118, 42. [Google Scholar] [CrossRef]
  8. van Rijthoven, M.; Balkenhol, M.; Siliņa, K.; van der Laak, J.; Ciompi, F. HookNet: Multi-resolution convolutional neural networks for semantic segmentation in histopathology whole-slide images. arXiv 2020, arXiv:2006.12230. [Google Scholar] [CrossRef] [PubMed]
  9. El Jurdi, R.; Petitjean, C.; Honeine, P.; Abdallah, F. BB-UNet: U-Net With Bounding Box Prior. IEEE J. Sel. Top. Signal Process. 2020, 14, 1189–1198. [Google Scholar]
  10. Prasad, P.J.R.; Elle, O.J.; Lindseth, F.; Albregtsen, F.; Kumar, R.P. Modifying U-Net for small dataset: A simplified U-Net version for liver parenchyma segmentation. In Proceedings of the Medical Imaging 2021: Computer-Aided Diagnosis, Virtual Event, 15–20 February 2021; Mazurowski, M.A., Drukker, K., Eds.; International Society for Optics and Photonics (SPIE): Bellingham, WA, USA, 2021; Volume 11597, pp. 396–405. [Google Scholar]
  11. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop (DLMIA 2018) and 8th International Workshop (ML-CDS 2018), Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  12. Gadosey, P.K.; Li, Y.; Agyekum, E.A.; Zhang, T.; Liu, Z.; Yamak, P.T.; Essaf, F. SD-UNet: Stripping down U-Net for Segmentation of Biomedical Images on Platforms with Low Computational Budgets. Diagnostics 2020, 10, 110. [Google Scholar] [CrossRef] [PubMed]
  13. Trebing, K.; Stanczyk, T.; Mehrkanoon, S. Smaat-UNet: Precipitation nowcasting using a small attention-UNet architecture. Pattern Recognit. Lett. 2021, 145, 178–186. [Google Scholar] [CrossRef]
  14. Wang, X.; Hu, Z.; Shi, S.; Hou, M.; Xu, L.; Zhang, X. A deep learning method for optimizing semantic segmentation accuracy of remote sensing images based on improved UNet. Sci. Rep. 2023, 13, 7600. [Google Scholar] [CrossRef] [PubMed]
  15. Tan, Y.; Zhao, S.X.; Yang, K.F.; Li, Y.J. A lightweight network guided with differential matched filtering for retinal vessel segmentation. Comput. Biol. Med. 2023, 160, 106924. [Google Scholar] [CrossRef] [PubMed]
  16. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.M.; Larochelle, H. Brain tumor segmentation with deep neural networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef] [PubMed]
  17. Shuvo, M.B.; Ahommed, R.; Reza, S.; Hashem, M. CNL-UNet: A novel lightweight deep learning architecture for multimodal biomedical image segmentation with false output suppression. Biomed. Signal Process. Control. 2021, 70, 102959. [Google Scholar] [CrossRef]
  18. Xu, G.; Wu, X.; Zhang, X.; He, X. Levit-unet: Make faster encoders with transformer for medical image segmentation. arXiv 2021, arXiv:2107.08623. [Google Scholar] [CrossRef]
  19. Yuan, D.; Xu, Z.; Tian, B.; Wang, H.; Zhan, Y.; Lukasiewicz, T. μ-Net: Medical image segmentation using efficient and effective deep supervision. Comput. Biol. Med. 2023, 160, 106963. [Google Scholar] [CrossRef]
  20. Jirik, M.; Gruber, I.; Moulisova, V.; Schindler, C.; Cervenkova, L.; Palek, R.; Rosendorf, J.; Arlt, J.; Bolek, L.; Dejmek, J.; et al. Semantic segmentation of intralobular and extralobular tissue from liver scaffold H&E images. Sensors 2020, 20, 7063. [Google Scholar]
  21. Brémond Martin, C.; Simon Chane, C.; Clouchoux, C.; Histace, A. AAEGAN Optimization by Purposeful Noise Injection for the Generation of Bright-Field Brain Organoid Images. In Proceedings of the Internal Conference On Image Proceccing Theory Tools and Application: Special Session Biological & Medical Image Analysis, Salzburg, Austria, 19–22 April 2022; p. 6. [Google Scholar]
  22. Gomez-Giro, G.; Arias-Fuenzalida, J.; Jarazo, J.; Zeuschner, D.; Ali, M.; Possemis, N.; Bolognin, S.; Halder, R.; Jäger, C.; Kuper, W.F.E.; et al. Synapse alterations precede neuronal damage and storage pathology in a human cerebral organoid model of CLN3-juvenile neuronal ceroid lipofuscinosis. Acta Neuropathol. Commun. 2019, 7, 222. [Google Scholar] [CrossRef]
  23. Gilroy, K.D. Overcoming Shot noise Limitations with Bright Field Mode. Vis. Res. 2019, 1, 74242. [Google Scholar]
  24. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
  25. Boyat, A.K.; Joshi, B.K. A Review Paper: Noise Models in Digital Image Processing. Signal Image Process. Int. J. 2015, 6, 63–75. [Google Scholar] [CrossRef]
  26. Yushkevich, P.A.; Piven, J.; Hazlett, H.C.; Smith, R.G.; Ho, S.; Gee, J.C.; Gerig, G. User-guided 3D active contour segmentation of anatomical structures: Significantly improved efficiency and reliability. NeuroImage 2006, 31, 1116–1128. [Google Scholar] [CrossRef] [PubMed]
  27. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man, Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  28. Adams, R.; Bischof, L. Seeded region growing. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 641–647. [Google Scholar] [CrossRef]
  29. Vora, P.; Oza, B. A survey on k-mean clustering and particle swarm optimization. Int. J. Sci. Mod. Eng. 2013, 1, 24–26. [Google Scholar]
  30. Menet, S.; Saint-Marc, P.; Medioni, G. Active contour models: Overview, implementation and applications. In Proceedings of the 1990 IEEE International Conference on Systems, Man, and Cybernetics, Los Angeles, CA, USA, 4–7 November 1990; pp. 194–199. [Google Scholar]
  31. Meziou, L.; Histace, A.; Precioso, F.; Matuszewski, B.; Carreiras, F. 3D Confocal Microscopy data analysis using level-set segmentation with alpha-divergence similarity measure. In Proceedings of the International Conference on Computer Vision Theory and Applications, Rome, Italy, 24–26 February 2012; pp. 861–864. [Google Scholar]
  32. Sharma, A.K.; Nandal, A.; Dhaka, A.; Koundal, D.; Bogatinoska, D.C.; Alyami, H. Enhanced watershed segmentation algorithm-based modified ResNet50 model for brain tumor detection. BioMed Res. Int. 2022, 2022, 7348344. [Google Scholar] [CrossRef]
  33. Malhotra, P.; Gupta, S.; Koundal, D.; Zaguia, A.; Enbeyle, W. Deep neural networks for medical image segmentation. J. Healthc. Eng. 2022, 2022, 9580991. [Google Scholar] [CrossRef]
  34. El Jurdi, R.; Petitjean, C.; Honeine, P.; Cheplygina, V.; Abdallah, F. High-level prior-based loss functions for medical image segmentation: A survey. Comput. Vis. Image Underst. 2021, 210, 103248. [Google Scholar] [CrossRef]
  35. Hoang, H.H.; Trinh, H.H. Improvement for Convolutional Neural Networks in Image Classification Using Long Skip Connection. Appl. Sci. 2021, 11, 2092. [Google Scholar] [CrossRef]
  36. Qi, G.J.; Luo, J. Small data challenges in big data era: A survey of recent progress on unsupervised and semi-supervised methods. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 2168–2187. [Google Scholar] [CrossRef] [PubMed]
  37. Yi, X.; Walia, E.; Babyn, P. Generative Adversarial Network in Medical Imaging: A Review. Med. Image Anal. 2019, 58, 101552. [Google Scholar] [CrossRef]
  38. Hao, X.; Yin, L.; Li, X.; Zhang, L.; Yang, R. A Multi-Objective Semantic Segmentation Algorithm Based on Improved U-Net Networks. Remote. Sens. 2023, 15, 1838. [Google Scholar] [CrossRef]
  39. Arano-Martinez, J.A.; Martínez-González, C.L.; Salazar, M.I.; Torres-Torres, C. A framework for biosensors assisted by multiphoton effects and machine learning. Biosensors 2022, 12, 710. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Mu-Net architecture. Arrows represent operations, and blocks represent convolutional or deconvolutional layers constituted by two 3 × 3 convolutions (except for the latest blue block, which is constituted by two 3 × 3 and one 1 × 1 deconvolution). The input consists of 250 × 250 pixel images.
Figure 1. Mu-Net architecture. Arrows represent operations, and blocks represent convolutional or deconvolutional layers constituted by two 3 × 3 convolutions (except for the latest blue block, which is constituted by two 3 × 3 and one 1 × 1 deconvolution). The input consists of 250 × 250 pixel images.
Biomedicines 11 02687 g001
Figure 2. Experimental scheme of the leave-one-out strategy to test the impact of various data augmentation methods on segmentation. The synthetic and traditional datasets are generated at each loop without the leave-one-out image.
Figure 2. Experimental scheme of the leave-one-out strategy to test the impact of various data augmentation methods on segmentation. The synthetic and traditional datasets are generated at each loop without the leave-one-out image.
Biomedicines 11 02687 g002
Figure 3. Qualitative comparisons of segmentations methods (a) Original image from [22]; (b) Ground truth; (c) Threshold; (d) Active contour; (e) Kmeans; (f) Region; (g) Watershed; (h) Mu-Net.
Figure 3. Qualitative comparisons of segmentations methods (a) Original image from [22]; (b) Ground truth; (c) Threshold; (d) Active contour; (e) Kmeans; (f) Region; (g) Watershed; (h) Mu-Net.
Biomedicines 11 02687 g003
Table 1. Sample view for each dataset. (1) is from [22], (3) and (4) samples are respectfully from [6,21]. Only samples from (2) are purposely transformed from the original dataset (1).
Table 1. Sample view for each dataset. (1) is from [22], (3) and (4) samples are respectfully from [6,21]. Only samples from (2) are purposely transformed from the original dataset (1).
DatasetEffectivesSample
(1) Original40Biomedicines 11 02687 i001
(2) Transformed40Biomedicines 11 02687 i002
(3) Loss optimization AAE240Biomedicines 11 02687 i003
(4) Noise injection AAE160Biomedicines 11 02687 i004
Table 2. Summary of differences between U-Net and derivatives architectures.
Table 2. Summary of differences between U-Net and derivatives architectures.
Update NameUnetUNet-MiniMu-Net
Layer Number544
Filter[64, 128, 256, 512, 1024][16, 32, 64, 64][16, 32, 64, 64]
Stridedefault (1)default (1)1 or 2
ActivationSigmoidSoftmaxSigmoid
Skip connectionsConcatenationAdditionConcatenation
Table 3. Mu-Net Model. ‘k’ corresponds to the kernel size, ‘s’ to the stride, ‘a’ to activation, ‘d’ to dropout, ‘p’ to pool size, ‘sz’ to upsampling size. Input is images of 250 × 250 pixels.
Table 3. Mu-Net Model. ‘k’ corresponds to the kernel size, ‘s’ to the stride, ‘a’ to activation, ‘d’ to dropout, ‘p’ to pool size, ‘sz’ to upsampling size. Input is images of 250 × 250 pixels.
NameFilterParameters
Conv16[k = 3, s = 1, a = relu]
Conv16[k = 3, s = 2, a = relu]
BatchNorm
Conv32[k = 3, s = 1, a = relu]
Conv32[k = 3, s = 2, a = relu]
BatchNorm
Conv64[k = 3, s = 1, a = relu]
Conv64[k = 3, s = 2, a = relu]
BatchNorm
Conv64[k = 3, s = 1, a = relu]
Conv64[k = 3, s = 2, a = relu]
BatchNorm
Dropout [d = 0.5]
Maxpooling [p = 2.2]
Deconv64[k = 3, s = 1, a = relu]
Deconv64[k = 3, s = 2, a = relu]
Upsampling64[k = 2, a = relu, sz = 2.2]
Concatenation
Deconv64[k = 3, a = relu]
Deconv64[k = 3, a = relu]
Upsampling32[k = 2, a = relu, sz = 2.2]
Concatenation
Deconv32[k = 3, a = relu]
Deconv32[k = 3, a = relu]
Upsampling16[k = 2, a = relu, sz = 2.2]
Concatenation
Deconv16[k = 3, a = relu]
Deconv16[k = 3, a = relu]
Conv2D2[k = 3, a = relu]
Conv2D1[k = 1, a = sigmoid]
Table 4. Observation of segmentation achieved with several noise optimization data augmentation strategies while training U-Net and its derivative architectures. GT corresponds to the ground truth segmentation. False-positive pixels are colored in purple, false negatives in cyan, true negatives in black and true positives in white.
Table 4. Observation of segmentation achieved with several noise optimization data augmentation strategies while training U-Net and its derivative architectures. GT corresponds to the ground truth segmentation. False-positive pixels are colored in purple, false negatives in cyan, true negatives in black and true positives in white.
GTClassicGaussianSalt & Pepp.SpeckleShot
U-NetBiomedicines 11 02687 i005Biomedicines 11 02687 i006Biomedicines 11 02687 i007Biomedicines 11 02687 i008Biomedicines 11 02687 i009Biomedicines 11 02687 i010
Unet-MiniBiomedicines 11 02687 i011Biomedicines 11 02687 i012Biomedicines 11 02687 i013Biomedicines 11 02687 i014Biomedicines 11 02687 i015Biomedicines 11 02687 i016
Mu-NetBiomedicines 11 02687 i017Biomedicines 11 02687 i018Biomedicines 11 02687 i019Biomedicines 11 02687 i020Biomedicines 11 02687 i021Biomedicines 11 02687 i022
LegendBiomedicines 11 02687 i047Biomedicines 11 02687 i048Biomedicines 11 02687 i049Biomedicines 11 02687 i050
Table 5. Observation of segmentations achieved with several loss injections for the data augmentation strategy while training U-Net and its derivative architectures. GT corresponds to the ground truth segmentation. False-positive pixels are colored in purple, false negatives in cyan, true negatives in black and true positives in white.
Table 5. Observation of segmentations achieved with several loss injections for the data augmentation strategy while training U-Net and its derivative architectures. GT corresponds to the ground truth segmentation. False-positive pixels are colored in purple, false negatives in cyan, true negatives in black and true positives in white.
GTClassicBCEBCE + L1LSPoissonWass.P.Wass.
U-NetBiomedicines 11 02687 i023Biomedicines 11 02687 i024Biomedicines 11 02687 i025Biomedicines 11 02687 i026Biomedicines 11 02687 i027Biomedicines 11 02687 i028Biomedicines 11 02687 i029Biomedicines 11 02687 i030
Unet-MiniBiomedicines 11 02687 i031Biomedicines 11 02687 i032Biomedicines 11 02687 i033Biomedicines 11 02687 i034Biomedicines 11 02687 i035Biomedicines 11 02687 i036Biomedicines 11 02687 i037Biomedicines 11 02687 i038
Mu-NetBiomedicines 11 02687 i039Biomedicines 11 02687 i040Biomedicines 11 02687 i041Biomedicines 11 02687 i042Biomedicines 11 02687 i043Biomedicines 11 02687 i044Biomedicines 11 02687 i045Biomedicines 11 02687 i046
Table 6. Dice scores for segmentation with several loss optimizations of data augmentation strategies during training. The best results per row are bolded.
Table 6. Dice scores for segmentation with several loss optimizations of data augmentation strategies during training. The best results per row are bolded.
ClassicBCEBCE + L1LSPoissonWass.P. Wass.
U-Net0.680.870.870.860.870.880.90
Layer0.830.890.880.870.900.880.91
Filter0.850.560.840.820.830.850.83
Kernel0.850.170.840.850.680.850.64
Ewise0.510.690.720.800.660.680.73
Layer + Filter0.840.870.870.850.880.870.88
Layer + Kernel0.340.350.680.680.350.500.70
Layer + Ewise0.330.210.280.280.250.320.37
Filter + Kernel0.710.780.720.590.700.710.74
Filter + Ewise0.670.390.770.760.590.310.65
Kernel + Ewise0.170.650.550.590.640.650.56
Layer + Filter + Kernel0.830.910.880.860.870.880.87
Layer + Kernel + Ewise0.350.330.160.160.210.310.37
Layer + Filter + Ewise0.420.280.190.360.130.290.41
Filter + Kernel + Ewise0.500.340.290.290.450.370.21
Layer + Filter + Kernel + Ewise (Mu-Net)0.850.880.870.890.870.890.88
Layer + Filter + Kernel + Stride0.840.830.850.780.840.820.81
Layer + Filter + Kernel + Stride + Activation + Ewise (Mini-UNet)0.860.840.870.880.850.880.82
Table 7. Dice scores for segmentation with several noise optimization data augmentation strategies during training. The best results per row are bolded.
Table 7. Dice scores for segmentation with several noise optimization data augmentation strategies during training. The best results per row are bolded.
ClassicGaussianSalt & Pepp.SpeckleShot
U-Net0.680.830.820.740.74
Layer0.830.910.830.830.85
Filter0.850.830.520.730.74
Kernel0.850.640.720.580.75
Ewise0.510.730.220.180.38
Layer + Filter0.840.880.830.690.80
Layer + Kernel0.340.700.770.490.40
Layer + Ewise0.330.200.540.200.32
Filter + Kernel0.710.740.590.550.62
Filter + Ewise0.670.650.450.100.21
Kernel + Ewise0.170.560.450.110.26
Layer + Filter + Kernel0.830.870.810.770.74
Layer + Kernel + Ewise0.350.420.370.260.32
Layer + Filter + Ewise0.420.180.140.080.14
Filter + Kernel + Ewise0.500.210.290.260.16
Layer + Filter + Kernel + Ewise (Mu-Net)0.850.880.800.810.83
Layer + Filter + Kernel + Stride0.840.810.590.750.81
Layer + Filter + Kernel + Stride + Activation + Ewise (Mini-UNet)0.860.820.740.480.77
Table 8. Comparison of training time, memory usage, model parameters and FLOPs per segmentation architecture. Training time is given for both a single image (1 launch) and the entire leave-one-out process (LOO on 40 images).
Table 8. Comparison of training time, memory usage, model parameters and FLOPs per segmentation architecture. Training time is given for both a single image (1 launch) and the entire leave-one-out process (LOO on 40 images).
ArchitectureTime
1 Launch
(s)
Time
LOO
(s)
Memory
GPU
(GB)
Memory
Model
(MB)
Model
Param.
Flops
(MFLOP)
Flops/
Param.
U-Net [5]99039,6133511817 M10929
UNet-Mini [20]15762827.351.31128 k40316
Mu-Net (ours)15160447.832.60680 k4558
Table 9. Comparison of segmentation methods.
Table 9. Comparison of segmentation methods.
ThresholdActiveKmeanRegionWatershedMu-Net
Dice<0.10.620.89<0.10.610.88
F1-score0.400.660.720.200.610.77
Sensitivity0.390.690.770.190.530.90
Specificity0.960.840.890.860.940.86
Accuracy0.830.900.900.70.880.92
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Brémond Martin, C.; Simon Chane, C.; Clouchoux, C.; Histace, A. Mu-Net a Light Architecture for Small Dataset Segmentation of Brain Organoid Bright-Field Images. Biomedicines 2023, 11, 2687. https://doi.org/10.3390/biomedicines11102687

AMA Style

Brémond Martin C, Simon Chane C, Clouchoux C, Histace A. Mu-Net a Light Architecture for Small Dataset Segmentation of Brain Organoid Bright-Field Images. Biomedicines. 2023; 11(10):2687. https://doi.org/10.3390/biomedicines11102687

Chicago/Turabian Style

Brémond Martin, Clara, Camille Simon Chane, Cédric Clouchoux, and Aymeric Histace. 2023. "Mu-Net a Light Architecture for Small Dataset Segmentation of Brain Organoid Bright-Field Images" Biomedicines 11, no. 10: 2687. https://doi.org/10.3390/biomedicines11102687

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop