Next Article in Journal
Exploration of the X-ray Dark-Field Signal in Mineral Building Materials
Next Article in Special Issue
Data Augmentation in Classification and Segmentation: A Survey and New Strategies
Previous Article in Journal
Existing and Emerging Approaches to Risk Assessment in Patients with Ascending Thoracic Aortic Dilatation
Previous Article in Special Issue
Graphical Image Region Extraction with K-Means Clustering and Watershed
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GEMA—An Automatic Segmentation Method for Real-Time Analysis of Mammalian Cell Growth in Microfluidic Devices

by
Ramiro Isa-Jara
1,2,*,
Camilo Pérez-Sosa
1,3,
Erick Macote-Yparraguirre
1,3,
Natalia Revollo
1,4,
Betiana Lerner
5,6,7,
Santiago Miriuka
1,8,
Claudio Delrieux
1,4,
Maximiliano Pérez
5,6,7,* and
Roland Mertelsmann
9,*
1
CONICET—National Scientific and Technical Research Council, Buenos Aires C1004, Argentina
2
Faculty of Informatics and Electronic, ESPOCH—Polytechnic School of Chimborazo, Riobamba 060155, Ecuador
3
IREN Center, UTN—National Technological University, Pacheco, Buenos Aires B1618, Argentina
4
Department of Electrical and Computer Engineering, USN (National University of the South), Bahía Blanca B8000, Argentina
5
Department of Electrical and Computer Engineering, FIU (Florida International University), Miami, FL 33174, USA
6
Faculty of Engineering, Institute of Engineering Biomedical, UBA (University of Buenos Aires), Buenos Aires C1053, Argentina
7
Collaborative Research Institute Intelligent Oncology, Hermann-Herder-Straße 4, 79104 Freiburg im Breisgau, Germany
8
LIAN-CONICET-FLENI—Foundation for the Fight against Childhood Neurological Diseases, Belén de Escobar B1625, Argentina
9
Department of Medicine I, Medical Center, Faculty of Medicine, University of Freiburg, 79085 Freiburg im Breisgau, Germany
*
Authors to whom correspondence should be addressed.
J. Imaging 2022, 8(10), 281; https://doi.org/10.3390/jimaging8100281
Submission received: 28 July 2022 / Revised: 15 September 2022 / Accepted: 20 September 2022 / Published: 14 October 2022
(This article belongs to the Special Issue Image Segmentation Techniques: Current Status and Future Directions)

Abstract

:
Nowadays, image analysis has a relevant role in most scientific and research areas. This process is used to extract and understand information from images to obtain a model, knowledge, and rules in the decision process. In the case of biological areas, images are acquired to describe the behavior of a biological agent in time such as cells using a mathematical and computational approach to generate a system with automatic control. In this paper, MCF7 cells are used to model their growth and death when they have been injected with a drug. These mammalian cells allow understanding of behavior, gene expression, and drug resistance to breast cancer. For this, an automatic segmentation method called GEMA is presented to analyze the apoptosis and confluence stages of culture by measuring the increase or decrease of the image area occupied by cells in microfluidic devices. In vitro, the biological experiments can be analyzed through a sequence of images taken at specific intervals of time. To automate the image segmentation, the proposed algorithm is based on a Gabor filter, a coefficient of variation (CV), and linear regression. This allows the processing of images in real time during the evolution of biological experiments. Moreover, GEMA has been compared with another three representative methods such as gold standard (manual segmentation), morphological gradient, and a semi-automatic algorithm using FIJI. The experiments show promising results, due to the proposed algorithm achieving an accuracy above 90% and a lower computation time because it requires on average 1 s to process each image. This makes it suitable for image-based real-time automatization of biological lab-on-a-chip experiments.

1. Introduction

Over the past few years, microscopy has been an important technique to study biological environments. Recent advancements in high-resolution fluorescent and brightfield microscopy have been of great importance for detailed visualization of the cells and their subcellular structures. Moreover, improvements in automated microscopy such as high-content screening and high-performance time-lapse, enable automatically computing of phenotypic changes of cells under a wide array of dissimilar environmental conditions and obtaining a better understanding of biological function through the study of cellular dynamics, respectively [1].
Microfluid devices have automated microscope hardware to parallelize experiments to both increase data resolution and test countless biological conditions [2]. Lab-on-a-chip technology has been broadly accepted by biological and medical scientific communities as a promising tool for the control of the microenvironment at the molecular, cellular, and tissue levels [3]. Due to the amount of data from microfluidic chips, it is necessary to develop new algorithms to allow the analysis of data and images with a high performance and efficient use of computational resources.
In [4], an image processing software, Python based image analysis for cell growth (PIACG) has been developed, that is able to calculate the total area of the well occupied by cells with fusiform and rounded morphology in response to different concentrations of fetal bovine serum in microfluidic chips, from microscopy images in transmission light, in a highly efficient way. The software is able to segment and quantify images with diverse cellular densities and different cellular morphology automatically, without parameter adjustment and considerable time-saving in data analysis, through a simple and sensible user interface, providing in a quick and simple way a multitude of statistical data.
Analyzing and developing different methods could be relevant in the research of future treatments against diseases such as breast cancer. Breast cancer is the most common malignancy in women around the world [5]. In quantitative image-based cell biology, cellular assessments are mainly derived from detected objects applying image segmentation methods. This requires automated segmentation which includes algorithm design, evaluation, and computational scalability in high-throughput applications [6].
With this aim, several automatic segmentation algorithms have been developed to work with phase contrast and differential interference contrast data. The most common approaches involve thresholding, active contours, and Watershed techniques [7]. Software packages commercial have also been developed to compound many of these ordinary approaches, suitable across various cell lines and modalities.
In [8], CellProfiler works adequately only for reconstructed images. In [9], ImageJ is open source software for processing and analyzing scientific images. This tool is developed in Java. In [10], BioImageXD is a multi-purpose post-processing tool for bioimaging. BioImageXD (version 1.0) was published in the July 2012 issue of Nature Methods. In [11], Icy is an open community platform for bioimage informatics to visualize, annotate and quantify bioimaging data. In [12], FIJI is a distribution of ImageJ which includes many useful plugins contributed by the community. This tool is used for biological image processing. In [13], FarSight was fruitless with the automatic threshold during foreground segmentation. In [14], Fogbank has been successfully applied to image segmentation using phase contrast, bright field, and fluorescence microscopy and binary images.
At the same time, due to the difficulty in accessing a large amount of labeled data, semi-supervised learning is becoming an attractive solution in medical image segmentation. Their success heavily depends on a large amount of annotated data which are hard to be available in the real world, especially in the field of medical image segmentation due to their expensive and time-consuming nature.
In [15], to diminish the limitation caused by label scarcity, a tripled-uncertainty guided semi-supervised model for medical image segmentation was proposed, which can effectively utilize the unlabeled data to improve the segmentation performance. Based on a mean teacher architecture, this model explores the relationship between the segmentation task, the foreground and background reconstruction task, and the signed distance field prediction task. The experimental results on two public medical datasets have demonstrated the feasibility and superiority of this method.
In [16], FastER was unsuccessful as a consequence of the nature of its maximally stable extremal region (MSER) detector. In [17,18,19], the segmentation algorithms—CellStar, SuperSegger, CellSerpent—were completely inappropriate for label-free non-round adherent cells with Dice coefficient < 0.1 [20]. Additional algorithms for segmentation are presented in [21]. However, it is necessary to improve the performance of segmentation algorithms when a specific application is aborded like biological experiments.
In this case, an advanced machine learning-based algorithms employ image gradients and pixel traces to enhance threshold values, but it still needs the criteria of the user as input to select a background elimination factor according to a dataset [22]. On the other hand, most deep learning-based methods output a probability map and use a hand-crafted threshold to generate the final segmentation result, while how confident the network is of the probability map remains unclear.
In [23], an uncertainty-guided network called UG-Net for automatic medical image segmentation was proposed. This network can learn from and contend with uncertainty by itself in an end-to-end manner. Experimental results show that the proposed network outperforms the state-of-the-art methods for nasopharyngeal carcinoma (NPC) segmentation, lung segmentation, optic disc segmentation, and retinal vessel detection. Although deep learning is currently dominant in medical image segmentation, there is no unified network due to the different target sizes of different tasks.
According to the state of art, most segmentation algorithms could fail under different circumstances which include the variance of contrast, luminescence, and cell densities when they approximate the confluency stage [24]. Unluckily, microscope image analysis remains a bottleneck for image segmentation. While plenty of software packages and algorithms have been developed focused on biological cells, the majority of them crave manual inputs from the experimenter and are designed for further processing which limits both accuracy and performance. Another issue in image analysis is related to real-time applications. It depends on the available resources, both experimentally and computationally. In this sense, it is necessary to develop algorithms capable of processing images optimally and adapted to the environment of biological experimentation.
In this paper, a real-time automatic cell segmentation method called GEMA (Gabor filter in biological experiments with morphology operations for segmentation applications) is proposed. In Figure 1, a general description of the computational system and hardware devices used to analyze the behavior of mammalian cells during experiments is presented.
To overcome the challenges mentioned previously, GEMA is a real-time automatic segmentation method for cell segmentation that uses linear regression to compute the optimal parameters to binarize images using an adaptive threshold method. In addition, linear regression allows to describe and predict the behavior of the experiment according to the image sequence through the computation of the percentage of area covered by cells in each image. This percentage of area represents the evolution of biological agents in time. Evolution is analyzed in two main stages: apoptosis and confluency [25,26].
This paper is organized as follows: Section 2 shows the biological framework. In this section, a description of the MCF7 cells with the different parameters used during experiments is presented. Section 3 shows the proposed method. In the first part, the mathematical description and pseudocode of GEMA are presented. In the second part, three relevant methods used to segment images such as manual segmentation, morphological gradient, and semi-automatic algorithms are presented. These methods are used to compare their performance in the segmentation process of biological images. Section 4 shows the experimental results obtained in the analysis and segmentation of two biological datasets. Moreover, the results are compared to evaluate the performance and accuracy of GEMA in front of the three algorithms. Section 5 shows the discussion about these experimental results, conclusions, and future work.

2. Biological Framework and Materials

The biological framework is focused on analyzing and developing different methods which could be relevant in the research of future treatments against diseases such as breast cancer. This kind of cancer is the most common malignancy in women around the world [5,27].

2.1. Cell Line—MCF7

Over the last few years, breast cancer has become a critical public health problem. In order to research this disease in depth, several cell lines have been designed for in vitro studies. This allows understanding behavior, gene expression, and drug resistance, among other variables [28].
One of these cell lines is MCF7 (Michigan Cancer Foundation-7), which has been extensively studied for 40 years. This type of cell has been studied in both in vitro and in vivo systems, to understand their cell-to-cell interaction, their metabolism, angiogenesis, etc. [29]. However, the maintenance of this cell line is inexpensive and does not usually require extra factors to maintain its characteristics. In addition, it has shown great stability along cell passages. This allows its growth to be homogeneous and the repeatability of the experiments to be plausible, while another point to be noted is its resistance to drugs that make the line candidate for tests with different drugs such as the apoptosis inducer used in this study, camptothecin (CPT) [30].

2.2. Apoptosis of the MCF7

The CPT drug was used to induce apoptosis of the MCF7 cell line. It was dissolved in the medium at concentrations 1 to 5 X and with exposure times of 10, 18, and 24 h. The injection was carried out with flows of 250 , 275 , and 300   μ L / min testing cell behavior in front of different drug concentrations.
Another important feature to note is that these cells are adaptable to various surfaces, both adhesive plastic and glass without the need for a pre-surface treatment. This makes the MCF7 cell line very versatile, especially for the present study where microdevices with a glass surface without adherent treatment were used.

2.3. Design and Fabrication of Microfluidic Devices

According to the Olmos et al., the microdevices were designed with Layout Editor software and transferred to TIL with a 2400 PPI infrared source; then TIL was laminated onto an unexposed photopolymer plate, photopolymer plate was exposed to UVA light at 0.45 J at the back for 10 s, a part of the photopolymer was covered with a mask plate at the back, and the photopolymer plate was exposed to 0.45 J UVA light at the back for 20 s. The above exposures were repeated one at a time.
In the next step, the front part was exposed to UVA light at 19 J for 360 s. After removing the TIL, the plate was washed with PROSOL N-1 solvent (supplied by Eastman Kodak) at 360 mm min−1 was dried in an oven for 30 min. at 50 °C, the plate had a last exposure to UVC light at 10 J for 17 min., and UVA light was applied at 4 J for 2 min. on the cold side with the mold manufactured briefly, a mixture of epoxy resin and curing agent (Crystal -T ack, Novarchem, Villa Martelli, Argentina) on the female mold to replicate the high relief design.
After curing, the epoxy resin mold (ERmold) was separated from the Fmold to form the male mold. Subsequently, a mixture of PDMS and curing agent in a 10:1 weight ratio (Sylgard 184 silicone elastomer kit, Dow Corning, Midland, MI, USA) was poured onto the ER mold and cured in an oven at 40 °C overnight [15]. In Figure 2, the final design and microfluid device PDMS used in the experiments are presented.

2.4. Cell Culture in Microdevices

The MCF7 cells were cultured in DMEN f12 Gibco basal medium, supplemented with 10 % fetal bovine serum, 1 % GlutaMAX and 1 % Penicillin-Streptomycin Gibco as shown in Figure 2. The cell cultures were maintained at 37 °C in a saturated atmosphere with 95 % of air and 5 % of CO 2 .
A total of 100,000 cells were seeded per mL, with flows of 300 μL/min he total volume 140 injected in each device was 500   μ L . Cells were cultured within the microdevices for up to 7 days. The change of medium was performed every 24 h, to stimulate growth exponentially, prior to the induction of apoptosis by CPT.
According to these characteristics, the designed system and biological framework allowed for maintaining the cells in the long term. Therefore, it is necessary to automate some of the processes that are involved in the experiment. In general, a microscope catches an image in an interval of time. This generates an image sequence that is analyzed by computational algorithms to evaluate the state of the cell culture at a specific time.

3. Algorithms for Image Segmentation

According to the introduction, several methods have been developed for automatic segmentation using different techniques. In this work, GEMA is proposed as a novel approach for using in segmentation of biological images. This algorithm is designed to work in real time which is useful both to analyze the increase or decrease in the number of the cells and to describe the behavior in time of a biological experiment during the administration of different drugs.

3.1. Datasets and Software Description

GEMA is programmed in Python using Google Colaboratory (Colab) [31]. It is publicly available to facilitate distribution and adaptation by other researchers. The source code and examples can be found in [32]. Three different methods are compared with GEMA. Firstly, the gold standards are obtained from a manual segmentation by experts. Secondly, a segmentation method based on a morphological gradient operator. And finally, a semi-automatic method using FIJI which is a library widely used in biology.
In order to test the performance of algorithms, 1775 images were used which come from two different experiments called Dataset_1 and Dataset_2. These experiments are available in Kaggle repository [33]. Each dataset contains two different sequences of images: the first one during the growth of cell culture to achieve the confluency stage and the second one during the cell death when the CPT drug was injected. Datasets were acquired from 24 to 72 h.

3.2. Manual Segmentation for Gold Standard Images

Experts have performed a manual segmentation to separate the total area occupied by cells from the image background for obtaining the gold standard. These images have been used to validate the results obtained by GEMA and to compare them with gradient and semi-automatic methods.
The VGG Image Annotator (VIA) program allowed the manual annotation to define the interest regions in images [34]. These regions can be described using geometrical shapes such as rectangles, circles, ellipses, polygons, etc. In this case, the polygon shape was used to select and enclose the cell regions as shown in Figure 3.

3.3. Algorithm Using FIJI for Semi-Automatic Segmentation

A semi-automatic method to quantitate cells by micropore was programmed using the macro developed by Rosero et al. [35] for the ImageJ platform version 1.49 (http://imagej.nih.gov/ij/ (accessed on 23 June 2021)) provided in the public domain by the National Institutes of Health, Bethesda, MD, USA, and coded the script using IJ1 programming language on Fiji image processing software (http://fiji.sc/ Fiji in the public domain (accessed on 23 June 2021)).
After file importation to the macro, the whole stack of the digital RGB images was converted to 8-bit. The convolve tool allows for increasing resolution of the image and brightness of each image to quantify cells. The matrix to increase the intensity used the center that corresponds to the source pixel and the other elements corresponding to neighboring pixels.
The destination pixel is calculated by multiplying each source pixel by its corresponding kernel coefficient and adding the results. Then, variance was used in the stack by replacing each pixel with the neighborhood. Finally, to quantify cells, the final stack was converted to binary in order to measure the percentage of cell area.

3.4. Gradient Operator for Automatic Segmentation

Morphological operators enhance the variation of pixel intensity in each neighborhood [36]. Each pixel of a gradient image measures the change in intensity of that same point in the original image. To segment the cancer cell areas, morphological gradient as gradient operators tend to enhance intensity variations on images is used.
The morphological gradient is defined as the arithmetic difference between the dilation and the erosion operation of an image with a structuring element B. The mathematical definition is presented in Equation (1).
g I = δ ρ B I ϵ ρ B I
where δ ρ B ,   ϵ ρ B are the dilation and the erosion operations, respectively.
The gradient image contains neighborhood-based information: the difference of a kernel of 5 × 5 pixel dilation and erosion. This difference is computed through the grayscale intensity I from RGB channels using I = R + G + B 3 .

3.5. GEMA for Automatic Segmentation

Gradient method and GEMA have a preprocessing stage which is an important step in image analysis. Contrast limited adaptive histogram equalization (CLAHE) [37] and Gaussian filter for smoothing are used to improve the quality of the images.
The GEMA flowchart is presented in Figure 4 where the main processes are: (a) Gabor filtering to identify regions with the highest response. (b) Compute the coefficient of variation (CV) of pixels values in an image. This coefficient will be modeled through a linear regression. (c) Binary conversion using an Adaptive method with a variable parameter, and (d) Morphological operations are applied over the binary image to connect interest regions and/or delete non-interest regions.

3.5.1. Gabor Filter Bank

The Gabor Filter has real and imaginary components representing orthogonal directions [38]. In two dimensions (2D), the Gabor filter is defined as shown in Equation (2):
g x ,   y = f 2 π γ η   e x p x 2 f 2 γ 2 + y 2 f 2 η 2   e j 2 π f   x
where x = x   c o s θ + y   s i n θ , y = x   s i n θ + y c o s θ   . Furthermore x   ,     y are the coordinates of frequencies in the reference system, x ,   y in θ orientation of Gabor function, f represents the distance from origin to center of Gaussian function, γ and η characterize the sharpness of Gaussian function along the major and minor axes.
A set of Gabor filters is used for detecting features in a given input image. This set can be computed in a space of M scales (frequencies) and N rotations to ensure invariance [39]. GEMA uses a set that is composed of 4 filters with 1 scale value and 4 different angles for rotation values. This set avoids the redundancy in features and optimizes the processing time. The real component of this set is presented in Figure 5.

3.5.2. Coefficient of Variation and Linear Regression

The coefficient of variation is a standardized measure of the dispersion of a probability distribution or frequency distribution [40]. A mathematical definition for a population of data is presented in Equation (3):
C V = σ μ
where CV is the coefficient of variation which describes the ratio of standard deviation σ to the mean μ .
In image processing, CV can be used to evaluate the dispersion of pixels values according to the histogram and it is used to analyze the evolution of biological experiments. In the confluence and apoptosis stages, the CV has the minimum value. Figure 6a shows the evolution in time of Dataset_1-CPT described by the CV parameter.
In Figure 6b, the data set { x i ,   y i } with i = 1 ,   ,   n where n is the length of Dataset is presented. These points represent the position in a 2 D space where x i is the number of frame and y i is its CV value. To model this coefficient, a linear regression [41] is used which is based on Equation (4):
Y = m X + b
where Y is the set of predicted values, X is the set of real values, m is the slope and b is the intercept. Both are the adjustable parameters for regression.

3.5.3. Adaptive Method and Morphological Operations

In the adaptive method, the threshold value for binarization is calculated for smaller regions. Different regions will have different threshold values. GEMA uses an adaptive method based on Gaussian and binary functions with their respective threshold values [42].
The threshold value for binary function is computed according to the evolution of the biological experiments. This parameter will take integer values in the i n t e r v a l = 5 ,     11 depending on the value of CV. Finally, the closing and dilate morphological operations are used to enlarge the boundaries of foreground regions and shrink background color holes in an image using a structuring element [37].

4. Results

In manual segmentation, the total segmented area (in pixels) is computed using the formula of Gaussian area [43]. The mathematical algorithm is very useful to find the area of any irregular polygon with whatever number of sides, both in the case of concave and convex polygons as shown in Equation (5):
A = 1 2         i = 1 n 1 x i y i + 1 + x n y 1   i = 1 n 1 x i + 1 y i + x 1 y n
where A is the area, n is the number of sides and ( x i ,   y i ) with i = 1 ,   ,   n are the sorted vertices of the polygon.
Each image has a specific area which depends on its height and width. This area is considered the percentage of 100%. Every manually or automatically segmented image allows for computing the area of MCF7 cells detected in the image in terms of percentage as shown in Equation (6).
P c e l l s = A c e l l s A i m g × 100
where A c e l l s is the area in pixels of the cell regions detected, A i m g is the total area in pixels of image and P c e l l s represents the percentage of cells to describe the evolution of experiment.
In Figure 7, the evolution of the biological experiments is presented through the percentage of area occupied by cells in each image in two Datasets used to test the algorithms.
The gradient approach uses a threshold parameter with the size of 25 to binarization according to the values of the gradient image. In GEMA, images in the confluency stage can be binarized successfully with a threshold value equal to or close to 5 and images in apoptosis stages with a threshold value equal to or close to 11. Therefore, the linear regression allows for updating this value automatically when the CV is greater than 5%.
As GEMA is an algorithm to be used in real-time experiments, the linear regression will be fitting with an error value in r a n g e = 5 % ,   7 % which allows to compute different slope and intercept values according to the evolution of biological experiment. This process is presented in Figure 6b. An example of image segmentation for both methods is presented in Figure 8 and Figure 9, respectively.
In Figure 10 and Figure 11, the results obtained by Gradient and GEMA approaches are presented with a comparison of Manual and FIJI methods during the evaluation of two datasets. A concentration curve was performed to determine the optimal concentration and exposure time for the induction of apoptosis in the culture. Both viability and the percentage of cell death were measured using the previously described TRYPAN-BLUE methodology. Regarding the action of CTP, both in conventional plates and in microdevices, the apoptotic effect began to be seen after 2 h of exposure to the compound and its maximum effect at 18 h.
In Figure 10b, the segmentation obtained for FIJI is better than other methods. This occurs because FIJI is a semi-automatic method that includes manually setting different values to threshold parameters during the analysis of each image of the dataset according to the contrast and pixel distribution in each image. However, the accuracy evaluated in Table 1 shows that the difference between GEMA and FIJI is very low. Furthermore, in the global analysis of the results, GEMA proves to be the best algorithm for image segmentation.
According to the results, the performance of GEMA is the highest compared with other approaches. In order to evaluate GEMA, in front of gold standard values, gradient, and FIJI, two metrics are used Accuracy [44] and Dice score (also known as F1-score) [45].
Accuracy and Dice score are performance metrics for image segmentation problems. The main difference might be the fact that accuracy takes into account true negatives while the Dice score only handles true negatives as uninteresting defaults. In Equations (7) and (8), mathematical expressions for computing accuracy and Dice score are presented:
Accuracy = TP + TN n
Dice   score = 2 × TP 2 × TP + FP +   FN
where TP is the among of True Positives, TN is the among of True Negatives, FP is the among of False Positives, FN is the among of False Negatives and n is the total of analyzed values.
In Table 1 and Table 2, the accuracy and Dice score values are presented which are computed by each method in front of gold standard images. These metrics have been evaluated for every 100 images in each dataset and experiment. Therefore, a total of 19 binary images is used to compute the final results by each metric. Finally, the mean of accuracy and dice score are computed to compare the performance of GEMA with FIJI and Gradient methods.
According to the results, the mean accuracy has been computed through an average of individual accuracy of the total set of segmented images by each method in front of gold standard images. Moreover, Table 1 shows that general performance of algorithms in terms of accuracy in all experiments; GEMA is about 0.86, Gradient is about 0.76 and FIJI is about 0.79.
Instead, Table 2 shows that general performance of algorithms in terms of dice score in all experiments; GEMA is about 0.86, Gradient is about 0.81 and FIJI is about 0.79.

GEMA in Real-Time Applications

In real-time applications, a critical parameter is the time of processing. This parameter has been measured in the segmentation process by each image acquired during the biological experiments.
In Figure 12, the box plots show the distribution of the processing time results which are used to compare and evaluate the performance of GEMA in front of the Gradient and FIJI methods.
According to these results, GEMA shows the best performance related to the Processing Time parameter with a mean value close to 1 s per image. Moreover, GEMA has the less variability, which makes it more stable in real-time applications. In Table 3, the summary of statistical parameters is presented.
To evaluate GEMA in real experiments, a GUI interface called ANSIS Real-time control interface version 0.1.2 was developed using Python version 3.8. This software was programmed by Ramiro Isa-Jara (Ecuador) and it has been shared on the GitHub platform [46]. In Table 4, the general description about inputs/outputs of ANSIS is presented.
In Figure 13, the ANSIS GUI is presented, which allows us to use the GEMA algorithm to detect the percentage of area occupied by MCF7 cell in real-time. Moreover, a syringe pump (Figure 1) can be controlled to inject different fluid values into a microfluid device according to increment or decrement in area of cells during the biological experiments.

5. Discussion and Conclusions

In this paper, a novel algorithm called GEMA for image segmentation is presented. This proposed algorithm has three main components as Gabor filter bank, the Coefficient of variation (CV), and a linear regression method.
The Gabor filter was used due to its capability in being applied in texture analysis. In biological images, it could be necessary for a frequency analysis because they could have high homogeneity in the pixel values. GEMA uses the Gabor filter bank with four directions which allow the identification of image regions with the high-frequency response and avoids redundancy in the output image. The redundancy is a relevant parameter especially in real-time algorithms.
The CV is a useful statistical measure when the dispersion should be analyzed. In this case, CV is a robust parameter because it is independent of the scale of the pixel values in each image which allows describing of the biological experiment. In the confluency and apoptosis stage, the CV has a minimum value (Figure 6a). Therefore, as the experiment progresses, the algorithm analyzes the decrease in CV to obtain the parameters of the regression model.
The last component, the linear regression method, allows for modeling of the experiment (Figure 6b.). This method is used because it is computationally efficient due to two parameters that must be computed only as the slope and intercept. It reduces the processing time focused on real-time applications. Moreover, the segmentation results in terms of biological interpretations were correctly validated by experts. Therefore, GEMA has been applied successfully in biological images to describe an experiment at a specific time. The description is performed through the segmentation of cell regions in a sequence of images taken in real-time.
According to the cell growth and development of the breast cancer line, MCF7 cells have the good performance on the plate as well as in conventional culture controls. The manufactured devices were able to maintain the cell cultures for a long period of time. The system demonstrated high performance for remote and real-time culture control, and in this context, the applications, and versatility of the system make it applicable to multiple branches of cell biology, such as development and cell differentiation, drug testing, etc.
According to the accuracy and dice mean presented in Table 1 and Table 2, GEMA shows the best performance compared with FIJI and Gradient methods. The average of accuracy in all the experiments is: G E M A = 0.86 , F I J I = 0.79 and G r a d i e n t = 0.76 . The average of dice score in all the experiments is: G E M A = 0.86 , F I J I = 0.79   and G r a d i e n t = 0.81 . These results show that the proposed method outperforms FIJI by 8 % and Gradient by above 10 % .
Therefore, the process described by GEMA is successfully adapted to analysis in real time of biological images. Furthermore, GEMA has the advantage of segmenting cells of the images in real-time which involves “understanding” the evolution of the biological experiment. The Coefficient of Variation allowed us to describe this evolution in time, i.e., the increase or decrease among cells in images. This parameter can be modeled using linear regression.
GEMA performs several regressions during the experiment to adjust the prediction with a maximum error of 5 % . In comparison with the morphological gradient approach, GEMA does not require external parameters since it is capable to fix its behavior in order to obtain the best possible results. Finally, the computing time required by the proposed method is on average 1.05 s per image. According to Table 3, GEMA shows a stable statistical distribution. In reference to standard deviation, GEMA has 84 % and 66 % values better than Gradient and FIJI methods, respectively. These are main characteristics in order to apply it in real-time applications.
Future work will focus on processing different sequences of images using another drug, cell lines, or different environments to study the performance of different drug treatments in real time. Another alternative is to use another machine learning technique, such as deep learning, to improve these results.

Author Contributions

Conceptualization, C.P.-S. and N.R.; Data curation, C.P.-S. Formal analysis, C.D. and R.M. Funding acquisition, B.L. and M.P.; Investigation, R.I.-J. and C.P.-S.; Methodology, C.P.-S., S.M. and R.M.; Project administration, B.L., S.M. and M.P.; Resources, S.M. and M.P.; Software, R.I.-J. and E.M.-Y.; Supervision, R.M.; Validation, N.R. and R.M.; Visualization, R.I.-J. and Erick E.M.-Y.; Writing—original draft, E.M.-Y.; Writing—review & editing, R.I.-J. and C.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Agencia de Promoción Cientifica y Tecnologica PICT E-2018-0277, PID-2018-0023, PICT-2018-02804, PID UTN 7662 and Mertelsmann Foundation gGmbH.

Institutional Review Board Statement

Ethical review and approval are not applicable. The MCF7 breast cancer tumor cell line described in the work was obtained from the American Type Culture Collection (ATCC) and maintained by the Institute of Biology and Experimental Medicine (IBYME), Buenos Aires, Argentina. This cell line was previously used in similar works [47]. This cell line is intended for laboratory research use only. It is not intended for any animal or human therapeutic use, any human or animal consumption, or any diagnostic use. References and other information relating to this material are available at www.atcc.org (accessed on 22 June 2022).

Informed Consent Statement

Not applicable.

Data Availability Statement

The data described in this article are openly available in the https://www.kaggle.com/datasets/ramiroisajara/dataset-gema/ (accessed on 22 June 2022).

Acknowledgments

The authors thank the CONICET, UTN, ESPOCH and USN.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zanella, F.; Lorens, J.B.; Link, W. High content screening: Seeing is believing. Trends Biotechnol. 2010, 28, 237–245. [Google Scholar] [CrossRef] [PubMed]
  2. Lugagne, J.-B.; Lin, H.; Dunlop, M.J. DeLTA: Automated cell segmentation, tracking, and lineage reconstruction using deep learning. PLoS Comput. Biol. 2020, 16, e1007673. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Xu, Z.; Gao, Y.; Hao, Y.; Li, E.; Wang, Y.; Zhang, J.; Wang, W.; Gao, Z.; Wang, Q. Application of a microfluidic chip-based 3D co-culture to test drug sensitivity for individualized treatment of lung cancer. Biomaterials 2013, 34, 4109–4117. [Google Scholar] [CrossRef]
  4. Georg, M.; Fernández-Cabada, T.; Bourguignon, N.; Karp, P.; Peñaherrera, A.B.; Helguera, G.; Lerner, B.; Pérez, M.S.; Mertelsmann, R. Development of image analysis software for quantification of viable cells in microchips. PLoS ONE 2018, 13, e0193605. [Google Scholar] [CrossRef]
  5. Uddin, N.; Wang, X. Identification of Breast Cancer Subtypes Based on Gene Expression Profiles in Breast Cancer Stroma. Clin. Breast Cancer 2022, 22, 521–537. [Google Scholar] [CrossRef]
  6. Bajcsy, P.; Cardone, A.; Chalfoun, J.; Halter, M.; Juba, D.; Kociolek, M.; Majurski, M.; Peskin, A.; Simon, C.; Simon, M.; et al. Survey statistics of automated segmentations applied to optical imaging of mammalian cells. BMC Bioinform. 2015, 16, 330. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Rahali, R.; Dridi, N.; Ben Salem, Y.; Descombes, X.; Debreuve, E.; De Graeve, F.; Dahman, H. Biological Image Segmentation Using Region-Scalable Fitting Energy with B-Spline Level Set Implementation and Watershed. IRBM 2022, in press. [Google Scholar] [CrossRef]
  8. Carpenter, A.E.; Jones, T.R.; Lamprecht, M.R.; Clarke, C.; Kang, I.H.; Friman, O.; Guertin, D.A.; Chang, J.H.; Lindquist, R.A.; Moffat, J.; et al. CellProfiler: Image analysis software for identifying and quantifying cell phenotypes. Genome Biol. 2006, 7, R100. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Schneider, C.A.; Rasband, W.S.; Eliceiri, K.W. NIH Image to ImageJ: 25 Years of image analysis. Nat. Methods 2012, 9, 671–675. [Google Scholar] [CrossRef] [PubMed]
  10. Kankaanpää, P.; Paavolainen, L.; Tiitta, S.; Karjalainen, M.; Päivärinne, J.; Nieminen, J.; Marjomäki, V.; Heino, J.; White, D.J. BioImageXD: An open, general-purpose and high-throughput image-processing platform. Nat. Methods 2012, 9, 683–689. [Google Scholar] [CrossRef] [PubMed]
  11. De Chaumont, F.; Dallongeville, S.; Chenouard, N.; Hervé, N.; Pop, S.; Provoost, T.; Meas-Yedid, V.; Pankajakshan, P.; LeComte, T.; Le Montagner, Y.; et al. Icy: An open bioimage informatics platform for extended reproducible research. Nat. Methods 2012, 9, 690–696. [Google Scholar] [CrossRef] [PubMed]
  12. Schindelin, J.; Arganda-Carreras, I.; Frise, E.; Kaynig, V.; Longair, M.; Pietzsch, T.; Preibisch, S.; Rueden, C.; Saalfeld, S.; Schmid, B.; et al. Fiji: An open-source platform for biological-image analysis. Nat. Methods 2012, 9, 676–682. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Al-Kofahi, Y.; Lassoued, W.; Lee, W.; Roysam, B. Improved Automatic Detection and Segmentation of Cell Nuclei in Histopathology Images. IEEE Trans. Biomed. Eng. 2009, 57, 841–852. [Google Scholar] [CrossRef] [PubMed]
  14. Chalfoun, J.; Majurski, M.; Dima, A.; Stuelten, C.; Peskin, A.; Brady, M. FogBank: A single cell segmentation across multiple cell lines and image modalities. BMC Bioinform. 2014, 15, 431. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Wang, K.; Zhan, B.; Zu, C.; Wu, X.; Zhou, J.; Zhou, L.; Wang, Y. Semi-supervised medical image segmentation via a tripled-uncertainty guided mean teacher model with contrastive learning. Med. Image Anal. 2022, 79, 102447. [Google Scholar] [CrossRef] [PubMed]
  16. Hilsenbeck, O.; Schwarzfischer, M.; Loeffler, D.; Dimopoulos, S.; Hastreiter, S.; Marr, C.; Theis, F.J.; Schroeder, T. fastER: A user-friendly tool for ultrafast and robust cell segmentation in large-scale microscopy. Bioinformatics 2017, 33, 2020–2028. [Google Scholar] [CrossRef] [Green Version]
  17. Stylianidou, S.; Brennan, C.; Nissen, S.B.; Kuwada, N.J.; Wiggins, P.A. SuperSegger: Robust image segmentation, analysis and lineage tracking of bacterial cells. Mol. Microbiol. 2016, 102, 690–700. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Bredies, K.; Wolinski, H. An active-contour based algorithm for the automated segmentation of dense yeast populations on transmission microscopy images. Comput. Vis. Sci. 2011, 14, 341–352. [Google Scholar] [CrossRef]
  19. Versari, C.; Stoma, S.; Batmanov, K.; Llamosi, A.; Mroz, F.; Kaczmarek, A.; Deyell, M.; Lhoussaine, C.; Hersen, P.; Batt, G. Long-term tracking of budding yeast cells in brightfield microscopy: CellStar and the Evaluation Platform. J. R. Soc. Interface 2017, 14, 20160705. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Vicar, T.; Balvan, J.; Jaros, J.; Jug, F.; Kolar, R.; Masarik, M.; Gumulec, J. Cell segmentation methods for label-free contrast microscopy: Review and comprehensive comparison. BMC Bioinform. 2019, 20, 360. [Google Scholar] [CrossRef] [PubMed]
  21. Eliceiri, K.W.; Berthold, M.R.; Goldberg, I.; Ibáñez, L.; Manjunath, B.S.; Martone, M.; Murphy, R.F.; Peng, H.; Plant, A.L.; Roysam, B.; et al. Biological imaging software tools. Nat. Methods 2012, 9, 697–710. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Fantuzzo, J.A.; Mirabella, V.R.; Hamod, A.H.; Hart, R.P.; Zahn, J.D.; Pang, Z.P. Intellicount: High-Throughput Quantification of Fluorescent Synaptic Protein Puncta by Machine Learning. Eneuro 2017, 4. [Google Scholar] [CrossRef] [Green Version]
  23. Tang, P.; Yang, P.; Nie, D.; Wu, X.; Zhou, J.; Wang, Y. Unified medical image segmentation by learning from uncertainty in an end-to-end manner. Knowl. Based Syst. 2022, 241, 108215. [Google Scholar] [CrossRef]
  24. Loewke, N.O.; Pai, S.; Cordeiro, C.; Black, D.; King, B.L.; Contag, C.H.; Chen, B.; Baer, T.M.; Solgaard, O. Automated Cell Segmentation for Quantitative Phase Microscopy. IEEE Trans. Med Imaging 2017, 37, 929–940. [Google Scholar] [CrossRef] [PubMed]
  25. Elmore, S. Apoptosis: A review of programmed cell death. Toxicol. Pathol. 2007, 35, 495–516. [Google Scholar] [CrossRef] [PubMed]
  26. Chang, W.-H.; Yang, Z.-Y.; Chong, T.-W.; Liu, Y.-Y.; Pan, H.-W.; Lin, C.-H. Quantifying Cell Confluency by Plasmonic Nanodot Arrays to Achieve Cultivating Consistency. ACS Sens. 2019, 4, 1816–1824. [Google Scholar] [CrossRef]
  27. Waks, A.G.; Winer, E.P. Breast Cancer Treatment: A Review. JAMA 2019, 321, 288–300. [Google Scholar] [CrossRef] [PubMed]
  28. Martin, L.-A.; Ribas, R.; Simigdala, N.; Schuster, E.; Pancholi, S.; Tenev, T.; Gellert, P.; Buluwela, L.; Harrod, A.; Thornhill, A.; et al. Discovery of naturally occurring ESR1 mutations in breast cancer cell lines modelling endocrine resistance. Nat. Commun. 2017, 8, 1865. [Google Scholar] [CrossRef] [Green Version]
  29. Comşa, Ş.; Cîmpean, A.M.; Raica, M. The Story of MCF-7 Breast Cancer Cell Line: 40 years of Experience in Research. Anticancer Res. 2015, 35, 3147–3154. [Google Scholar]
  30. Uchino, M.; Kojima, H.; Wada, K.; Imada, M.; Onoda, F.; Satofuka, H.; Utsugi, T.; Murakami, Y. Nuclear β-catenin and CD44 upregulation characterize invasive cell populations in non-aggressive MCF-7 breast cancer cells. BMC Cancer 2010, 10, 414. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Bisong, E. Google Colaboratory. In Building Machine Learning and Deep Learning Models on Google Cloud Platform; Apress: Berkeley, CA, USA, 2019. [Google Scholar] [CrossRef]
  32. Isa-Jara, R. GEMA Algorithm. 2022. Available online: https://colab.research.google.com/drive/1A8HItyE1CvgaexnpB5vqhn_O88s0p-R6?usp=sharing (accessed on 30 July 2022).
  33. Perez-Sosa, C.; Isa-Jara, R. GEMA Datasets. 2022. Available online: https://www.kaggle.com/datasets/ramiroisajara/dataset-gema/ (accessed on 15 June 2022).
  34. Dutta, A.; Zisserman, A. The VIA Annotation Software for Images, Audio and Video. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2276–2279. [Google Scholar] [CrossRef] [Green Version]
  35. Olmos, C.M.; Rosero, G.; Fernandez-Cabada, T.; Booth, R.; Der, M.; Cabaleiro, J.M.; Debut, A.; Cumbal, L.; Perez, M.S.; Lerner, B. Hybrid microchannel-solid state micropore device for fast and optical cell detection. RSC Adv. 2020, 10, 5361–5370. Available online: https://pubs.rsc.org/en/content/articlelanding/2020/RA/C9RA09939E (accessed on 21 March 2021). [CrossRef] [PubMed]
  36. Haralick, R.M.; Sternberg, S.R.; Zhuang, X. Image Analysis Using Mathematical Morphology. IEEE Trans. Pattern Anal. Mach. Intell. 1987, 9, 532–550. [Google Scholar] [CrossRef] [PubMed]
  37. Yadav, G.; Maheshwari, S.; Agarwal, A. Contrast limited adaptive histogram equalization based enhancement for real time video system. In Proceedings of the 2014 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Delhi, India, 24–27 September 2014; pp. 2392–2397. [Google Scholar] [CrossRef]
  38. Ilonen, J.; Kämäräinen, J.-K.; Kälviäinen, H.; Kamarainen, J.-K. Efficient Computation of Gabor Features. 2005. Available online: https://pdfs.semanticscholar.org/01cc/80c412d50129bec6ca545b52e3b1a1eac6de.pdf?_ga=2.48446436.351996994.1519063929-1052252548.1496673189 (accessed on 19 February 2018).
  39. Daugman, J.G. Uncertainty relation for resolution in space, spatial frequency, and orientation optimized by two-dimensional visual cortical filters. J. Opt. Soc. Am. A 1985, 2, 1160–1169. Available online: http://www.ncbi.nlm.nih.gov/pubmed/4020513 (accessed on 19 February 2018). [CrossRef] [PubMed]
  40. Brown, C.E. Coefficient of Variation. In Applied Multivariate Statistics in Geohydrology and Related Sciences; Springer: Berlin/Heidelberg, Germany, 1998. [Google Scholar] [CrossRef]
  41. Motulsky, H.; Christopoulos, A. Fitting Models to Biological Data Using Linear and Nonlinear Regression: A Practical Guide to Curve Fitting; Oxford University Press: Oxford, UK, 2004; Available online: https://global.oup.com/academic/product/fitting-models-to-biological-data-using-linear-and-nonlinear-regression-9780195171792?cc=ec&lang=en& (accessed on 12 September 2021).
  42. Roy, P.; Dutta, S.; Dey, N.; Dey, G.; Chakraborty, S.; Ray, R. Adaptive thresholding: A comparative study. In Proceedings of the 2014 International Conference on Control, Instrumentation, Communication and Computational Technologies (ICCICCT), Kanyakumari, India, 10–11 July 2014; pp. 1182–1186. [Google Scholar]
  43. Braden, B. The Surveyor’s Area Formula. Coll. Math. J. 1986, 17, 326–337. [Google Scholar] [CrossRef]
  44. Hasnain, M.; Pasha, M.F.; Ghani, I.; Imran, M.; Alzahrani, M.Y.; Budiarto, R. Evaluating Trust Prediction and Confusion Matrix Measures for Web Services Ranking. IEEE Access 2020, 8, 90847–90861. [Google Scholar] [CrossRef]
  45. Bertels, J.; Eelbode, T.; Berman, M.; Vandermeulen, D.; Maes, F.; Bisschops, R.; Blaschko, M.B. Optimizing the Dice Score and Jaccard Index for Medical Image Segmentation: Theory and Practice. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2019; Springer: Cham, Switzerland, 2019; Volume 11765, pp. 92–100. [Google Scholar] [CrossRef] [Green Version]
  46. Isa-Jara, R.F. ANSIS Real-Time Control Interface; GitHub: Riobamba, Ecuador, 2021; Available online: https://github.com/RamiroIsaJ/ANSIS_GUI (accessed on 23 July 2021).
  47. La Greca, A.D.; Pérez, N.; Castañeda, S.; Milone, P.M.; Scarafía, M.A.; Möbbs, A.; Miriuka, S.G. celldeath: A tool for detection of cell death in transmitted light microscopy images by deep learning-based visual recognition. PLoS ONE 2021, 16, e0253666. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The computational system and hardware devices used in biological experiments.
Figure 1. The computational system and hardware devices used in biological experiments.
Jimaging 08 00281 g001
Figure 2. The PDMS device used during experiments. (a) Design view; (b) Physical view.
Figure 2. The PDMS device used during experiments. (a) Design view; (b) Physical view.
Jimaging 08 00281 g002
Figure 3. The MCF7 breast cancer cell line. (a) Image from Dataset 2—CPT_24H_72H; (b) manual segmentation of cell regions.
Figure 3. The MCF7 breast cancer cell line. (a) Image from Dataset 2—CPT_24H_72H; (b) manual segmentation of cell regions.
Jimaging 08 00281 g003
Figure 4. The GEMA flowchart used to segment MCF7 cells from image background.
Figure 4. The GEMA flowchart used to segment MCF7 cells from image background.
Jimaging 08 00281 g004
Figure 5. Parameters of the Gabor filter set: f = 3.0 ,   s i z e = 21 , 4 rotations: θ 0 = 0 ° , θ 1 = 45 ° , θ 2 = 90 ° ,   θ 3 = 135 ° and γ = 10 ,   η = 0.5 .
Figure 5. Parameters of the Gabor filter set: f = 3.0 ,   s i z e = 21 , 4 rotations: θ 0 = 0 ° , θ 1 = 45 ° , θ 2 = 90 ° ,   θ 3 = 135 ° and γ = 10 ,   η = 0.5 .
Jimaging 08 00281 g005
Figure 6. Dataset-1 with CPT for inducing apoptosis from confluency stage. (a) Evolution of CV in time. (b) Modeling CV using a linear regression to compute and predict the parameter for adaptive threshold.
Figure 6. Dataset-1 with CPT for inducing apoptosis from confluency stage. (a) Evolution of CV in time. (b) Modeling CV using a linear regression to compute and predict the parameter for adaptive threshold.
Jimaging 08 00281 g006
Figure 7. Gold Standard values: percentages of cells obtained with manual segmentation. (a) Evolution of cells to the confluency stage; (b) cells injected with CPT to generate the apoptosis stage.
Figure 7. Gold Standard values: percentages of cells obtained with manual segmentation. (a) Evolution of cells to the confluency stage; (b) cells injected with CPT to generate the apoptosis stage.
Jimaging 08 00281 g007
Figure 8. The segmentation of cells using gradient approach. (a) Grayscale image from Dataset_2_CONFL in apoptosis stage; (b) image applied morphological gradient; (c) binarization using a threshold value equal to 25.
Figure 8. The segmentation of cells using gradient approach. (a) Grayscale image from Dataset_2_CONFL in apoptosis stage; (b) image applied morphological gradient; (c) binarization using a threshold value equal to 25.
Jimaging 08 00281 g008
Figure 9. The segmentation of cells using the GEMA approach. (a) Grayscale image from Dataset_1_CONFL in apoptosis stage; (b) image applied the Gabor filter set; (c) binarization using a closing and dilate morphological operations.
Figure 9. The segmentation of cells using the GEMA approach. (a) Grayscale image from Dataset_1_CONFL in apoptosis stage; (b) image applied the Gabor filter set; (c) binarization using a closing and dilate morphological operations.
Jimaging 08 00281 g009
Figure 10. Results from Dataset_1 (a) The evolution to confluency stage; (b) the evolution to apoptosis stage.
Figure 10. Results from Dataset_1 (a) The evolution to confluency stage; (b) the evolution to apoptosis stage.
Jimaging 08 00281 g010
Figure 11. Results from Dataset_2 (a) The evolution to confluency stage; (b) the evolution to apoptosis stage.
Figure 11. Results from Dataset_2 (a) The evolution to confluency stage; (b) the evolution to apoptosis stage.
Jimaging 08 00281 g011
Figure 12. The box plots of the processing time used by each method during the segmentation process in two datasets.
Figure 12. The box plots of the processing time used by each method during the segmentation process in two datasets.
Jimaging 08 00281 g012
Figure 13. The ANSIS interface to control and analyze the evolution of biological experiment in real-time using GEMA.
Figure 13. The ANSIS interface to control and analyze the evolution of biological experiment in real-time using GEMA.
Jimaging 08 00281 g013
Table 1. Accuracy computed through the confusion matrix by Gradient, FIJI and GEMA approaches relative to Gold-standard images.
Table 1. Accuracy computed through the confusion matrix by Gradient, FIJI and GEMA approaches relative to Gold-standard images.
Dataset and ApproachAccuracyAccuracy
Mean
0100200300400500
1-CPTFIJI0.6520.7490.785---0.73
Gradient0.8840.8530.699---0.81
GEMA0.8640.8980.979---0.91
2-CPTFIJI0.8890.8420.7870.7810.794-0.82
Gradient0.9590.6170.6240.5940.557-0.67
GEMA0.7360.7930.8320.8410.814-0.80
1-CONFFIJI0.7960.8160.8060.7320.616-0.75
Gradient0.7330.7110.7040.8410.918-0.78
GEMA0.8160.8710.8980.9010.821-0.86
2-CONFFIJI0.8660.8580.8190.8440.8500.8650.85
Gradient0.6800.7570.6530.7740.8600.9060.77
GEMA0.8710.8810.7980.8300.9100.9180.87
Table 2. Dice score computed through the confusion matrix by Gradient, FIJI and GEMA approaches relative to Gold-standard images.
Table 2. Dice score computed through the confusion matrix by Gradient, FIJI and GEMA approaches relative to Gold-standard images.
Dataset and ApproachDice ScoreDice Score
Mean
0100200300400500
1-CPT FIJI0.7450.8050.810---0.79
Gradient0.9330.9060.785---0.87
GEMA0.9200.9150.859---0.90
2-CPT FIJI0.9400.8300.7420.7250.747-0.80
Gradient0.9790.6940.6590.6310.620-0.72
GEMA0.8420.7970.8020.8030.781-0.80
1-CONF FIJI0.7900.8130.8000.7890.737-0.79
Gradient0.7840.7630.7510.9000.956-0.83
GEMA0.8330.8460.8250.8580.863-0.85
2-CONF FIJI0.8440.8720.8380.8820.9020.9180.88
Gradient0.7120.8220.7450.8520.9190.9480.83
GEMA0.8420.8930.8290.8820.9450.9550.89
Table 3. A summary of statistical measures by each method according to the processing time.
Table 3. A summary of statistical measures by each method according to the processing time.
MethodMeanMedianStandard Deviation
GEMA1.051.110.14
Gradient4.315.010.88
FIJI0.951.200.41
Table 4. A general description of the ANSIS Real-time control interface.
Table 4. A general description of the ANSIS Real-time control interface.
Inputs/OutputsDescriptionType/Value
Operative systemTo run the program Windows, Linux (64 bits)
Image typeRead and process imagesJPG, PNG, TIFF
SettingsMax fluid value to pump 250 300   μL/min
Min fluid value to pump 200 250   μL/min
Time to process a new image 0 5 min
MethodsAlgorithms for segmentationGEMA, Gradient
Type of analysis Online (real-time), Offline
DirectoriesRead images and save resultsSource, Destiny
Serial portConnect to syringe pumpSerial COM
Baud rate value 9600   bpm
GraphicsVisualizationImage segmentation, Real-time of cell growth
ResultsPercentage of cell areaMean, Current, Previous
Name of imagesImage
State of pumpON/OFF
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Isa-Jara, R.; Pérez-Sosa, C.; Macote-Yparraguirre, E.; Revollo, N.; Lerner, B.; Miriuka, S.; Delrieux, C.; Pérez, M.; Mertelsmann, R. GEMA—An Automatic Segmentation Method for Real-Time Analysis of Mammalian Cell Growth in Microfluidic Devices. J. Imaging 2022, 8, 281. https://doi.org/10.3390/jimaging8100281

AMA Style

Isa-Jara R, Pérez-Sosa C, Macote-Yparraguirre E, Revollo N, Lerner B, Miriuka S, Delrieux C, Pérez M, Mertelsmann R. GEMA—An Automatic Segmentation Method for Real-Time Analysis of Mammalian Cell Growth in Microfluidic Devices. Journal of Imaging. 2022; 8(10):281. https://doi.org/10.3390/jimaging8100281

Chicago/Turabian Style

Isa-Jara, Ramiro, Camilo Pérez-Sosa, Erick Macote-Yparraguirre, Natalia Revollo, Betiana Lerner, Santiago Miriuka, Claudio Delrieux, Maximiliano Pérez, and Roland Mertelsmann. 2022. "GEMA—An Automatic Segmentation Method for Real-Time Analysis of Mammalian Cell Growth in Microfluidic Devices" Journal of Imaging 8, no. 10: 281. https://doi.org/10.3390/jimaging8100281

APA Style

Isa-Jara, R., Pérez-Sosa, C., Macote-Yparraguirre, E., Revollo, N., Lerner, B., Miriuka, S., Delrieux, C., Pérez, M., & Mertelsmann, R. (2022). GEMA—An Automatic Segmentation Method for Real-Time Analysis of Mammalian Cell Growth in Microfluidic Devices. Journal of Imaging, 8(10), 281. https://doi.org/10.3390/jimaging8100281

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop