Next Article in Journal
Screening for Prognostic Biomarkers in Metastatic Adrenocortical Carcinoma by Tissue Micro Arrays Analysis Identifies P53 as an Independent Prognostic Marker of Overall Survival
Next Article in Special Issue
Early Prediction of Planning Adaptation Requirement Indication Due to Volumetric Alterations in Head and Neck Cancer Radiotherapy: A Machine Learning Approach
Previous Article in Journal
LPIN1 Induces Gefitinib Resistance in EGFR Inhibitor-Resistant Non-Small Cell Lung Cancer Cells
Previous Article in Special Issue
A Novel Machine Learning 13-Gene Signature: Improving Risk Analysis and Survival Prediction for Clear Cell Renal Cell Carcinoma Patients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Cancer Cell Taxonomy Using an Ensemble of Deep Neural Networks

1
Department of Medical IT Convergence Engineering, Kumoh National Institute of Technology, Gumi 39177, Korea
2
Department of IT Convergence Engineering, Kumoh National Institute of Technology, Gumi 39177, Korea
3
Department of Data Science, Seoul National University of Science and Technology, Seoul 01811, Korea
*
Author to whom correspondence should be addressed.
Cancers 2022, 14(9), 2224; https://doi.org/10.3390/cancers14092224
Submission received: 27 October 2021 / Revised: 18 April 2022 / Accepted: 26 April 2022 / Published: 29 April 2022
(This article belongs to the Collection Artificial Intelligence and Machine Learning in Cancer Research)

Abstract

:

Simple Summary

As of recently, cancer is considered a major cause of death in developed and developing countries. Therefore, there is an urgent need for improvements in existing diagnostic methods for effective early diagnosis. However, cross-contamination of cancer cell lines results in the development of inappropriate treatments that cannot be administered to patients. To address this issue, we propose an automatic cancer cell taxonomy with high accuracy using optical images of cells obtained through low-scale benchtop optical microscopy. Specifically, we built a deep-learning-based framework to classify cervical, hepatocellular, breast, and lung cancer cells. The experimental results demonstrated that the proposed deep-learning-based approach facilitates the automatic identification of cancer cells. Moreover, our findings provide important insights into the design of convolutional neural networks for various clinical tasks that utilize microscopic images.

Abstract

Microscopic image-based analysis has been intensively performed for pathological studies and diagnosis of diseases. However, mis-authentication of cell lines due to misjudgments by pathologists has been recognized as a serious problem. To address this problem, we propose a deep-learning-based approach for the automatic taxonomy of cancer cell types. A total of 889 bright-field microscopic images of four cancer cell lines were acquired using a benchtop microscope. Individual cells were further segmented and augmented to increase the image dataset. Afterward, deep transfer learning was adopted to accelerate the classification of cancer types. Experiments revealed that the deep-learning-based methods outperformed traditional machine-learning-based methods. Moreover, the Wilcoxon signed-rank test showed that deep ensemble approaches outperformed individual deep-learning-based models (p < 0.001) and were in effect to achieve the classification accuracy up to 97.735%. Additional investigation with the Wilcoxon signed-rank test was conducted to consider various network design choices, such as the type of optimizer, type of learning rate scheduler, degree of fine-tuning, and use of data augmentation. Finally, it was found that the using data augmentation and updating all the weights of a network during fine-tuning improve the overall performance of individual convolutional neural network models.

1. Introduction

Recently, cancer has begun being considered a major cause of death in developed and developing countries; thus, the American Cancer Association and GLOBOCAN estimate the number of new cancer cases and deaths each year and aggregate the most recent data on population-based cancer incidence [1,2]. According to the report, 1,806,590 new cancer patients and 606,520 cancer patients were expected to die in the United States by 2020. Specifically, breast cancer is the most common and leading cause of death in women around the world, and the number of patients increases with age, but early diagnosis can increase breast cancer survival by up to 80 percent [3]. Therefore, there is an urgent need for improvements in existing diagnostic methods for effective early diagnosis.
Typically, to diagnose cancer, a radiologist identifies suspicious locations through diagnostic equipment such as X-rays, Magnetic Resonance Imaging (MRI), Computed Tomography (CT), etc., and then conducts a biopsy to check for abnormalities under a microscope [4,5,6,7,8]. Biopsy during clinical diagnosis is an efficient and accurate diagnostic method for cancer detection, and plays an important role in breast cancer as well as in other types of cancer [8,9]. In this approach, a pathologist analyzes a tissue sample of a suspected cancer cell metastasis under a microscope for the detection and classification of tumors. While pathologists familiar with clinical tissues can determine two types of lesions, benign and malignant, manual analysis of microscopic images is a very complex, challenging task, and sometimes misjudged [10]. Therefore, extensive research on computer-aided diagnosis has been actively conducted to increase the accuracy of diagnosis [11,12,13,14,15].
For decades, microscopic image analysis methods have been widely used for biological studies and diagnosis of various diseases, including specific cell counts [16], cell location [12,13], cell shape [17], and cell categorization [14,15]. In particular, microscopic images acquired of tissue or cells facilitate the validation of the presence of certain diseases [18], the categorization of tumor types [19], and the interpretation of cell and molecular genetic mechanisms [20]. However, mis-authentication of cell lines due to cross-contamination has been acknowledged as a serious problem over the past 50 years [21,22]. Generally, cross-contamination of cancer cell lines can be caused by incorrect labeling, cross-use of pipette tips, and sharing of cell culture media [23,24].
Due to the changed or contaminated cell lines, researchers perform experiments using inappropriate cells, resulting in the development of treatments that cannot be administered to patients [25]. Therefore, institutions such as the National Institutes of Health and The International Cell Line Authentication Committee have required additional tests to authenticate the type of cells they are trying to use before conducting relevant research [26,27,28]. Various molecular interpretation trials have been used to solve these problems and identify cell lines, and alternative methods have been actively studied [23,29]. The most widely used at present is short tandem repeat (STR) analysis, which reveals the number of repeated DNA traces of particular DNA motifs [30]. Each sample cell is amplified and processed during STR profiling, and the resulting value is determined to be the same as the standardized cell line profile with approximately 80% similarity [30]. However, STR profiling must be implemented by an experienced professional and is not readily available to users due to its relatively high cost and limited use. In addition, because STR profiling is only suitable for distinguishing cell lines from a single species, researchers need specialized knowledge of the biological differences in each cell [31]. However, even with STR analysis, it was confirmed that 15–20% of the currently used cell lines were incorrectly identified [32]. For example, previous work has confirmed that up to 96 cell lines were misidentified when 482 different human tumor cell lines were analyzed using STR profiling, thus finding that STR profiling alone is insufficient [28]. Therefore, we need an alternative approach for cancer cell line classification that can be easily applied by non-experts in the laboratory, and several artificial-intelligence-based taxonomies of cancer cell lines have been introduced [9,16,17,33].
Recently, convolutional neural networks (CNNs) that can independently extract and construct discriminative features from the data have garnered widespread interest from researchers [34,35,36]. However, in order to utilize images of cells obtained through optical microscopy in deep learning, researchers use expensive and customized equipment such as high-scale microscopy [37], high-frequency single-beam acoustic tweezers [38], hyperspectral imaging systems [39,40], and time-stretch quantitative phase imaging systems [15]. Furthermore, considerable time and effort are required to prepare images stained in various colors as training and test data [9,41].
Therefore, in this work, we propose an automatic cancer cell taxonomy using optical images of cells obtained through low-scale benchtop optical microscopy that is typically used in laboratories. For the automatic classification of four cancer cell types, various deep learning models were trained using a transfer learning approach. We also presented a pipeline for ensemble approach based on both individual deep learning models and multiple heterogeneous models. The main contributions of this study are threefold:
  • We proposed a deep-learning-based approach to prevent cross-contamination of several heterogeneous cancer cell lines.
  • The experimental results showed that the proposed deep-learning-based approach identifies with an accuracy over 97%, demonstrating that our method can be a promising alternative approach to STR for the automated cancer cell taxonomy.
  • We presented and discussed the effects of various design choices on the overall performance of CNN architectures for various clinical tasks that utilize microscopic images.
The rest of this paper is as follows. Section 2 describes the details of the proposed approach. Section 3 presents various experimental results, and Section 4 provides discussions on the experimental results. Finally, we summarize and conclude our work in Section 5.

2. Method

Figure 1 depicts the proposed framework consisting of four phases: image preparation, image preprocessing, training, and testing of the CNNs. The next subsections describe the details of each step.

2.1. Image Preparation

The cancer cell lines were cultured for seven days, and the bright-field images were acquired every day using an inverted fluorescent microscope (IX73 with DP80, Olympus Corp., Tokyo, Japan). Four cells were used in the experiment: HeLa (human, cervical cancer cells), MCF-7 (human, female, 69 years old, Caucasian, breast cancer cells), Huh7 (human, liver cancer cells), and NCI-H1299 (human, lung cancer cells). All cells were purchased from Korean Cell Line Bank (Seoul, Korea) and cultured in the following manner. The cell lines were cultured in a high-glucose Dulbecco’s Modified Eagle Medium containing 10% Fetal Bovine Serum with 1% penicillin streptomycin. The prepared cells were incubated at 37 C in a humidified incubator with 5% C O 2 [42]. The prepared cells were trypsinized when 80% confluence was achieved, washed three times with phosphate buffer solution (PBS) [43] to separate the cells, and prepared with an approximate concentration of 1 × 10 6 cells/mL. In total, 889 cell images were collected through the microscope for seven days after starting the cell culture: 247 images in HeLa, 281 images in Huh7, 149 images in MCF7, and 212 images in NCI-H1299.

2.2. Image Preprocessing

To obtain several morphological types of cell images which will be used for training and testing various CNNs, the acquired images were preprocessed using OpenCV and scikit-image libraries, which are popular open-source libraries used for computer vision and image preprocessing tasks, such as scale transformation, denoising, and adaptive thresholds for Region Of Interest (ROI) of bright field images. The preprocessing steps of the cell images are summarized in Figure 2. First, the brightfield cell image acquired through the microscope (Figure 2a) was converted to grayscale (Figure 2b) and then translated into the binary image using adaptive thresholding. Subsequently, noise removal was performed using the dilation function with a 2 × 2 kernel (Figure 2c). The processed image allows the identification of each cell’s contour and the creation of the bounding boxes (Figure 2d). The size of a bounding box is proportional to the size and number of cells, and uninformative cells or floating debris (the sum of width and height less than 100 pixels) were excluded. The final segmented image patches are depicted in Figure 2e. A total of 27,200 samples were collected by segmenting 889 cell brightfield images. Finally, before feeding the image patches to the CNNs, we apply a different normalization step designed for each CNN architecture that will be introduced in the next subsection. Specifically, the normalization methods include (1) scaling the input pixel values to [0, 1] and then normalizing each channel with respect to the ImageNet, (2) converting the colorspace from RGB to BGR first and then zero-centering the pixel values with respect to the ImageNet without scaling, and (3) scaling the pixel values to [−1, 1] or [0, 1] sample-wise. Therefore, all the images are normalized differently according to the CNN architecture used. More details of the image preprocessing step can be found from Table S1 in Supplementary Data.

2.3. Training CNNs for Cancer Classification

Generally, training CNNs from scratch requires a significant amount of data and resources to achieve high performance. Therefore, for efficient training in various domains, transfer learning is widely used, where the weights of a model pretrained on a large-scale dataset are used for solving a new/related task [44]. In this study, we adopted a transfer learning approach wherein the pretrained models are tuned from the general domain (i.e., ImageNet database [45]) to the medical domain (i.e., cancer cell images). Various CNN models pretrained on ImageNet, such as DenseNet121 [46], MobileNetV2 [47], EfficentNetB2 [48], InceptionV3 [49], and ResNet-50 [50], were used as our base networks. All models were trained for 50 epochs using the categorical cross-entropy loss. Moreover, design choices for various learning strategies, such as data augmentation, degree of fine-tuning, optimizer and learning rate scheduler, and ensemble configurations, were considered.

2.3.1. Data Augmentation

Data augmentation is considered one of the most promising techniques to improve the robustness and performance of the CNNs by increasing the amount of data with various transformations toward the original images. Of the several available augmentation techniques in the image domain [51], rotation (random rotation between 0–90 degrees), translation (shifting by 2 pixels), and vertical flip methods were considered to adjust spatial parameters of the original image. Figure 3 shows a set of examples, where rotation (Figure 3B), translation (Figure 3C), vertical flip (Figure 3D), and a combination of these methods are applied (Figure 3E) to the original image (Figure 3A), respectively. During the experiments, the original images or images augmented with a combination method (Figure 3E) were used for training CNNs.

2.3.2. Degree of Fine-Tuning

Generally, training deep learning models from scratch requires a large amount of high-quality data as well as computing resources to achieve a high performance. Therefore, transfer learning with fine-tuning has been popular in training deep learning models, as it can transfer the knowledge learned from a large-scale image dataset to the new/similar domain/task. Moreover, it can help to build a more accurate model with less time and data consumed. While adopting a transfer learning method in our pipeline, we considered the following fine-tuning strategies during the training process: (1) updating all parameters in the pretrained model, or (2) freezing the first 25% of the layers in the pretrained model and updating the rest. Figure 4 shows the difference between these strategies. In contrast to the fine-tuning strategy where all the weights are updated to fit a new domain/task (Figure 4a), the second strategy (Figure 4b) utilizes the fixed weight of early layers learned from ImageNet and updates the rest to suit our domain.

2.3.3. Optimizer and Learning Rate Scheduler

An optimizer is one of the most important components that can affect the training speed and accuracy of the CNNs. Of the several available optimizers [52], we selected the Stochastic Gradient Descent (SGD) optimizer, the most popular gradient-descent-based method, and the Adaptive Gradient (AdaGrad) optimzer [53], one of the most popular adaptive methods, as our candidate optimizers. Specifically, the SGD optimizer updates parameters based on the gradient-descent-based optimization using mini-batch data. On the other hand, AdaGrad works similarly to SGD but adaptively controls a learning rate based on the magnitude of previous gradients.
A learning rate is also an important hyper-parameter that determines the extent to which the parameters should be updated. In this study, we used a fixed learning rate of 0.001 or an exponential decay [53] scheduler that allows adaptive scaling of the learning rate at each iteration, which is defined as:
η ( t ) = η ( 0 ) × e t / r
where η ( 0 ) is the initial learning rate (0.001), e is the decay rate (0.96), t is the current step, and r is the decay step (10,000). The use of a learning rate scheduler allows an adaptive scaling of the learning rate per each training iteration/epoch.

2.4. Ensemble of CNNs

An ensemble approach is a well-known method to improve the performance of a machine-learning-based system by exploiting multiple classification models. A guiding principle in designing ensemble methods has been ’many heads are better than one’ [54]. An ensemble approach typically consists of a set of individual models that predict their own labels for a given sample and therefore can be categorized based on how individual base classifiers are built. Traditionally, in terms of building multiple classifiers, an ensemble approach can be classified into bagging-, boosting-, and stacking-based methods [54]. In bagging, individual base classifiers are trained with a subset of data sampled randomly with replacement [55]. The final prediction is then made by aggregating the result from each base classifier. In this aggregation step, various voting approaches can be considered. Examples of the voting approaches include (1) majority voting (i.e., the predicted target label of the ensemble is the mode of the distribution of individually predicted labels), (2) soft voting (i.e., the predicted target label of the ensemble is the class with the largest sum of probabilities from models), and (3) weighted voting schemes (i.e., the result from each base model is weighted by the model’s importance). Conversely, in the boosting-based method, models are trained sequentially, where subsequent models focus on previous mis-classified samples [56]. Finally, an additional meta-learner can be trained to optimally combine the predictions made by base models in the stacking-based method [57].
From the perspective of a deep learning pipeline, an ensemble approach also can be categorized based on if the ensemble is made across multiple models or within a single model [54]. In the former case, multiple and often independent deep learning models with different model architectures, image preprocessing steps, and pretrained weights are trained and aggregated. Sometimes, each individual model can be trained on a particular subset of training dataset to increase the model diversity. Conversely, the ensemble within a single model is generally achieved by implicit ensembles where a set of neurons, layers, and blocks in the network is deactivated randomly.
In this study, we propose two ensemble pipelines for the classification of cancer cells: (i) a single-architecture and (ii) a multi-architecture approach. In the single-architecture approach, a set of the same CNN models trained with different strategies (i.e., different hyper-parameters) is utilized as illustrated in Figure 5a. For example, the MobileNet-based ensemble is composed of a set of MobileNet models trained with different hyper-parameters. Given the test sample, individual MobileNet networks compute their own probabilities for the test sample, and then a voting step is performed to make the final prediction. The multi-architecture ensemble approach is similar to the single-architecture ensemble approach, except that a set of different CNN architectures are included in the ensemble. As depicted in Figure 5b, the class probabilities from different CNN architectures are aggregated for voting. The final prediction is made based on soft voting, where the result is computed by the class probabilities from individual networks. Therefore, our ensemble approach can be considered a kind of bagging ensemble across multiple independent models with soft voting. Additionally, various ensemble configurations are considered to determine the optimal networks to be included in the ensemble. The network selection rule for each ensemble approach is described below:
-
Single-architecture ensemble (single-arch, hereafter): As shown in Table 1, there are 16 available configurations for each CNN architecture. In this approach, we select the top-4, top-8, and top-16 best-performing configurations in terms of classification accuracy. Therefore, we can build three ensembles for each model, for a total of 15 single-arch ensemble prediction pipelines.
-
Multi-architecture ensemble (multi-arch, hereafter): In contrast to the single-arch pipeline, the multi-arch approach is composed of heterogeneous CNN architectures. To establish this pipeline, we select the top-1, top-2, and top-3 best-performing configurations from each model. Therefore, top-1, top-2, and top-3 multi-arch ensemble pipelines include 5, 10, and 15 individual classification models from different architectures, respectively.

3. Experimental Results

3.1. Experimental setup

All the experiments were conducted using a GPU server with two NVIDIA RTX 3090 GPUs, 128 GB RAM, and an Intel i9-10940X CPU. We used Tensorflow framework with Keras backend for the training and evaluation of the CNNs. The experiments were conducted using fivefold cross-validation to report precision, recall, accuracy, and F1-score.

3.2. Performance evaluation

Table 2 summarizes the performance of the classification models in terms of effectiveness. Note that the reported values are from the best-performing configuration of each CNN model (Table 3). More details of the performance evaluation of all the configurations of each model can be found in Table S2 in the Supplementary Data. In addition to CNNs, we report the performance of traditional machine learning algorithms, such as Support Vector Machine (SVM), Random Forest (RF), Linear Discriminant Analysis (LDA), and K-Nearest Neighbor (k-NN). Similar to the methods proposed in previous studies [58,59], traditional machine learning algorithms used in our experiment were trained with conventional visual features, such as histograms of gradients (HOG), extracted from each cell image separately. The experiments with traditional machine learning algorithms were also conducted using fivefold cross-validation to report precision, recall, accuracy, and F1-score.
First, the results in Table 2 clearly demonstrate that traditional machine learning approaches fail to achieve superior performance. Specifically, machine learning methods showed an average accuracy of 49.39%. Conversely, CNNs achieved significant performance gain when compared to the machine learning methods, yielding up to 97.735% classification accuracy (from multi-arch ensemble with the top-3 configuration). Moreover, it is evident that both the single-arch (avg. 96.868%) and multi-arch (avg. 97.657%) ensemble approaches outperformed individual CNN models (avg. 96.071%) in terms of accuracy (p < 0.001, Wilcoxon Signed-rank test). Among the ensemble approaches, multi-arch approaches performed better than single-arch approaches, with a performance gain of 0.789%p on average. In the case of individual CNN models, DenseNet121 outperformed the other models with an average performance improvement of 0.844%p in terms of accuracy. Furthermore, the DenseNet121-based single-arch ensemble approach also produced the best result with an accuracy of 97.64%, beating other single-arch models. Finally, Figure 6 and Figure 7 represent the classification accuracy and loss per epoch during training and testing, respectively. As shown in the figures, the validation accuracy and loss of DenseNet121 and ResNet50 converged within 20 epochs, yielding stable performances much earlier than the other networks. The confusion matrices of each individual CNN model with the best-performing configuration are presented in Figure S1 in the Supplementary Data.
Next, we present the number of trainable parameters for each CNN architecture used in the experiments. As shown in Table 4, InceptionV3 and ResNet50 are the heaviest ones with 21–23 M parameters to be updated. In contrast, MobileNetV2 has the smallest number of trainable parameters (∼2.2 M), while DenseNet121 and EfficientNetB2 have 6.7 M and 7.7 M parameters, respectively. Taking into account the number of trainable parameters and classification accuracy, it can be inferred that DenseNet121 would be the best choice for a single CNN model considering that it can provide both moderate model size as well as high effectiveness.

4. Discussion

4.1. Performance of Deep-Learning-Based Approaches

First, we discuss the effectiveness of each model for automatic cancer cell taxonomy. As summarized in Table 2, all the traditional machine learning approaches failed to achieve superior performance in terms of all the metrics. Specifically, the machine learning methods showed an average accuracy of 49.39%, which is not practical for real-world situations. The SVM classifier yielded the best accuracy of at most 58.7%, which reveals a significant gap between the ML approach and deep learning approaches. Considering that traditional approaches generally utilize a set of classic hand-crafted features, their low performance implies that they are no longer cost-effective. Table 2 also shows that the introduction of deep learning approaches resulted in a significant performance improvement compared to the traditional methods. Moreover, we could observe that the proposed ensemble approach was more effective than individual CNN models for the classification of cancer cells, which was statistically significant (p < 0.001). These results imply that CNN models can be effectively applied to the domain of cancer cell microscopic images and can deliver superior performance in the classification of cell types.
On the other hand, Table 2 and Table 4 suggest interesting points regarding the relationship between the classification performance and the model size. It is worth noting that the number of trainable parameters did not significantly affect classification accuracy in the case of our domain. For example, the performance of MobileNetV2 with 2 M parameters and InceptionV3 with 21 M parameters did not show a significant difference (i.e., 95.412% from MobileNetV2 and 95.57% from InceptionV3). Moreover, single-arch ensemble approaches based on MobileNetV2 and InceptionV3 yielded similar classification accuracies of 96.533% and 96.342%, respectively.

4.2. Network Design Choice

In this section, we discuss how the different design choices of each hyper-parameter affect the overall performance in terms of the accuracy of the individual CNN models. The statistical significance based on Wilcoxon Signed-Rank test for each network design choice is depicted in Figure 8 with star marks (* (p < 0.05), ** (p< 0.01), and *** (p< 0.001)).
Optimizer: First, the difference in classification accuracy between the model with the SGD optimizer and the model with the AdaGrad optimizer is presented. As shown in Figure 8A, an optimal choice that worked best for all networks was non-existent. Regardless of the optimizer used, DenseNet121 and InceptionV3 performed equivalently. MobileNetV2 performed better with the SGD optimizer, while EfficientNetB2 and ResNet50 benefited from the use of the AdaGrad optimizer. Therefore, in this domain, an optimizer should be considered based on the type of CNN architecture.
Data augmentation: Second, the effects of the use of data augmentation on the overall performance are presented. In contrast to the use of optimizers, the use of data augmentation significantly affects the overall performance of individual CNN models. As shown in Figure 8B, it is obvious that applying data augmentation improves the classification accuracy of all types of networks (p < 0.001). Specifically, the networks with data augmentation achieved an average of 2.85%p higher classification accuracy when compared to those without data augmentation.
Learning rate scheduler: Third, the possible effects of the use of a learning rate scheduler on the performance are discussed. As presented in Figure 8C, it is clear that there is no significant difference in performance between the models with and without the learning rate schedulers. The results indicate that (i) the use of a learning rate scheduler does not significantly affect the performance and (ii) the default choice (0.001) is adequate to achieve high performance.
Fine-tuning: Next, the performance difference between the models trained by updating all weights and the models trained by freezing the first 25% layer and updating just the rest is examined. As shown in Figure 8D, DenseNet121 (p < 0.001), InceptionV3 (p < 0.001), and ResNet50 (p < 0.05) showed significant differences while the performances of MobileNetV2 and EfficientNetB2 were not affected by the degree of fine-tuning.
Ensemble: Finally, the possible effects of selection criterion of the ensemble pipeline on the performance of the ensemble prediction are discussed. We first show the difference in performance of the single-arch ensemble approach according to the ensemble configuration. As mentioned in Section 2.4, the single-arch ensemble pipeline can be built using the top-4, top-8, and top-16 models from the same CNN architecture. Figure 9A summarizes the effect of the ensemble configuration on the classification accuracy of the single-arch approach. The result implies that the performance degrades when more networks are involved. Every network showed a similar pattern, where the top-4 or top-8 configuration resulted in the best performance. Basically, the diversity of each base model is important to establish a successful ensemble pipeline. In the case of the single-model approach, the diversity of the base model is relatively low, even though we applied different training strategies, because the base architecture is the same. Therefore, adding more models in this case just resulted in the inclusion of poor models (low-ranked ones), thereby adversely affecting the overall performance. More details on the classification accuracy of each single-arch ensemble approach are presented in Table S3 in the Supplementary Data. In contrast, the multi-arch ensemble pipeline can be built using the top-1, top-2, and top-3 configurations from all types of CNN architectures. Figure 9B shows that the performance of the multi-arch ensemble approach improves as more networks are included in the ensemble. In contrast to the single-arch ensemble approach, the diversity of the models included in the multi-model ensemble is relatively high because their base network architecture and training strategies are totally different. By adding more models in this case, we can include top-performing models with different architectures, thereby increasing the model diversities of the ensemble which can contribute to performance improvement. Finally, Table S4 in the Supplementary Data summarizes a fold-wise classification accuracy of the multi-arch ensemble approach.

4.3. Comparison with Previous Studies

Representative CNN studies related to the classification of cancer cells are summarized in Table 5, which shows that our proposed method may provide advantages over the above-mentioned studies. Since the Papanicolaou (Pap) smear test is one of the most essential screening methods for cervical cancer detection [60], it commonly appeared in datasets in the related research [58,61,62,63,64,65]. Despite this popularity, the image acquisition procedure of a Pap smear or a Hematoxylin and Eosin (H&E) stained sample is a labor-intensive and time-consuming process which relies on expert cytologists [58,59,61,62,63,64,65]. In addition, expensive and specialized equipment such as low-coherence off-axis holography [33] or confocal immunofluorescence microscopy [38] is often used to acquire the images, but there is a lack of sufficient datasets to be faced. On the other hand, the strengths of our proposed method include the use of bright field images of the cancer cells from cell culture flasks obtained through the low-scale benchtop optical microscopy that is typically used in laboratories. The other advantage of our method is that it requires no additional wet bench work using fluorescent/staining dyes or biochemical markers. Since the annotated cancer cell lines used in this study were provided directly from cell line provider Korean Cell Line Bank (Seoul, Korea) and cultured in different flasks for each cell line, the training dataset for each cell line serves as the ground truth. Finally, a relatively simple and fast preparation procedure enables researchers to create a large number of datasets for multiple cancer cell lines in their own use. Quantitatively, among the related studies that used the specialized imaging systems, Rubin et al. [33] obtained a maximum accuracy of 90–99% and Oei et al. [38] attained an accuracy of 97.2%. Other studies based on images with staining [58,61,62,63,64,65] reported accuracies of 82.9–96.73%.
Our proposed method achieved a test accuracy of 97.735%, a precision of 97.74, and a recall of 97.74. From these comparisons, it can be inferred that our proposed method outperforms the other classification of cancer cells studies, even though the prepared cancer cell images used in training and evaluation steps require no additional biochemical staining process or expensive image acquisition system compared with these previous studies.

5. Conclusions

In this paper, we presented deep-learning-based approaches for the classification of the type of microscopic cancer cell images. We constructed a framework to exploit individual and ensemble CNN pipelines to solve a four-class classification task. The experimental results validated the feasibility of the proposed approach. Specifically, all the CNN models achieved a high classification accuracy of 96.07 (±0.58)%, outperforming traditional machine learning classifiers. In particular, the ensemble approach with a multi-arch strategy achieved the best results, with an accuracy of 97.735%, validating the feasibility of the proposed framework. Moreover, our experimental results indicate that the network design choice and ensemble configuration can affect the overall performance. The results indicated that (i) AdaGrad optimizer is helpful to boost up the performance of EfficientNet-B2 (p < 0.01) and RestNet-50 (p < 0.01), (ii) data augmentation is always useful for all the networks (p < 0.001), (iii) the use of a learning rate scheduler does not make a significant performance difference, and (iv) only DenseNet121 and InceptionV3 benefit from the fine-tuning of all weights rather than freezing part of a network (p < 0.001). Based on the experimental results, we believe that the proposed method can reduce the cost of identifying cancer cells, and even users without expertise can identify cell types. Furthermore, our approach does not require expensive equipment and can identify cross-infection among cancer cells using low-scale benchtop microscopy without any additional bench work.
However, additional studies are still required to overcome the limitations of our current approach.
First, four types of cancer cell lines with high mortality were selected to perform label-free cell classification in this study. The annotated cancer cell lines used in this study were provided directly from the official cell line provider, and the same type of cancer cell line was cultured in the individual flask. In other words, pathologically trained experts are not required for validating the test dataset and additional wet bench work to classify cell types. Thus, a relatively simple and fast validation procedure enables us to shorten the preparation time and provide a cost-effective analysis method. On the contrary, randomly mixed cancer cell lines in a single flask may be considered a more realistic model, and it increases the role of the pathologist to validate or identify cell types through fluorescent staining or an H&E staining procedure. Therefore, we plan to apply the proposed framework to mixed-cell images obtained from a single culture flask to provide more practical solutions.
Second, more advanced classification and prediction methods will be required to address various clinical tasks under the aforementioned environments. For example, a transformer architecture, which was very effective for natural language processing tasks, is now widely applied to the computer vision tasks due to its robust and scalable learning capabilities [66,67,68]. In addition, researchers have recently proposed various applications based on self-supervised learning techniques in the computer vision domain and demonstrated effective learning of underlying image representations [69,70,71]. It is also expected that adopting the recent advances in deep learning for computer vision tasks will help in addressing various challenging tasks in the medical domain.
Finally, even though we achieved a higher classification accuracy using ensembles of multiple deep learning architectures with different training strategies, the computational and storage cost required for our models could be another kind of burden for practical use. Therefore, our future work will also focus on improving the computational efficiency as well as classification accuracy by adopting the recent advances in deep learning techniques, for example, knowledge distillation [72,73] from multiple teachers.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers14092224/s1, Table S1: Pseudo code of the image preprocessing step, Table S2: Fivefold cross-validation accuracy of each CNN model with each network configuration, Table S3: Classification accuracy of each single-arch ensemble approach, Table S4: Classification accuracy of the multi-arch ensemble approach, Figure S1: Confusion matrix of each individual CNN model with the best-performing configuration.

Author Contributions

Conceptualization, S.-w.C. and J.-W.J.; methodology, S.-w.C., H.-Y.Y., J.-Y.J., and J.-W.J.; software, S.-w.C., H.-Y.Y., J.-Y.J., J.P. and J.-W.J.; validation, S.-w.C., H.-Y.Y., J.-Y.J., J.P. and J.-W.J.; investigation, S.-w.C., H.-Y.Y., J.-Y.J. and J.-W.J.; resources, S.-w.C.; data curation, S.-w.C., H.-Y.Y., J.-Y.J., J.P. and J.-W.J.; writing—original draft preparation, S.-w.C. and J.-W.J.; writing—review and editing, S.-w.C. and J.-W.J.; visualization, S.-w.C., H.-Y.Y., J.-Y.J. and J.-W.J.; supervision, S.-w.C. and J.-W.J.; project administration, S.-w.C. and J.-W.J.; funding acquisition, S.-w.C. and J.-W.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) [NRF-2019R1F1A1062397, NRF-2021R1F1A1059665] and Brain Korea 21 FOUR Project (Dept. of IT Convergence Engineering at Kumoh National Institute of Technology, Dept. of Data Science at Seoul National University of Science and Technology).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Siegel, R.L.; Miller, K.D.; Jemal, A. Cancer statistics, 2020. CA Cancer J. Clin. 2020, 70, 7–30. [Google Scholar] [CrossRef] [PubMed]
  2. Bray, F.; Ferlay, J.; Soerjomataram, I.; Siegel, R.L.; Torre, L.A.; Jemal, A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2018, 68, 394–424. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. World Health Organization. WHO Position Paper on Mammography Screening; World Health Organization: Geneva, Switzerland, 2014. [Google Scholar]
  4. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [CrossRef] [PubMed]
  5. Kainz, P.; Pfeiffer, M.; Urschler, M. Segmentation and classification of colon glands with deep convolutional neural networks and total variation regularization. PeerJ 2017, 2017, e3874. [Google Scholar] [CrossRef]
  6. Aubreville, M.; Knipfer, C.; Oetter, N.; Jaremenko, C.; Rodner, E.; Denzler, J.; Bohr, C.; Neumann, H.; Stelzle, F.; Maier, A. Automatic Classification of Cancerous Tissue in Laserendomicroscopy Images of the Oral Cavity using Deep Learning. Sci. Rep. 2017, 7, 11979. [Google Scholar] [CrossRef] [Green Version]
  7. Teramoto, A.; Tsukamoto, T.; Kiriyama, Y.; Fujita, H. Automated Classification of Lung Cancer Types from Cytological Images Using Deep Convolutional Neural Networks. BioMed Res. Int. 2017, 2017. [Google Scholar] [CrossRef]
  8. Khan, S.U.; Islam, N.; Jan, Z.; Ud Din, I.; Rodrigues, J.J. A novel deep learning based framework for the detection and classification of breast cancer using transfer learning. Pattern Recognit. Lett. 2019, 125, 1–6. [Google Scholar] [CrossRef]
  9. Chekkoury, A.; Khurd, P.; Ni, J.; Bahlmann, C.; Kamen, A.; Patel, A.; Grady, L.; Singh, M.; Groher, M.; Navab, N.; et al. Automated Malignancy Detection in Breast Histopathological Images; Medical Imaging 2012: Computer-Aided Diagnosis; van Ginneken, B., Novak, C.L., Eds.; SPIE: San Diego, CA, USA, 2012; Volume 8315, p. 831515. [Google Scholar] [CrossRef]
  10. Pollanen, I.; Braithwaite, B.; Ikonen, T.; Niska, H.; Haataja, K.; Toivanen, P.; Tolonen, T. Computer-aided breast cancer histopathological diagnosis: Comparative analysis of three DTOCS-based features: SW-DTOCS, SW-WDTOCS and SW-3-4-DTOCS. In Proceedings of the 2014 4th International Conference on Image Processing Theory, Tools and Applications, IPTA, Paris, France, 14–17 October 2014. [Google Scholar] [CrossRef]
  11. Toratani, M.; Konno, M.; Asai, A.; Koseki, J.; Kawamoto, K.; Tamari, K.; Li, Z.; Sakai, D.; Kudo, T.; Satoh, T.; et al. A convolutional neural network uses microscopic images to differentiate between mouse and human cell lines and their radioresistant clones. Cancer Res. 2018, 78, 6703–6707. [Google Scholar] [CrossRef] [Green Version]
  12. Xing, F.; Yang, L. Robust nucleus/cell detection and segmentation in digital pathology and microscopy images: A comprehensive review. IEEE Rev. Biomed. Eng. 2016, 9, 234–263. [Google Scholar] [CrossRef]
  13. Xing, F.; Xie, Y.; Su, H.; Liu, F.; Yang, L. Deep Learning in Microscopy Image Analysis: A Survey. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 4550–4568. [Google Scholar] [CrossRef]
  14. Irshad, H.; Veillard, A.; Roux, L.; Racoceanu, D. Methods for nuclei detection, segmentation, and classification in digital histopathology: A review-current status and future potential. IEEE Rev. Biomed. Eng. 2014, 7, 97–114. [Google Scholar] [CrossRef] [PubMed]
  15. Hu, C.; He, S.; Lee, Y.J.; He, Y.; Kong, E.M.; Li, H.; Anastasio, M.A.; Popescu, G.; Anastasio, M. Live-dead assay on unlabeled cells using phase imaging with computational specificity. Nat. Commun. 2022, 13, 1–8. [Google Scholar] [CrossRef] [PubMed]
  16. Xie, W.; Noble, J.A.; Zisserman, A. Microscopy cell counting and detection with fully convolutional regression networks. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2018, 6, 283–292. [Google Scholar] [CrossRef]
  17. Wainberg, M.; Merico, D.; Delong, A.; Frey, B.J. Deep learning in biomedicine. Nat. Biotechnol. 2018, 36, 829–838. [Google Scholar] [CrossRef]
  18. O’Connor, T.; Hawxhurst, C.; Shor, L.M.; Javidi, B. Red blood cell classification in lensless single random phase encoding using convolutional neural networks. Opt. Express 2020, 28, 33504. [Google Scholar] [CrossRef]
  19. Coates, A.S.; Winer, E.P.; Goldhirsch, A.; Gelber, R.D.; Gnant, M.; Piccart-Gebhart, M.J.; Thürlimann, B.; Senn, H.J.; André, F.; Baselga, J.; et al. Tailoring therapies-improving the management of early breast cancer: St Gallen International Expert Consensus on the Primary Therapy of Early Breast Cancer 2015. Ann. Oncol. 2015, 26, 1533–1546. [Google Scholar] [CrossRef]
  20. Solnica-Krezel, L. Conserved patterns of cell movements during vertebrate gastrulation. Curr. Biol. 2005, 15, R213–R228. [Google Scholar] [CrossRef] [Green Version]
  21. Gartler, S.M. Apparent HeLa cell contamination of human heteroploid cell lines. Nature 1968, 217, 750–751. [Google Scholar] [CrossRef]
  22. Lande, R. Natural Selection and Random Genetic Drift in Phenotypic Evolution. Evolution 1976, 30, 314. [Google Scholar] [CrossRef]
  23. Capes-Davis, A.; Theodosopoulos, G.; Atkin, I.; Drexler, H.G.; Kohara, A.; MacLeod, R.A.; Masters, J.R.; Nakamura, Y.; Reid, Y.A.; Reddel, R.R.; et al. Check your cultures! A list of cross-contaminated or misidentified cell lines. Int. J. Cancer 2010, 127, 1–8. [Google Scholar] [CrossRef]
  24. Neimark, J. Line of attack. Science 2015, 347, 938–940. [Google Scholar] [CrossRef] [PubMed]
  25. Wang, R.; Wang, D.; Kang, D.; Guo, X.; Guo, C.; Dongye, M.; Zhu, Y.; Chen, C.; Zhang, X.; Long, E.; et al. An artificial intelligent platform for live cell identification and the detection of cross-contamination. Ann. Transl. Med. 2020, 8, 697. [Google Scholar] [CrossRef] [PubMed]
  26. Lorsch, B.J.R.; Collins, F.S.; Lippincott-schwartz, J. Fixing problems with cell lines. Science 2014, 346, 1452–1453. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Masters, J.R. Cell-line authentication: End the scandal of false cell lines. Nature 2012, 492, 186. [Google Scholar] [CrossRef]
  28. Bian, X.; Yang, Z.; Feng, H.; Sun, H.; Liu, Y. A Combination of Species Identification and STR Profiling Identifies Cross-contaminated Cells from 482 Human Tumor Cell Lines. Sci. Rep. 2017, 7, 9774. [Google Scholar] [CrossRef] [Green Version]
  29. Almeida, J.L.; Cole, K.D.; Plant, A.L. Standards for Cell Line Authentication and Beyond. PLoS Biol. 2016, 14, e1002476. [Google Scholar] [CrossRef]
  30. Masters, J.R.; Thomson, J.A.; Daly-Burns, B.; Reid, Y.A.; Dirks, W.G.; Packer, P.; Toji, L.H.; Ohno, T.; Tanabe, H.; Arlett, C.F.; et al. Short tandem repeat profiling provides an international reference standard for human cell lines. Proc. Natl. Acad. Sci. USA 2001, 98, 8012–8017. [Google Scholar] [CrossRef] [Green Version]
  31. Poetsch, M.; Petersmann, A.; Woenckhaus, C.; Protzel, C.; Dittberner, T.; Lignitz, E.; Kleist, B. Evaluation of allelic alterations in short tandem repeats in different kinds of solid tumors - Possible pitfalls in forensic casework. Forensic Sci. Int. 2004, 145, 1–6. [Google Scholar] [CrossRef]
  32. Lohar, P.S. Textbook of Biotechnology; MJP Publisher: Chennai, India, 2019. [Google Scholar]
  33. Rubin, M.; Stein, O.; Turko, N.A.; Nygate, Y.; Roitshtain, D.; Karako, L.; Barnea, I.; Giryes, R.; Shaked, N.T. TOP-GAN: Stain-free cancer cell classification using deep learning with a small training set. Med. Image Anal. 2019, 57, 176–185. [Google Scholar] [CrossRef] [Green Version]
  34. Cruz-Roa, A.; Gilmore, H.; Basavanhally, A.; Feldman, M.; Ganesan, S.; Shih, N.N.; Tomaszewski, J.; González, F.A.; Madabhushi, A. Accurate and reproducible invasive breast cancer detection in whole-slide images: A Deep Learning approach for quantifying tumor extent. Sci. Rep. 2017, 7, 46450. [Google Scholar] [CrossRef] [Green Version]
  35. Wang, J.; Yang, X.; Cai, H.; Tan, W.; Jin, C.; Li, L. Discrimination of Breast Cancer with Microcalcifications on Mammography by Deep Learning. Sci. Rep. 2016, 6, 27327. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Ayana, G.; Dese, K.; Choe, S.W. Transfer Learning in Breast Cancer Diagnoses via Ultrasound Imaging. Cancers 2021, 13, 738. [Google Scholar] [CrossRef] [PubMed]
  37. Meng, N.; Lam, E.Y.; Tsia, K.K.; So, H.K.H. Large-Scale Multi-Class Image-Based Cell Classification with Deep Learning. IEEE J. Biomed. Health Inform. 2019, 23, 2091–2098. [Google Scholar] [CrossRef] [PubMed]
  38. Oei, R.W.; Hou, G.; Liu, F.; Zhong, J.; Zhang, J.; An, Z.; Xu, L.; Yang, Y. Convolutional neural network for cell classification using microscope images of intracellular actin networks. PLoS ONE 2019, 14, e0213626. [Google Scholar] [CrossRef] [PubMed]
  39. Choe, S.W.; Terman, D.S.; Rivers, A.E.; Rivera, J.; Lottenberg, R.; Sorg, B.S. Drug-loaded sickle cells programmed ex vivo for delayed hemolysis target hypoxic tumor microvessels and augment tumor drug delivery. J. Control. Release 2013, 171, 184–192. [Google Scholar] [CrossRef] [Green Version]
  40. Cho, K.; Seo, J.H.; Heo, G.; Choe, S.W. An Alternative Approach to Detecting Cancer Cells by Multi-Directional Fluorescence Detection System Using Cost-Effective LED and Photodiode. Sensors 2019, 19, 2301. [Google Scholar] [CrossRef] [Green Version]
  41. Nelissen, B.G.L.; van Herwaarden, J.A.; Moll, F.L.; van Diest, P.J.; Pasterkamp, G. SlideToolkit: An Assistive Toolset for the Histological Quantification of Whole Slide Images. PLoS ONE 2014, 9, e110289. [Google Scholar] [CrossRef]
  42. Choe, S.W.; Choi, H. Suppression technique of hela cell proliferation using ultrasonic power amplifiers integrated with a series-diode linearizer. Sensors 2018, 18, 4248. [Google Scholar] [CrossRef] [Green Version]
  43. Choi, H.; Ryu, J.-M.; Choe, S.-W. A novel therapeutic instrument using an ultrasound-light-emitting diode with an adjustable telephoto lens for suppression of tumor cell proliferation. Measurement 2019, 147, 106865. [Google Scholar] [CrossRef]
  44. Tan, C.; Sun, F.; Kong, T.; Zhang, W.; Yang, C.; Liu, C. A survey on deep transfer learning. In Proceedings of the International Conference on Artificial Neural Networks, Rhodes, Greece, 4–7 October, 2018; Springer: Rhodes, Greece, 2018; Volume 11141, pp. 270–279. [Google Scholar] [CrossRef] [Green Version]
  45. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef] [Green Version]
  46. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef] [Green Version]
  47. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 July 2018; pp. 4510–4520. [Google Scholar] [CrossRef] [Green Version]
  48. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019. [Google Scholar]
  49. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2818–2826. [Google Scholar] [CrossRef] [Green Version]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. IEEE Computer Society, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  51. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  52. Ruder, S. An overview of gradient descent optimization algorithms. arXiv 2016, arXiv:1609.04747. [Google Scholar]
  53. Duchi, J.; Hazan, E.; Singer, Y. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  54. Cao, Y.; Geddes, T.A.; Yang, J.Y.H.; Yang, P. Ensemble deep learning in bioinformatics. Nat. Mach. Intell. 2020, 2, 500–508. [Google Scholar] [CrossRef]
  55. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  56. Bartlett, P.; Freund, Y.; Lee, W.S.; Schapire, R.E. Boosting the margin: A new explanation for the effectiveness of voting methods. Ann. Stat. 1998, 26, 1651–1686. [Google Scholar] [CrossRef]
  57. David, H.W. Stacked generalization. Neural Netw. 1992, 5, 241–259. [Google Scholar] [CrossRef]
  58. Sophea, P.; Handayani, D.O.D.; Boursier, P. Abnormal cervical cell detection using hog descriptor and SVM classifier. In Proceedings of the 2018 Fourth International Conference on Advances in Computing, Communication & Automation (ICACCA), Subang Jaya, Malaysia, 26–28 October 2018; pp. 1–6. [Google Scholar]
  59. Kumar, R.; Srivastava, R.; Srivastava, S. Detection and classification of cancer from microscopic biopsy images using clinically significant and biologically interpretable features. J. Med. Eng. 2015, 2015, 457906. [Google Scholar] [CrossRef] [PubMed]
  60. Follen, M.; Richards-Kortum, R. Emerging Technologies and Cervical Cancer. JNCI J. Natl. Cancer Inst. 2000, 92, 363–365. [Google Scholar] [CrossRef] [Green Version]
  61. Shi, J.; Wang, R.; Zheng, Y.; Jiang, Z.; Zhang, H.; Yu, L. Cervical cell classification with graph convolutional network. Comput. Methods Programs Biomed. 2021, 198, 105807. [Google Scholar] [CrossRef]
  62. Chankong, T.; Theera-Umpon, N.; Auephanwiriyakul, S. Automatic cervical cell segmentation and classification in Pap smears. Comput. Methods Programs Biomed. 2014, 113, 539–556. [Google Scholar] [CrossRef]
  63. Sharma, M.; Singh, S.K.; Agrawal, P.; Madaan, V. Classification of clinical dataset of cervical cancer using KNN. Indian J. Sci. Technol. 2016, 9, 1–5. [Google Scholar] [CrossRef] [Green Version]
  64. Gençtav, A.; Aksoy, S.; Önder, S. Unsupervised segmentation and classification of cervical cell images. Pattern Recognit. 2012, 45, 4151–4168. [Google Scholar] [CrossRef] [Green Version]
  65. Marinakis, Y.; Dounias, G.; Jantzen, J. Pap smear diagnosis using a hybrid intelligent scheme focusing on genetic algorithm based feature selection and nearest neighbor classification. Comput. Biol. Med. 2009, 39, 69–78. [Google Scholar] [CrossRef] [PubMed]
  66. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  67. Linh, T.D.; Nhi, H.L.; Toan, B.T.; Vuong, M.N.; Phuong, T.N. Detection of tuberculosis from chest X-ray images: Boosting the performance with vision transformer and transfer learning. Expert Syst. Appl. 2021, 184, 115519. [Google Scholar] [CrossRef]
  68. Beal, J.; Kim, E.; Tzeng, E.; Park, D.H.; Zhai, A.; Kislyuk, D. Toward transformer-based object detection. arXiv 2020, arXiv:2012.09958. [Google Scholar]
  69. Ayush, K.; Uzkent, B.; Meng, C.; Tanmay, K.; Burke, M.; Lobell, D.; Ermon, S. Geography-Aware Self-Supervised Learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 10181–10190. [Google Scholar]
  70. Yan, K.; Cai, J.; Jin, D.; Miao, S.; Harrison, A.P.; Guo, D.; Tang, Y.; Xiao, J.; Lu, J.; Lu, L. Self-supervised learning of pixel-wise anatomical embeddings in radiological images. arXiv 2020, arXiv:2012.02383. [Google Scholar] [CrossRef]
  71. Lin, L.; Song, S.; Yang, W.; Liu, J. Ms2l: Multi-task self-supervised learning for skeleton based action recognition. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2490–2498. [Google Scholar]
  72. Hinton, G.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  73. Lan, X.; Zhu, X.; Gong, S. Knowledge Distillation by On-the-Fly Native Ensemble. In Advances in Neural Information Processing Systems; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
Figure 1. Workflow of the proposed approach.
Figure 1. Workflow of the proposed approach.
Cancers 14 02224 g001
Figure 2. Image preprocessing step. (a) Captured microscope image, (b) grayscale image, (c) noise removed image, (d) identified cell contour, (e) segmented image patches.
Figure 2. Image preprocessing step. (a) Captured microscope image, (b) grayscale image, (c) noise removed image, (d) identified cell contour, (e) segmented image patches.
Cancers 14 02224 g002
Figure 3. Example of data augmentation: (A) original, (B) rotation, (C) translation, (D) vertical flip, (E) all.
Figure 3. Example of data augmentation: (A) original, (B) rotation, (C) translation, (D) vertical flip, (E) all.
Cancers 14 02224 g003
Figure 4. Degree of fine-tuning.
Figure 4. Degree of fine-tuning.
Cancers 14 02224 g004
Figure 5. Overview of ensemble approaches.
Figure 5. Overview of ensemble approaches.
Cancers 14 02224 g005
Figure 6. Classification accuracy per epoch: (A) training accuracy, (B) validation accuracy.
Figure 6. Classification accuracy per epoch: (A) training accuracy, (B) validation accuracy.
Cancers 14 02224 g006
Figure 7. Loss per epoch: (A) training loss, (B) validation loss.
Figure 7. Loss per epoch: (A) training loss, (B) validation loss.
Cancers 14 02224 g007
Figure 8. Network design choices (the statistical significance is represented using * (p < 0.05), ** (p < 0.01), and *** (p < 0.001)): (A) optimizer, (B) data augmentation, (C) learning rate scheduler, (D) degree of fine-tuning.
Figure 8. Network design choices (the statistical significance is represented using * (p < 0.05), ** (p < 0.01), and *** (p < 0.001)): (A) optimizer, (B) data augmentation, (C) learning rate scheduler, (D) degree of fine-tuning.
Cancers 14 02224 g008
Figure 9. Performance change according to the ensemble configuration: (A) single-arch ensemble, (B) multi-arch ensemble.
Figure 9. Performance change according to the ensemble configuration: (A) single-arch ensemble, (B) multi-arch ensemble.
Cancers 14 02224 g009
Table 1. Summary of hyper-parameters.
Table 1. Summary of hyper-parameters.
ParameterOptionNote
Data augmentationORotation, translation, and vertical flip
XWithout any augmentation
Fine-tuningWithout freezeAll weights are updated
25% freezeOnly 75% of weights are updated
OptimizerSGDStochastic gradient descent
AdaGradAdaptive gradient-based optimization
Learning rate schedulerOExponential decay
XLearning rate is fixed to 0.001
Table 2. Comparison of the model performance (Best scores in each algorithm are marked in bold).
Table 2. Comparison of the model performance (Best scores in each algorithm are marked in bold).
AlgorithmModelAccuracyPrecisionRecallF1-Score
Machine LearningSVM58.7 ± 0.7458.34 ± 0.7658.7 ± 0.7458.52 ± 0.75
RF49.55 ± 0.3249.01 ± 0.3349.55 ± 0.3249.3 ± 0.32
LDA46.26 ± 0.9844.81 ± 0.9845.26 ± 0.9845.03 ± 0.98
KNN44.05 ± 0.9245.86 ± 1.144.05 ± 0.9244.93 ± 0.94
Average49.39 ± 5.9449.51 ± 5.5249.39 ± 5.9449.44 ± 5.72
Deep LearningDenseNet12196.915 ± 0.07296.916 ± 0.07796.915 ± 0.07296.915 ± 0.075
EfficientNetB296.195 ± 0.2396.23 ± 0.27296.176 ± 0.19496.203 ± 0.232
ResNet5096.265 ± 0.13896.274 ± 0.1396.265 ± 0.13896.269 ± 0.134
InceptionV395.57 ± 0.32295.604 ± 0.37695.556 ± 0.29895.58 ± 0.336
MobileNetV295.412 ± 0.22395.446 ± 0.22995.412 ± 0.22495.429 ± 0.226
Average96.071 ± 0.58496.1 ± 0.5896.06 ± 0.58196.08 ± 0.58
Ensemble (Single-architecture)DenseNet12197.64 ± 0.1697.643 ± 0.1697.64 ± 0.1697.641 ± 0.16
EfficientNetB296.757 ± 0.20296.763 ± 0.29496.757 ± 0.29496.76 ± 0.294
ResNet5097.066 ± 0.14897.073 ± 0.14597.066 ± 0.14896.07 ± 0.147
InceptionV396.342 ± 0.19696.345 ± 0.20296.342 ± 0.19696.343 ± 0.199
MobileNetV296.533 ± 0.20996.55 ± 0.22696.533 ± 0.20996.541 ± 0.217
Average96.868 ± 0.596.875 ± 0.596.868 ± 0.596.871 ± 0.5
Ensemble (Multi-architecture)Top-197.563 ± 0.14597.568 ± 0.14597.563 ± 0.14597.565 ± 0.145
Top-297.673 ± 0.12297.677 ± 0.12497.673 ± 0.12297.675 ± 0.123
Top-397.735 ± 0.13297.74 ± 0.1497.74 ± 0.13297.74 ± 0.134
Average97.657 ± 0.14497.661 ± 0.14997.657 ± 0.14497.659 ± 0.144
Table 3. Configuration of the best-performing individual deep learning models.
Table 3. Configuration of the best-performing individual deep learning models.
AlgorithmData AugmentationDegree of Fine-TuningOptimizerLearning Rate Scheduler
DenseNet121OAll weightsSGDX
EfficientNetB2OAll weightsAdaGradX
ResNet50OAll weightsAdaGradX
InceptionV3OAll weightsSGDO
MobileNetV2OFreeze the early 25% layersSGDO
Table 4. Number of trainable parameters.
Table 4. Number of trainable parameters.
Degree of Fine-TuningAll WeightsFreeze the First 25% Layers
Model
DenseNet1216,957,9566,716,740
MobileNetV22,228,9962,197,060
EfficientNetB27,706,6307,700,858
InceptionV321,776,54821,348,836
ResNet5023,542,78823,315,972
Table 5. Comparison with previous studies (“CNN” denotes “Convolutional Neural Network”, “GAN” denotes “Generative Adversarial Network”, “ML” denotes “Machine Learning”, “ANN” denotes “Artificial Neural Network”, “GA” denotes “Generic Algorithm”).
Table 5. Comparison with previous studies (“CNN” denotes “Convolutional Neural Network”, “GAN” denotes “Generative Adversarial Network”, “ML” denotes “Machine Learning”, “ANN” denotes “Artificial Neural Network”, “GA” denotes “Generic Algorithm”).
Ref.TaskImage AcquisitionMethodNum. of ClassesMetricPerformanceFeature
Rubin et al. [33]Cancer cell classificationLow-coherence off-axis holography without statiningGAN-based approach4 classes (healthy skin, melanoma cells, colorectal adenocarcinoma colon cells, metastatic colorectal adenocarcinoma cells)Accuracy90–99%CNN feature
Oei et al. [38]Breast cancer cell detectionConfocal immunofluorescence microscopy images with stainingCNN2 classes (breast normal cells and cancer cells)Accuracy97.2%CNN feature
Kumar et al. [59]Cervical cancer cell detectionMicroscopic biopsy images with stainingRF, SVM, KNN, fuzzy KNN2 classese (noncancerous, cancerous)Accuracy92.19%Texture features, morphology and shape features, HOG, wavelet features, etc.
Shi et al. [61]Cervical cancer cell classificationMicroscopic images of Pap smear slides with stainingGraph neural network5 types of cervical cancer cells (superficial–intermediate, parabasal, koilocytotic, dyskeratotic, and metaplastic cells)Accuracy94.93%CNN feature
Sophea et al. [58]HOG + SVM2 classes (normal and abnormal)Accuracy94.7%HOG
Chankong et al. [62]Bayes, LDA, KNN, ANN, SVM7 classes (superficial squamous, intermediate squamous, columnar, mild dysplasia, moderate dysplasia, severe dysplasia, and carcinoma in situ)Accuracy93.78%Hand-crafted features (area of cucleus, nucleus-to-cytoplasm ratio, etc.)
Sharma et al. [63]KNNAccuracy82.9%
Gençtav et al. [64]Bayesian, decision tree, SVMPrecision91.7%
Marinakis et al. [65]GAAccuracy96.73%
Our proposed methodCancer cell classificationMicroscopic images of cell culture flask without stainingCNN ensemble4 classes of cell culture flask (HeLa, MCF-7, Huh7, and NCI-H1299)Accuracy97.735%CNN feature
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Choe, S.-w.; Yoon, H.-Y.; Jeong, J.-Y.; Park, J.; Jeong, J.-W. Automatic Cancer Cell Taxonomy Using an Ensemble of Deep Neural Networks. Cancers 2022, 14, 2224. https://doi.org/10.3390/cancers14092224

AMA Style

Choe S-w, Yoon H-Y, Jeong J-Y, Park J, Jeong J-W. Automatic Cancer Cell Taxonomy Using an Ensemble of Deep Neural Networks. Cancers. 2022; 14(9):2224. https://doi.org/10.3390/cancers14092224

Chicago/Turabian Style

Choe, Se-woon, Ha-Yeong Yoon, Jae-Yeop Jeong, Jinhyung Park, and Jin-Woo Jeong. 2022. "Automatic Cancer Cell Taxonomy Using an Ensemble of Deep Neural Networks" Cancers 14, no. 9: 2224. https://doi.org/10.3390/cancers14092224

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop