Next Article in Journal
Small Bowel Endoscopic Features of Eosinophilic Gastroenteritis
Previous Article in Journal
Interpretable Machine Learning Techniques in ECG-Based Heart Disease Classification: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Bone Marrow Cell Classification for Haematological Disease Diagnosis Using Siamese Neural Network

1
Centre for Cyber Physical Systems, Vellore Institute of Technology, Chennai 600127, India
2
School of Computer Science and Engineering, Vellore Institute of Technology, Chennai 600127, India
3
School of Information and Data Sciences, Nagasaki University, Nagasaki 8528521, Japan
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(1), 112; https://doi.org/10.3390/diagnostics13010112
Submission received: 10 November 2022 / Revised: 10 December 2022 / Accepted: 13 December 2022 / Published: 29 December 2022
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
The critical structure and nature of different bone marrow cells which form a base in the diagnosis of haematological ailments requires a high-grade classification which is a very prolonged approach and accounts for human error if performed manually, even by field experts. Therefore, the aim of this research is to automate the process to study and accurately classify the structure of bone marrow cells which will help in the diagnosis of haematological ailments at a much faster and better rate. Various machine learning algorithms and models, such as CNN + SVM, CNN + XGB Boost and Siamese network, were trained and tested across a dataset of 170,000 expert-annotated cell images from 945 patients’ bone marrow smears with haematological disorders. The metrics used for evaluation of this research are accuracy of model, precision and recall of all the different classes of cells. Based on these performance metrics the CNN + SVM, CNN + XGB, resulted in 32%, 28% accuracy, respectively, and therefore these models were discarded. Siamese neural resulted in 91% accuracy and 84% validation accuracy. Moreover, the weighted average recall values of the Siamese neural network were 92% for training and 91% for validation. Hence, the final results are based on Siamese neural network model as it was outperforming all the other algorithms used in this research.

1. Introduction

The use of optical microscopes to evaluate and categorize samples of bone marrow cells is a century-old practice for diagnosing blood diseases. The strategy of searching for uncommon but diagnostically essential cells is well-established, but time-consuming and difficult. Highly skilled individual with long term experience is a necessity [1]. However, the conclusion of those findings is subjective, and there is no statistical standard in place yet. The increased density of contacting and overlapping cells, as well as the greater diversity and complexity of cell morphologies, make detection and categorization in BMA smears substantially more difficult. Without good cell identification and localisation of cell borders, classification is challenging, and the minor distinctions in cytologic properties required to differentiate various cell types found in bone marrow make it even more difficult. [2]. In the diagnosis of malignant and non-malignant illnesses of the hematopoietic system, the examination and distinction of bone marrow cell morphologies are critical [3].
Previous studies have been applied to the bone marrow problem, but they were all predicated on the premise that manually segmented pictures were accessible. Machine Learning techniques adopted for this purpose were mainly focused on cherry picking a particular subset of classes consisting of more data points, which may fail to classify those excluded data points [4,5] Other models which took the whole dataset in consideration were able to achieve state of the art precision in classes consisting of a large number of datapoints, but failed to achieve such figures in those classes which are deficient of enough data points [6,7,8]. Furthermore, most prior investigations of automated cytomorphologic identification concentrated on the classification of physiological cell types or peripheral blood smears as shown in Figure 1, restricting their applicability to the characterization of leukocytes in the BM for haematological malignancies diagnosis. Recent advancements in this field have focused mainly upon applying efficient Image segmentation and classification techniques; however, class imbalance as presented in Figure 2 played a major role in the classification task to achieve overall accuracy [9,10].
Bone marrow is a dynamic tissue compartment in the cavity of bones. In adults, haematopoietic cells are produced by the bone marrow cells in the large bones that account for 2–5% of an adult’s weight. Bone marrow imaging is a diagnostic tool for determination and staging of several haematological bone marrow disorders [3]. The sample acquisition process is mainly accomplished using either bone marrow aspiration or bone marrow biopsy techniques. These examinations are quite useful in finding out the root cause of problem with RBC’s (red blood cells), WBC’s (white blood cells) and platelets as well as monitoring related diseases such as polycythaemia vera, thrombocytopenia and anaemia including certain type of cancers such as lymphoma, leukaemia and multiple myeloma [1].
Because of the uneven staining and high-cell population, segmenting the nucleus and cytoplasm of leukocytes from bone marrow pictures is a complex procedure. The nucleus and cytoplasm of leukocytes were extracted using a combination of segmentation techniques such as thresholding, edge detection, pixel clustering and area growth. These approaches have been used since the photographs had homogeneous backgrounds and strong contrast, allowing the things of interest to be clearly distinguished. Samples generated for daily investigations in haematology laboratories, on the other hand, are frequently not consistent [6]. Few studies integrate training techniques such as SVM, mean-shift and coloured image segmentation to provide an automated AI method for finding models of various regions of an image. When compared to the thresholding and watershed algorithms, these approaches yield greater segmentation accuracy in complicated situations and are more resistant to colour confusion and alterations. Few strategies have been proposed to tackle the problem of overlaying blood cells. The underlying algorithm separates the cells either by employing dividing lines to combine concave points or by applying thresholding techniques and expanding the ROI while keeping the form [7].
Diagnosis of bone marrow is a time-consuming process as well as it requires proper infrastructure of diagnosis tools to spin up the process. Since it involves use of complex methodology and tools, the bearing cost is also quite high. Deep learning classification models currently available with high accuracy are not trained with all the classes and those which have been trained with all the classes fail to achieve acceptable accuracy to be utilised for medical diagnosis. This paper presents a way to fill this trade off gap by adopting a model that achieves high accuracy even with low density classes.
In our work, we have focused on improving the overall accuracy of classification. The dataset consists of more than 170,000 bone marrow images with very acute class imbalance. Thus, the main motivation was to find a way to eradicate the effect of this imbalance to train a deep neural network that has very high overall accuracy. The key tool that we adopted to counter the problem of class imbalance is the use of the Siamese network. The main key feature of this network is that it works well even when we are in shortage of enough data points, which is exactly the issue we are facing with some of the data classes. While testing the Siamese model, it outperformed XGB and SVM model with a training accuracy more than 91% and valid accuracy of over 84%. The model also performed well with individual class classification.
The remainder of the paper is laid out as follows: Section 2 contains a literature survey in the area of bone marrow morphology. Section 3 contains a proposed system and algorithms. Section 4 proposes the experimental results, discussion and performance analysis. The study is concluded in Section 5, with future directions.

2. Literature Review

Bone marrow cell classification is a key area in the prognosis of diseases related to blood. Many classification models have been proposed related to transfer learning CNN models, DC-GAN with ResNet combination and VGG based implementation for cell classification. In the following paragraphs we present an overview of those methods.
Study [8] proposed a model that combined DC-GAN with ResNet for classification of blood cells and adopted a transfer learning approach on ImageNet dataset. The results shows that there is an increase in accuracy by 1.2% on DC-GAN enhanced images with an overall test accuracy of 91.68%. This research [11] adopted a CNN based approach consisting of VGG16 and InceptionV3 to classify blood cell type against 17,902 digital images and eight classes. The overall accuracy was 90%; however, there was a large variation in true positives rates for individual classes in both the approaches. Another study [5] proposed a rapid localization of bone marrow and self-generating ROI with a patch-based deep learning model for the classification of 16 kinds of cell. The proposed model was trained on 12,426 annotated cells with an overall validation recall of more than 90.5%. The results of a test conducted on another batch of 3000 images, achieving accuracy of more than 98%, show that the model does not overfit the training data at all.
Currently, due to the exploding development in deep learning, increasing research has been conducted on the classification of multiple myeloma. The authors of [9] explored the effectiveness of detection algorithms such as RCNN and detection of multiple myeloma using unet, for which they used 85 high resolution microscopic images. The model possessed a state-of-the-art training loss of 0.05 after 400 epochs, but with a magnificent difference in validation loss. Despite the fact that Mask-RCNN is capable of addressing most of the issues associated with myeloma cell segmentation, it failed to detect some of the under-stained cells when the colour contrast between cytoplasm and background is lower.
At the same time, DL systems for the concurrent detection and classification in histology image analysis deal with the location and distinct types of nuclei/cells. Traditionally, these two challenges have been tackled separately, necessitating more training time. To solve this time constraint issue, this research [10] came up with a concatenated asymmetric DL structure (Syn-AHDA) for efficiently accounting of pictures with deformed features and noise effects. The experimental dataset includes 10,496 annotated images. The proposed model achieved detection precision of 92.66% and classification precision of 87.12%. The main feature of the Syn-ADHA network is that it produced competitive accuracy but took almost two-thirds of the overall training time as compared to another network.
The studies [12,13,14] created an AI system that evaluates blood cell dysplasia on bone marrow smears as well as displays the outcome of AI forecast for one of the dysplasias that is most representative: reduced granules. The classifier’s performance was assessed using a five-fold cross-validation after the detector, which was built on a deep learning model pre-trained with natural pictures and resulted in 97.2% and 98.2% accuracy, respectively, when labels for DG1–3 are deemed positive and DG1 is not used. In order to do a differential cell count on bone marrow smears to aid in diagnosis, the authors of [15] built an autonomous system and pathologists using microscopes conducted the study in parallel with machine automatic recognition and traditional manual differential count. The result of classification was based on specificity, sensitivity and accuracy and the effectiveness of the approach in classifying the cells in BM smears was validated with an overall validation accuracy of 90.1%. It may be able to help with the clinical application of BM smear examination.
The purpose of this [16] study is to determine whether transcriptomic approach machine learning can predict the presence of AML without expert input. They took into account a variety of real-world situations, such as cross-study problems and prediction across several technology platforms, and discovered that reliable prediction is feasible in a variety of situations and, in many instances, with only a small number of training samples. However, they also demonstrated that, based on the instance and the corresponding prevalence, huge training sets might be necessary to achieve accuracies good enough to produce respectable PPVs. The findings indicate that it may be possible to attain decent performance in a nearly automated manner with current technologies. This [17] study focuses on the part that senescence plays in the bone marrow (BM) microenvironment’s ageing process. Numerous BM conditions are age-related illnesses that heavily rely on the bone marrow microenvironment. The study resulted in 97.6% accuracy in classifying AML bone marrow type v/s other types in dataset 1 comprising 2500 samples, 98% accuracy in 8348 samples from Affymetrix HG U133 2.0 Micro array and 99.1% accuracy from 1181 samples obtained using RNAseq.
This [18] study created a machine learning model for MDS prediction one year before the condition was clinically diagnosed. A total number of 790,470 patients participated in this trial, of whom 1428 received an MDS diagnosis in the end, whereas 789,042 did not. The XGB model was contrasted with two additional models: artificial neural networks and logistic regression, two alternative machine learning methods. This [19] study showed that cancer/MSC cell fusion, a relatively uncommon occurrence, can impart chemotherapy resistance in addition to other PR tumorigenic features to malignancies that have been known to attract MSC. They examined the association between MSC fusion and the pattern of gene expression in SCC-25 cancer cells. Other noteworthy findings were that only 21% of those in the negative class test set had previously been diagnosed with cancer, compared to 33% of those in the positive class set.
Although classifying the blood cells manually is frequently employed in clinics and hospitals, it is not time-effective and has an impact on clinical workflow. Automated cell classification technologies could help clinicians make diagnoses more quickly and accurately. Convolutional Neural Network and Support Vector Machine (CNN-SVM) was presented in this [20] study as a solution to this issue, with Convolutional Neural Network model being utilised to directly extract features from the pictures. Neutrophil, lymphocyte, monocyte, eosinophil and pathogenic white blood cells are the five categories of white blood cells that may be classified using the features that were collected. On 15,764 enhanced pictures, three distinct CNN models, including Alexnet, Resnet-101 and VGG-19, were tested to determine which one would be the most effective at extracting features. Resnet-101 was chosen to categorise five categories of white blood cells because it had the highest accuracy rate of (97.8%) when combined with SVM. Examining the bone marrow (BM) is a crucial first step in the diagnosis and treatment of many hematologic diseases. For the purpose of BM NDC WSI analysis, this study [9] presented an effective and entirely autonomous hierarchical deep learning system. The suggested hierarchical framework is made up of three parts: (1) a deep learning model for quick localization of BM particles and cellular trails that creates regions of interest (ROI) for additional analysis; (2) a patch-based deep learning approach for cell recognition of 16 cell types, that have not been shown in prior studies; and (3) a fast-stitching model for combining patch-based results. The suggested method outperforms the current benchmark small-image-based techniques in terms of recall, accuracy and computing efficiency. Table 1 lists the summary of finding from other papers.

3. Material and Methods

In this paper, we proposed a probabilistic Siamese network with triplet loss function for bone marrow cell classification. For justified comparison, 2 parallel models are trained against identical training, test and validation which include:
CNN + SVM
CNN + Xgboost
The Bone Marrow Cell Classification dataset available in Kaggle [21] was used for this work. The dataset comprises more than 170,000 annotated images taken from the bone marrow smears of 945 patients using the May-Grünwald-Giemsa/Pappenheim stain [22]. The images had been acquired using a brightfield microscope with 40x magnification and oil immersion. The samples had been processed in the Munich Leukaemia Laboratory (MLL), scanned using equipment developed at Fraunhofer IIS and post-processed using software developed at Helmholtz Munic [23].
The block diagram for the model is present in Figure 3. The proposed model is composed of:
Four convolution layers of paired input images accompanied by Max Pooling layer. A concatenation layer to feed forward the output of both the convolution + dense transforms. Five Dense layers to predict whether the provided paired images were of same class or not.
This methodology is quite effective in mitigating the vast accuracy variation due to imbalance in the dataset classes. In short, Siamese network learns differentiating features among other networks in a must faster rate as compared to other persisting models.

3.1. Siamese Network

Siamese network, often called twin neural network proposed by Chicco, Davide, which uses the identical training parameters for both the given inputs for accomplishing the classification task. Out of the given pair, one of the inputs has already been assigned with a particular label, thus limiting the job of the network to extract identical or differentiating features out of the pair.
For learning the parameters, the Siamese model applies an optimization function on the triplet loss function. Triplet loss function requires one anchor data point as well as one classification data point with either a positive or negative label. If the data point is positive, it means that both the fed inputs belong to the same class, and vice versa. The model includes two series of networks, one for generating the encodings out of the image, and the other solemnly dedicated to the classification task from those acquired encodings. The primary objective of triplet loss function is the difference between anchor and positive data point to be smaller than anchor and negative data point, i.e.,
f a n c h o r f p o s i t i v e < f a n c h o r f n e g a t i v e + ε
where,
f(n) is encodings of input data points
ε often known as margin to nullify the trivial solution.
Thus, the overall triplet loss function is defined as:
L a n c h o r ,   p o s i t i v e ,   n e g a t i v e = m a x f a n c h o r f p o s i t i v e f a n c h o r f n e g a t i v e + ε ,   0
The max function is applied intentionally, so that the loss function does not backfires itself when we obtain the desired negative loss, that is, when the overall loss
Loss = Σ   L anchor i ,   positive i ,   negative i
is positive the model will try to improve the parameters using the optimising algorithm provided to the model, and when the first argument to the max function turns out to be negative, it will take zero instead, so that loss function for that particular pair does not have any effect on the optimising algorithm.
The dataset consists of paired 50 × 50 images and corresponding similarity labels. We try to maximise this similarity label prediction through hidden representation generated by the encoding layer of the input. The conditional likelihood is given by:
P h 1 ,   h 2 ,   ε = ( g h 1 ,   h 2 ,   ϵ s   1 g h 1 ,   h 2 ,   ε 1 s )
where,
h1 = encoding generated by anchor image,
h2 = encoding generated by test image
g(h1, h2, e) = contrastive loss function
s = similarity label
ε = margin
After taking natural log both side in (1):
L s   |   h 1 ,   h 2 ,   ε = ( s l o g   g h 1 ,   h 2 ,   ε + 1 s l o g   1 g h 1 ,   h 2 ,   ε    
where, L = Contrastive loss function
The hidden encodings are generated using forward passing the image to the convolution layer followed by a flatten and dense layer with ReLU activation. In Siamese network, the model parameters, as well as the network parameters, are simultaneously optimised by taking the log likelihood w.r.t. similarity model given by:
L α = S m g m g m 1 g m   g h 1 m , h 2 m ,   α   α
L h 1 = S m g m g m 1 g m   g h 1 m , h 2 m ,   α   h 1
L h 2 = S m g m g m 1 g m   g h 1 m , h 2 m ,   α   h 2

3.2. Algorithm

First from the dataset of 170,000 images paired data are generated and then divided into mini batches. Stranding of similar and dissimilar pairs are kept equal for each of the class labels. The pairs are then passed through a model where the network parameters and conditional probability parameters are jointly updated.
Create training pair images
For n number of epochs
   For each image pair do
    Forward Propagation
    Compute f(anchor) and f(test)
    Compute h1k and h2k and gk
    Backward propagation computing derivatives for both network as well as conditional parameters
   End for
   Update parameters using adam optimizer function
End for

3.2.1. Forward Propagation

Convolution Layer

The input image is passed through a series of convolution units as shown in Figure 4 to extract key features from the image to generate optimal encoding for it. Pixel values in convoluted images are generated as the average weighted sum of kernel and neighbouring pixel values. The selection of the desired kernel has also been incorporated as a part of model training itself.

Max Pooling Layer

After series of convolution the model is passed through max pooling layer where input pixels are taken in a 2 × 2 grid manner and maximum out of it is retained in the output discarding the other 3-pixel values. After each max pooling layer, the dimension of output image becomes the half of input image as demonstrated in Figure 5.

Flatten Layer

This layer changes the input 2D matrix to a 1D vector which can be then fed to the Dense layer as shown in Figure 6. The stacking of pixel layers in 1D channel is performed in a row-wise manner.

Dense Layer

It is a deeply connected layer of neurons that takes upon the input and generates the result from the following mathematical formula:
O u t p u t = A c t i v a t i o n d o t W e i g h t ,   i n p u t + b a i s
wherein Weight and bais are model trainable parameters and ReLu activation has been used to introduce non-linearity in the model.

3.2.2. Backward Propagation

The Siamese network learns by back propagating the derivatives according to the chain rule as follows:
For the last Dense layer, derivatives are calculated as:
L W 5 = L h 1   h 1 W 5 + L h 2   h 2 W 5
where ∂h1/∂W and ∂h2/∂W is given by:
h 1 W 5 = h 1 Z 1 Z 1 W 5
h 2 W 5 = h 2 Z 2 Z 2 W 5
From this the overall set of back propagating derivatives can be derived as follows:
δ 1 5 = L h 1 f Z 1 5
δ 2 5 = L h 2 f Z 2 5
while for layers l = 1 to 4 is given as:
δ 1 L = W L   δ 1 L + 1   f z 1 L
δ 2 L = W L   δ 2 L + 1   f z 2 L

3.2.3. Update

The parameters are then updated as follows in the adam optimizer:
W L = W L η A 1 L   V δ 1 C o r r e c t e d S δ 1 C o r r e c t e d + ϵ + A 2 L   V δ 2 C o r r e c t e d S δ 2 C o r r e c t e d + ε
b L = b L η δ 1 L + 1 + δ 2 L + 1
where Sδi is given by,
V δ i c o r r e c t e d = V δ i 1 β 1
S δ i c o r r e c t e d =   S δ i 1 β 2
while Sδi and Vδi equals,
V δ i e p o c h = V δ i e p o c h 1   β 1 + 1 β 1   δ i
S δ i e p o c h = S δ i e p o c h 1   β 2 + 1 β 2   δ i   δ i
where η, β1, β2 are adjustable hyperparameters of the model and initially
Vδ1[0] = 0
Vδ2[0] = 0
Sδ1[0] = 0
Sδ2[0] = 0.

4. Experimental Results

The dataset used for this work comprises over 170,000 images of expert-annotated cells from the bone marrow smears of 945 patients stained using the May–Grünwald–Giemsa/Pappenheim stain. The images of the dataset were acquired using brightfield microscope with 40x magnification and oil immersion. The dataset comprises 21 unique class labels. The details of the class labels are provided in Table 2.
The development of data-driven, computational approaches in diagnostic medicine requires large datasets with high quality data capture and annotation. Only a few datasets are now publicly accessible for bone marrow cells morphology, a crucial diagnostic technique for a wide range of hematologic disorders. All samples were processed at the Munich Leukaemia Laboratory (MLL), which also used Fraunhofer IIS equipment for scanning and Helmholtz Munich software for post-processing.
The Metrics taken for evaluation of the models are-
Accuracy = T P + T N T P + T N + F N + F P
Precision = T P T P + F P
Rec   ( or   Sensitivity ) = T P T P + F N
F 1 = 2 1 P r e c i s i o n + 1 R e c a l l
where,
TP = True Positives
TN = True Negatives
FP = False Positives
FN = False Negatives.

4.1. Siamese Neural Network

4.1.1. Loss, Validation Loss, Accuracy and Validation Accuracy against Number of Epochs for Cell Classification

The Siamese neural network model was trained on a bone marrow cell class classification image dataset, while training the loss on training dataset validation loss on validation dataset was noted, as shown in Figure 7, in every epoch, and thus the graph is plotted according to the values.
The Siamese neural network model was trained on a bone marrow cell class classification image dataset, while training the accuracy on training dataset validation accuracy on validation dataset was noted as shown in Figure 8 in every epoch, and thus the graph is plotted according to the values.

4.1.2. Similarity between Class Images

Here to check the similarity between the images belonging to the same or different class, the parallel inputs A and B are created and the output, which is the distance based on loss function, is depicted as similarity percentage which depicts the confidence of images belonging to the same or different class of cells, as presented in Figure 9.
In the below-given Figure 10, the first three images in both the rows belongs to the same class, hence the similarity between them is 100%. The last three images in both rows belong to different classes of cells, and hence the similarity between A and B is 0%.

4.1.3. Actual and Predicted Class Images

After the final training of model with certain epochs the actual and predicted images of cells are given below represented as Figure 11.
Actual 100% signifies the images A and B belongs to same class and the prediction is performed by the model which is 88%, 21% and 98%, respectively, for the same class.
Whereas Actual 0% signifies the images, A and B belong to different classes, and the prediction is performed by the model, which is 1%, 0% and 2%, respectively, for different classes.

4.1.4. Confusion Matrix

Below, in the given confusion matrix, is the visualisation of the model performance which was trained on bone marrow cell images to classify the cells based on the class of the training dataset. The visual representation is shown in Figure 12.
A total of 109,670 images were taken in the train set, out of which 41 belonged to class LYI, 1753 of PEB, 30 FGC class images, 4196 of MYB, 188 of OTH, 7662 of BLA, 1955 of MMZ, 2585 of MON, 12,563 of ART, 282 of BAS, 7676 of PMO, 3765 of EOS, 26 of KSC, 5 of ABE, 4882 of PLM, 6379 of NGB, 261 of HAC, 17,532 of EBO, 2264 of NIF, 16,794 of LYT, and 18,831 of NGS.
According to the confusion matrix shown, the model correctly classifies 41 images out of 41, 1752 images out of 1753, 30 images out of 30, 4182 images out of 4196, 188 images out of 188, 7576 images out of 7662, 1943 images out of 1955, 2579 images out of 2585, 4931 images out of 12,563, 282 images out of 282, 7480 images out of 7676, 3765 images out of 3765, 26 images out of 26, 5 images out of 5, 4865 images out of 4882, 6315 images out of 6379, 261 images out of 261, 17,070 images out of 17,532, 2260 images out of 2264, 15,971 images out of 16,794, and 18,183 images out of 18,831 for each class, respectively.
Similarly, below the confusion matrix is the visualisation of the model performance which was trained on bone marrow cells images to classify the cells based on class on the validation dataset. The visual representation is shown in Figure 13.
A total of 30,837 images were taken in the validation set, out of which 11 belonged to class LYI, 493 of PEB, 8 FGC class images, 1180 of MYB, 52 of OTH, 2155 of BLA, 549 of MMZ, 727 of MON, 3533 of ART, 79 of BAS, 2158 of PMO, 1058 of EOS, 7 of KSC, 1 of ABE, 1373 of PLM, 1794 of NGB, 73 of HAC, 4931 of EBO, 636 of NIF, 4723 of LYT, and 5296 of NGS.
According to the confusion matrix shown, the model correctly classifies 10 images out of 11, 414 images out of 493, 2 images out of 8, 1050 images out of 1180, 11 images out of 52, 1928 images out of 2155, 445 images out of 549, 356 images out of 727, 3264 images out of 3533, 28 images out of 79, 2008 images out of 2158, 986 images out of 1058, 7 images out of 7, 1 image out of 1, 1282 images out of 1373, 1530 images out of 1794, 56 images out of 73, 4062 images out of 4931, 524 images out of 636, 2912 images out of 4723, and 5136 images out of 5296 for each class, respectively.

4.1.5. Evaluation Metrics

The evaluation metrics here, as shown in Table 3 and Table 4, are constructed on the precision, recall, f1 score and support values resulted from the implemented strategy for different cells’ class classification on the training and validation set. It can be observed that the weighted average recall score for training and validation of the Siamese network was 92% and 91%, respectively. As we know, recall score is a good indicator of the classification of false negatives and gives details about missed diagnostics. In that sense, a weighted average recall score of 92% for training and 91% for validation of Siamese network indicates that the model performs well in detecting anomalous cases.

4.1.6. Comparison among Various Models Used

Here, the comparison table shown in Table 5 between different models mentioned in this research is presented taking in the metric as an F1 score and the plotting of the confusion matrix represented in Figure 14, Figure 15 and Figure 16. Since F1= 2 1 P r e c i s i o n + 1 R e c a l l , It will take into account both precision and recall. The comparison given is for validation dataset containing 30,837 images accounting for 21 classes of bone marrow cells.

4.1.7. Confusion Matrix

As we can analyse from both the F1-score, accuracy, recall and as well as confusion matrix, the best performing model is Siamese neural network based on the validation set.

5. Discussion and Future Scope

In this paper, we present Siamese network to classify BM into 21 different classes. Substantial evaluation on 30,837 valid images has shown dominance and generalisation in predicting the labelled class with high accuracy. The model used unified Siamese architecture and CNN to classify the images. It is distinct from the existing classification models ResNeXt-50 and XGG-Boost in the sense that it does not rely on feature extraction from single image itself, rather it focuses more on similarity and dissimilarity among images between the same class and different class labels, respectively. The final accuracy on the Siamese neural network model resulted in 91% training and 84% validation accuracy. In future research, the major focus of interest will be investigation upon YOLO and Grad-CAM algorithms to focus specifically on region of interest in a single cell image, in order to reduce major dependency on convolutional unit to generate encoding for the image. In addition, we will explore various image augmentation algorithms to diminish the effect on accuracy due to class imbalance.

Author Contributions

Conceptualization, S.A., B.A. and M.S.K.; Methodology, S.A., B.A. and A.S; Software, S.A, P.G. and V.G.; Validation, A.B, M.S.K., A.S. and S.A.; Formal Analysis, S.A.; Investigation, S.A., P.G. and V.G.; Resources, S.A.; Data Curation, S.A.; Writing—original draft preparation, S.A., P.G. and V.G.; Writing—review and editing, B.A.; Visualization, S.A., P.G. and V.G.; Supervision, B.A.; Project Administration, B.A. and S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings are available on request. Moreover, the Bone Marrow Cell Classification dataset can be downloaded from Kaggle.

Acknowledgments

The authors wish to express their thanks to Vellore Institute of Technology (VIT) management for their extensive support during this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guo, L.; Huang, P.; Huang, D.; Li, Z.; She, C.; Guo, Q.; Zhang, Q.; Li, J.; Ma, Q.; Li, J. A classification method to classify bone marrow cells with class imbalance problem. Biomed. Signal Processing Control. 2022, 72, 103296. [Google Scholar] [CrossRef]
  2. Chandradevan, R.; Aljudi, A.A.; Drumheller, B.R.; Kunananthaseelan, N.; Amgad, M.; Gutman, D.A.; Cooper, L.A.D.; Jaye, D.L. Machine-based detection and classification for bone marrow aspirate differential counts: Initial development focusing on nonneoplastic cells. Lab. Investig. 2020, 100, 98–109. [Google Scholar] [CrossRef] [PubMed]
  3. Boes, K.; Durham, A. Bone Marrow, Blood Cells, and the Lymphoid/Lymphatic System1. Pathol. Basis Vet. Dis. 2017, 724–804.e2. [Google Scholar]
  4. Theera-Umpon, N. White blood cell segmentation and classification in microscopic bone marrow images. In Proceedings of the International Conference on Fuzzy Systems and Knowledge Discovery, Changsha, China, 27–29 August 2005; Springer: Berlin/Heidelberg, Germnay, 2005; pp. 787–796. [Google Scholar]
  5. Wang, C.W.; Huang, S.C.; Lee, Y.C.; Shen, Y.J.; Meng, S.I.; Gaol, J.L. Deep learning for bone marrow cell detection and classification on whole-slide images. Med. Image Anal. 2022, 75, 102270. [Google Scholar] [CrossRef]
  6. Dorini, L.B.; Minetto, R.; Leite, N.J. White blood cell segmentation using morphological operators and scale-space analysis. In Proceedings of the XX Brazilian Symposium on Computer Graphics and Image Processing (SIBGRAPI 2007), Belo Horizonte, Brazil, 7–10 October 2007; pp. 294–304. [Google Scholar]
  7. Wang, W.; Song, H. Cell cluster image segmentation on form analysis. In Proceedings of the Third International Conference on Natural Computation (ICNC 2007), Haikou, China, 24–27 August 2007; Volume 4, pp. 833–836. [Google Scholar]
  8. Ma, L.; Shuai, R.; Ran, X.; Liu, W.; Ye, C. Combining DC-GAN with ResNet for blood cell image classification. Med. Biol. Eng. Comput. 2020, 58, 1251–1264. [Google Scholar] [CrossRef]
  9. Vyshnav, M.T.; Sowmya, V.; Gopalakrishnan, E.A.; Menon, V.K. Deep learning based approach for multiple myeloma detection. In Proceedings of the 2020 11th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kharagpur, India, 1–3 July 2020; pp. 1–7. [Google Scholar]
  10. Waris, A.; Niazi, I.K.; Jamil, M.; Englehart, K.; Jensen, W.; Kamavuako, E.N. Multiday evaluation of techniques for EMG-based classification of hand motions. IEEE J. Biomed. Health Inform. 2018, 23, 1526–1534. [Google Scholar] [CrossRef] [Green Version]
  11. Acevedo, A.; Merino, A.; Boldú, L.; Molina, Á.; Alférez, S.; Rodellar, J. A new convolutional neural network predictive model for the automatic recognition of hypogranulated neutrophils in myelodysplastic syndromes. Comput. Biol. Med. 2021, 134, 104479. [Google Scholar] [CrossRef] [PubMed]
  12. Deng, Y.; Julaiti, A.; Ran, W.; He, Y. Bone marrow mesenchymal stem cells-derived exosomal microRNA-19b-3p targets SOCS1 to facilitate progression of esophageal cancer. Life Sci. 2021, 278, 119491. [Google Scholar] [CrossRef] [PubMed]
  13. Sarikhani, M.; Firouzamandi, M. Cellular senescence in cancers: Relationship between bone marrow cancer and cellular senescence. Mol. Biol. Rep. 2022, 49, 1–10. [Google Scholar] [CrossRef] [PubMed]
  14. Mori, J.; Kaji, S.; Kawai, H.; Kida, S.; Tsubokura, M.; Fukatsu, M.; Harada, K.; Noji, H.; Ikezoe, T.; Maeda, T.; et al. Assessment of dysplasia in bone marrow smear with convolutional neural network. Sci. Rep. 2020, 10, 1–8. [Google Scholar] [CrossRef] [PubMed]
  15. Jin, H.; Fu, X.; Cao, X.; Sun, M.; Wang, X.; Zhong, Y.; Yang, S.; Qi, C.; Peng, B.; He, X.; et al. Developing and preliminary validating an automatic cell classification system for bone marrow smears: A pilot study. J. Med. Syst. 2020, 44, 1–10. [Google Scholar] [CrossRef] [PubMed]
  16. Warnat-Herresthal, S.; Perrakis, K.; Taschler, B.; Becker, M.; Baßler, K.; Beyer, M.; Günther, P.; Schulte-Schrepping, J.; Seep, L.; Klee, K.; et al. Scalable prediction of acute myeloid leukemia using high-dimensional machine learning and blood transcriptomics. Iscience 2020, 23, 100780. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Hellmich, C.; Moore, J.A.; Bowles, K.M.; Rushworth, S.A. Bone marrow senescence and the microenvironment of hematological malignancies. Front. Oncol. 2020, 10, 230. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Radhachandran, A.; Garikipati, A.; Iqbal, Z.; Siefkas, A.; Barnes, G.; Hoffman, J.; Mao, Q.; Das, R. A machine learning approach to predicting risk of myelodysplastic syndrome. Leuk. Res. 2021, 109, 106639. [Google Scholar] [CrossRef] [PubMed]
  19. Liu, C.; Billet, S.; Choudhury, D.; Cheng, R.; Haldar, S.; Fernandez, A.; Biondi, S.; Liu, Z.; Zhou, H.; Bhowmick, N.A. Bone marrow mesenchymal stem cells interact with head and neck squamous cell carcinoma cells to promote cancer progression and drug resistance. Neoplasia 2021, 23, 118–128. [Google Scholar] [CrossRef] [PubMed]
  20. Truong, K.H.; Minh, D.N.; Trong, L.D. Automatic White Blood Cell Classification Using the Combination of Convolution Neural Network and Support Vector Machine. In Proceedings of the International Conference on Hybrid Intelligent Systems, Virtual Event, India, 14–16 December 2020; Springer: Cham, Switzerland, 2020; pp. 720–728. [Google Scholar]
  21. Matek, C.; Krappe, S.; Münzenmayer, C.; Haferlach, T.; Marr, C. Highly accurate differentiation of bone marrow cell morphologies using deep neural networks on a large image data set. Blood J. Am. Soc. Hematol. 2021, 138, 1917–1927. [Google Scholar] [CrossRef] [PubMed]
  22. Matek, C.; Krappe, S.; Münzenmayer, C.; Haferlach, T.; Marr, C. An Expert-Annotated Dataset of Bone Marrow Cytology in Hematologic Malignancies [Data set]. Cancer Imaging Arch. 2021. [Google Scholar] [CrossRef]
  23. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and Operating a Public Information Repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Bone marrow cells of 21 different classes.
Figure 1. Bone marrow cells of 21 different classes.
Diagnostics 13 00112 g001
Figure 2. Pie chart depicting the fraction of images for each class out of 170,000 images.
Figure 2. Pie chart depicting the fraction of images for each class out of 170,000 images.
Diagnostics 13 00112 g002
Figure 3. Block diagram of proposed Siamese neural network model.
Figure 3. Block diagram of proposed Siamese neural network model.
Diagnostics 13 00112 g003
Figure 4. Outline of operation principle of convolution layer with 3 × 3 kernel.
Figure 4. Outline of operation principle of convolution layer with 3 × 3 kernel.
Diagnostics 13 00112 g004
Figure 5. Illustration of function performed by kernel in max pooling layer for each 2 × 2 grid in a convolved image.
Figure 5. Illustration of function performed by kernel in max pooling layer for each 2 × 2 grid in a convolved image.
Diagnostics 13 00112 g005
Figure 6. Graphic demonstrating the role of the Flatten layer in converting a 2D grid to a 1D vector.
Figure 6. Graphic demonstrating the role of the Flatten layer in converting a 2D grid to a 1D vector.
Diagnostics 13 00112 g006
Figure 7. Loss v/s epoch graph for Siamese network model trained after training for 754 epochs.
Figure 7. Loss v/s epoch graph for Siamese network model trained after training for 754 epochs.
Diagnostics 13 00112 g007
Figure 8. Accuracy v/s epoch graph for Siamese network model trained after training for 754 epochs.
Figure 8. Accuracy v/s epoch graph for Siamese network model trained after training for 754 epochs.
Diagnostics 13 00112 g008
Figure 9. Block diagram illustrating end to end traversal in model right from the input of paired images.
Figure 9. Block diagram illustrating end to end traversal in model right from the input of paired images.
Diagnostics 13 00112 g009
Figure 10. Output of model with percentage confidence in similarity for given pairs of anchor and test input images.
Figure 10. Output of model with percentage confidence in similarity for given pairs of anchor and test input images.
Diagnostics 13 00112 g010
Figure 11. Confidence percentage of test images by Siamese model whether the input paired BM smear image belong to same class or not.
Figure 11. Confidence percentage of test images by Siamese model whether the input paired BM smear image belong to same class or not.
Diagnostics 13 00112 g011
Figure 12. Confusion matrix of predictions generated by Siamese model on training data.
Figure 12. Confusion matrix of predictions generated by Siamese model on training data.
Diagnostics 13 00112 g012
Figure 13. Confusion matrix of predictions generated by Siamese model on test data.
Figure 13. Confusion matrix of predictions generated by Siamese model on test data.
Diagnostics 13 00112 g013
Figure 14. Confusion matrix of predictions generated by CNN-XGBoost.
Figure 14. Confusion matrix of predictions generated by CNN-XGBoost.
Diagnostics 13 00112 g014
Figure 15. Confusion matrix of predictions made by CNN-SVM.
Figure 15. Confusion matrix of predictions made by CNN-SVM.
Diagnostics 13 00112 g015
Figure 16. Confusion matrix of predictions generated by Siamese model.
Figure 16. Confusion matrix of predictions generated by Siamese model.
Diagnostics 13 00112 g016
Table 1. A summary of the findings from other papers.
Table 1. A summary of the findings from other papers.
ReferenceAlgorithm/Model UsedPerformance MetricNumber of Bone Marrow Classes/LabelsRemark
7DC-GAN + ResnetAccuracy4 (eosinophils, lymphocytes, monocytes and neutrophils)91.70%
8CNN + VGG16Accuracy6 (neutrophils, eosinophils, basophils, monocytes, lymphocytes (T and B cells)) 94%
9ROI Processing+ DLRecall1684.20%
10RCNNRecall85 images92.68%
11Syn-ADHAPrecision1687.12%
18XGBoostAccuracy790,470 images88%
10UnetRecall85 images73.17%
20CNN + SVMAccuracy1697.80%
Table 2. Description of 21 different class labels in the dataset.
Table 2. Description of 21 different class labels in the dataset.
S. No.Class Label AcronymClass Label Description
1ABEAbnormal eosinophil
2ARTArtefact
3BASBasophil
4BLABlast
5EBOErythroblast
6EOSEosinophil
7FGCFaggott cell
8HACHairy cell
9KSCSmudge cell
10LYIImmature lymphocyte
11LYTLymphocyte
12MMZMetamyelocyte
13MONMonocyte
14MYBMyelocyte
15NGBBand neutrophil
16NGSSegmented neutrophil
17NIFNot identifiable
18PEBProerythroblast
19PLMPlasma cell
20PMOPromyelocyte
21OTHOther cell
Table 3. Evaluation metrics score for different cell classes on train data.
Table 3. Evaluation metrics score for different cell classes on train data.
Cell ClassPrecisionRecallF1-ScoreSupport
LYI1.000.670.81553
PEB1.000.770.872261
FGC1.000.560.71532
MYB1.000.900.944660
OTH1.000.690.81651
BLA0.990.940.968076
MMZ0.990.790.882455
MON1.000.830.913113
ART0.390.970.565065
BAS1.000.360.53778
PMO0.970.940.967954
EOS1.000.880.944265
KSC1.000.750.86529
ABE1.000.610.02492
PLM1.000.910.955349
NGB0.990.930.966817
HAC1.000.350.52740
EBO0.970.970.9717,526
NIF1.000.810.902778
LYT0.950.970.9616,413
NGC0.970.970.9718,663
Accuracy 0.91109,670
Weighted average 0.950.920.93109,670
Table 4. Evaluation metrics score for different cell classes on validation data.
Table 4. Evaluation metrics score for different cell classes on validation data.
Cell ClassPrecisionRecallF1-ScoreSupport
LYI0.910.740.81250
PEB0.840.630.72656
FGC0.530.670.59271
MYB0.890.790.841322
OTH0.210.650.31237
BLA0.890.890.892157
MMZ0.810.630.71711
MON0.490.610.54584
ART0.920.940.933477
BAS0.350.340.35274
PMO0.930.900.912234
EOS0.930.810.871219
KSC1.000.630.77239
ABE1.000.540.70218
PLM0.930.850.891515
NGB0.850.870.861760
HAC0.770.610.33262
EBO0.820.960.884249
NIF0.820.700.76751
LYT0.620.950.753078
NGC0.970.960.965373
Accuracy 0.8430,837
Weighted average 0.930.910.9130,837
Table 5. Comparison of different approaches used in this study based on F1-Score.
Table 5. Comparison of different approaches used in this study based on F1-Score.
ClassSiameseCNN-SVMCNN-XGBoost
LYI0.810.000.00
PEB0.720.000.08
FGC0.590.000.00
MYB0.840.000.00
OTH0.310.000.00
BLA0.890.000.30
MMZ0.710.000.00
MON0.540.000.00
ART0.930.090.26
BAS0.350.000.00
PMO0.910.070.31
EOS0.870.000.00
KSC0.770.000.00
ABE0.700.000.00
PLM0.890.000.00
NGB0.860.000.00
HAC0.330.000.00
EBO0.880.420.52
NIF0.760.000.02
LYT0.750.320.46
NGC0.960.390.50
Accuracy0.840.280.38
Weighted average0.910.200.31
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ananthakrishnan, B.; Shaik, A.; Akhouri, S.; Garg, P.; Gadag, V.; Kavitha, M.S. Automated Bone Marrow Cell Classification for Haematological Disease Diagnosis Using Siamese Neural Network. Diagnostics 2023, 13, 112. https://doi.org/10.3390/diagnostics13010112

AMA Style

Ananthakrishnan B, Shaik A, Akhouri S, Garg P, Gadag V, Kavitha MS. Automated Bone Marrow Cell Classification for Haematological Disease Diagnosis Using Siamese Neural Network. Diagnostics. 2023; 13(1):112. https://doi.org/10.3390/diagnostics13010112

Chicago/Turabian Style

Ananthakrishnan, Balasundaram, Ayesha Shaik, Shivam Akhouri, Paras Garg, Vaibhav Gadag, and Muthu Subash Kavitha. 2023. "Automated Bone Marrow Cell Classification for Haematological Disease Diagnosis Using Siamese Neural Network" Diagnostics 13, no. 1: 112. https://doi.org/10.3390/diagnostics13010112

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop