Next Article in Journal
Group Testing with Blocks of Positives and Inhibitors
Next Article in Special Issue
Improved Complementary Pulmonary Nodule Segmentation Model Based on Multi-Feature Fusion
Previous Article in Journal
The Analysis of the Power Law Feature in Complex Networks
Previous Article in Special Issue
Correction: Brochet et al. A Quantitative Comparison between Shannon and Tsallis–Havrda–Charvat Entropies Applied to Cancer Outcome Prediction. Entropy 2022, 24, 436
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Entropy Measurements for Leukocytes’ Surrounding Informativeness Evaluation for Acute Lymphoblastic Leukemia Classification

by
Krzysztof Pałczyński
*,†,
Damian Ledziński
and
Tomasz Andrysiak
Faculty of Telecommunications, Computer Science and Electrical Engineering, Bydgoszcz University of Science and Technology, 85-796 Bydgoszcz, Poland
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2022, 24(11), 1560; https://doi.org/10.3390/e24111560
Submission received: 14 September 2022 / Revised: 20 October 2022 / Accepted: 26 October 2022 / Published: 29 October 2022
(This article belongs to the Special Issue Application of Entropy to Computer Vision and Medical Imaging)

Abstract

:
The study of leukemia classification using deep learning techniques has been conducted by multiple research teams worldwide. Although deep convolutional neural networks achieved high quality of sick vs. healthy patient discrimination, their inherent lack of human interpretability of the decision-making process hinders the adoption of deep learning techniques in medicine. Research involving deep learning proved that distinguishing between healthy and sick patients using microscopic images of lymphocytes is possible. However, it could not provide information on the intermediate steps in the diagnosis process. As a result, despite numerous examinations, it is still unclear whether the lymphocyte is the only object in the microscopic picture containing leukemia-related information or if the leukocyte’s surroundings also contain the desired information. In this work, entropy measures and machine learning models were applied to study the informativeness of both whole images and lymphocytes’ surroundings alone for Leukemia classification. This work aims to provide human-interpretable features marking the probability of sickness occurrence. The research stated that the hue distribution of images with lymphocytes obfuscated alone is informative enough to facilitate 93.0% accuracy in healthy vs. sick classification. The research was conducted on the ALL-IDB2 dataset.

1. Introduction

Acute lymphoblastic lekemia (ALL) diagnosis is closely associated with morphological changes in white blood cells (WBC, or leukocytes). ALL, known in the group of blood diseases, is characterized by the overproduction and continuous proliferation of malignant and immature white blood cells (referred to as lymphoblasts or blasts). Although the number of leukocytes can often be considered an essential indicator of pathological changes in the morphological picture of the blood, it is not always sufficient.
The detection of ALL and its subtypes is often accomplished by examining blood or bone marrow smears. According to the French-American-British (FAB) classification standard, ALL is classified into the L1, L2, and L3 subtypes. Assignment to the correct subtype is carried out according to observation of the nucleus’s morphology, including the affected cell’s pattern and variation in its shape. This procedure is generally accepted and known from numerous works of authors researching the detection and classification of leukocytes.
In clinical practice, microscopic examination of blood smears to verify ALL is based primarily on counting different types of white blood cells. Equally important is the analysis of the nuclear features of leukocytes, which are often distinguished by their pastel blue coloration. Nevertheless, their further evaluation becomes complicated, as according to the FAB, features such as the size, shape variation, and texture should be considered. After all, leukocytes can be distinguished by their size, color characteristics, the ratio of the nuclei to the cytoplasm contained in them, etc.
Due to the morphological diversity of white blood cells, the classification into ALL subtypes may not always be realized correctly. The reason for this can perhaps be found in the complexity of the backgrounds of microscopic images. One of the essential aspects is played by the surroundings of white blood cells, which occur in the setting of other morphotic components of the blood such as red blood cells or platelets. However, according to FAB classification, the background does not add significant information from the perspective of classification, although it can significantly hinder the correct classification of ALL. Our work measures background informativeness to determine the influence of the lymphocytes’ surroundings on correct Leukemia classification.
In their article [1], Andrade et al. performed an extensive evaluation of the leukocyte segmentation techniques by artificial intelligence systems developed by well-established research teams. The authors performed experiments on five databases: ALL-IDB2 [2], BloodSeg [3], Leukocytes [4], JTSC Database, and CellaVision [5]. The leukocyte segmentation methods examined by the authors use Otsu threshold [6,7,8,9] (Madhloom et al., Arslan et al., Nazlibilek et al., and Prinyakupt et al.), K-means [10,11,12,13,14,15,16,17] (Nasir et al., Mohapatra et al., Madhukar et al., Amin et al., Sarrafzadeh et al., Vincent et al., Vogado et al., and Kumar et al.), region growing [7,10,18] (Nasir et al., Mohammed et al., and Arslan et al.), edge detector [18] (Mohammed et al.), Zack’s algorithm [19] (Abdeldaim et al.), and arithmetical image processing operations [3,6,9,14] (Madhloom et al., Mohamed et al., Prinyakupt et al., and Sarrafzadeh et al.). The methods were examined on the images encoded using RGB [9] (Prinyakupt et al.), grayscale [6,18] (Madhloom et al., and Mohammed et al.), HSI [10] (Nasir et al.), L*a*b* [11,12,14,15] (Mohapatra et al., Madhukar et al., Sarrafzadeh et al., and Vincent et al.), HSV [13,17] (Amin et al., and Kumar et al.), CMYK [19] (Abdeldaim et al.), and CMYK + L*a*b* color schemes [16] (Vogado et al.). The authors achieved satisfactory results for all of the datasets, in some cases reaching 97% accuracy. However, none of the methods examined proved to be the best on all datasets. It is also important to note that the method with the best results in this experiment achieved only a 58.44% leukocyte nuclei detection rate. In this article, a leukocyte was considered detected by computing the true positive rate metric T P R t with a threshold t = 0.9 .
In [1], Andrade et al. proved that leukocyte nuclei segmentation is a non-trivial task, and none of the well-established methods proved efficient in leukocyte detection. The authors stated in the article that leukocyte nuclei segmentation is performed to classify the presence of leukemia in the cells. The survey of image processing techniques and their results motivated us to perform research focusing on attempting to examine the amount of information contained in non-leukocytes for leukemia classification without performing image segmentation. Furthermore, the plethora of black box-type artificial intelligence systems applied with various evaluation results signified the importance of establishing the features’ quality.
The subject of leukemia classification was researched thoroughly using deep learning methods. The authors of [20,21,22,23,24] applied various types of convolutional neural networks (CNNs) [25]. In [20], Rehman et al. applied the AlexNet [26] architecture of CNN networks for leukemia classification, achieving 97.78% accuracy. Similar research was conducted by the authors of [21] (Prellberg et al.) using the ResNeXt50 [27] architecture of CNN networks pretrained on the ImageNet dataset [28]. The researchers achieved 89.7% accuracy. The non-binary classification of leukemia was conducted by Ahmed et al. [22]. The experiments were run to establish the ability of CNNs to discriminate in one vs. many mode against leukemia types such as acute myeloid leukemia (AML), chronic lymphocytic leukemia (CLL), chronic myeloid leukemia (CLM), and acute lymphoblastic leukemia (ALL). They achieved classification with 81% accuracy. The authors of [23] (Guo et al.) used Siamese networks [29] to achieve few-shot learning [30] with 89.96% accuracy. Similar research was conducted by Abhishek et al. [24]. In this work, the authors used transfer learning to compare the results of deep convolutional neural networks with support vector machines (SVMs) against SVM interpreting features extracted by local binary patterns (LBPs) [31] and the histogram of oriented gradients (HOG) [32]. The deep learning approach obtained 98% accuracy, SVM + LBP resulted in 83% accuracy, and SVM + HOG resulted in 50% accuracy. In their work, Rodrigues et al. [33] also applied deep learning for leukemia classification. What differs them from other works is the optimization of trained neural networks using a genetic algorithm. This procedure improved the results to 98.46%.
The results of leukemia classification using deep learning in [20,21,22,23] (Rehman et al., Prellberg et al., Ahmed et al., Guo et al., Abhishek et al., and Rodrigues et al.) provided substantial accuracy for healthy vs. sick discrimination. However, due to the black box nature of deep learning methods, it is unclear what pattern was extracted by neural networks to achieve such efficiency. The knowledge distillation [34] techniques are currently not advanced enough to determine the neural network’s reasoning, leading to classification in a manner humans can understand. Because of this issue, the adoption of deep learning and machine learning in medicine is slow due to the inability to verify the quality of the extracted features. Our work examines the ALL-IDB dataset samples to determine the features understandable by humans, allowing reliable classification and, at the same time, determining the usefulness of the leukocyte’s surroundings in leukemia classification.

1.1. Summary of Surveyed Research Works

The surveyed works are summarized in Table 1 to present various method and image encoding technique combinations applied in a readable format.

1.2. The Aim of This Work

This work is a continuation of the research described in [35] (Pałczyński et al.). This article aims to establish the informativeness of a lymphocyte’s surroundings for leukemia classification. The classification is conducted on the features extracted from images with lymphocytes obfuscated using black rectangles. The classification results without information regarding lymphocytes are compared against the quality of discrimination in the unmodified images. The features extracted from the image are deterministic, human-interpretable qualities. The discrimination is performed using both simple, divergence-based clusterization (mean squared error and cross-entropy [36]) and by applying machine learning algorithms such as logistic regression [37] and the XGBoost algorithm [38]. The images used in this research are encoded using RGB and HSV methods.
This work aims to quantify the amount of information stored in human-interpretable features computed from an image with the lymphocytes obfuscated. This work aims not to achieve the best classification results but to determine how well the classification can be performed using a limited amount of information while remaining interpretable by humans. In our previous article, deep neural networks were applied to the raw images to perform the classification. Although the results were satisfactory, the inherent black box nature of deep neural networks prevented us from acquiring human-interpretable knowledge on the nature of this particular classification problem. This work aims to provide such information.

1.3. Summary of Our Contributions

Our main contributions can be summarized as follows:
  • We examined the influence of lymphocyte obfuscation on acute lymphoblastic leukemia classification to evaluate its surroundings’ informativeness. The hue distribution of lymphocytes’ surroundings processed by the XGBoost algorithm resulted in classification with 93% accuracy.
  • We evaluated the informativeness of channels’ value distributions of both the RGB and HSV color encodings. We determined that the channel encoding color green contained the most information, with an XGBoost classification accuracy of 96%. The same evaluation of red and blue color channels resulted in classification accuracies of 87% and 83%, respectively. The hue, saturation, and value channels obtained classification results of 94%, 94%, and 84%, respectively.
  • The classification results of the XGBoost algorithm interpreting the distributions of individual channel values resulted in a classification quality similar to the effects of deep learning application on raw images performed by other researchers. As a result, we reduced the amount of input information by three orders of magnitude while achieving comparable results.
  • We evaluated the informativeness of the entropy measurements of each channel’s values distribution using the Shannon entropy. The Shannon entropy computed for the hue distribution of images with lymphocytes obfuscated resulted in a classification accuracy of 81% and 68% accuracy when using images without the lymphocytes being obfuscated. The results suggest that lymphocytes’ surroundings contain essential information for acute lymphoblastic leukemia classification.

1.4. Paper Organization

This work is divided into sections. Section 2 describes the materials and methods used in this research. This section describes the image preprocessing techniques, encoding, feature vectorization, experiments conducted, and metrics. Section 3 presents the results of the experiments described in Section 2. Section 4 provides interpretation of the results presented in Section 3, and Section 5 concludes the paper.

2. Materials and Methods

This section describes the materials and methods used in this research. It describes the data preprocessing, vectorization, and algorithms for generating the experimental results. The experimental procedure involved computing the Shannon entropy, measuring the cross-entropy score between the obtained value distributions, and fitting the machine learning models.

2.1. ALL-IDB Database

The research was conducted on the ALL-IDB database, containing microscopic images of lymphocyte cells documented from healthy people and patients with acute lymphoblastic leukemia. This database is made publicly available by Universitià degli Studi di Milano, and it contains annotation of which samples represent cases of leukemia and which were obtained from healthy patients. Oncologists performed the annotation.
The database contains 260 microscopic images. The dataset is balanced between classes, having 130 images of blood smears taken from healthy patients and the same amount from sick ones. The home website of this dataset is accessible at this link: http://homes.di.unimi.it/scotti/all/ (accessed on 15 October 2021).

2.2. Image Preprocessing

In this section, the techniques of image processing applied during the experiments are presented. The amount of information contained within the background of the image can be examined by removing the lymphocytes from the graphics. Removing lymphocyte information was performed by covering them with black rectangle-shaped bounding boxes. This type of obfuscation was chosen to remove all information from the lymphocytes and information regarding the cells’ shapes, which may have interfered with the experiment’s results. The resulting shrinkage of the background from such an obfuscation technique was not considered a concern. The results of this operation are presented on the Figure 1. The experiments were conducted using images both unmodified and obfuscated to compare the information stored in background of the image with all of the information contained.
The images were also subjected to data augmentation to determine the influence of commonly used image preprocessing techniques on the informativeness of the background. The data augmentation was performed before lymphocyte obfuscation. The modification methods examined in this research were the following:
  • Gaussian blur;
  • Median blur;
  • Gaussian noise.
Gaussian blur (also known as Gaussian smoothing) is a commonly used data augmentation technique in deep learning for increasing the number of images in the training set. It serves as a low-pass filter, reducing higher frequencies from the image and thus achieving the perceived effect of smoothness. The filter works by convolving the image with a matrix representation of a two-dimensional Gaussian function. Equation (1) presents the method for obtaining the filter matrix:
G ( x , y ) = 1 2 π σ 2 e x 2 + y 2 2 σ 2
where x and y represent the divergence from the center of the matrix. The bigger the matrix (kernel), the more precise the convolution is. The kernel lengths used in this research were 3, 9, 21, and 51. The results of applying this operation are presented in Figure 2.
Median blur has a similar purpose to Gaussian blur. This filter windows the image and returns the median value from each window. This operation reduces noise and creates a low-pass filter. The sizes used for the windows in this research were 3, 9, 21, and 51. The results of applying this operation are presented in Figure 3.
The last data augmentation technique used was adding Gaussian noise. Compared with the previous two techniques, this method increases the amount of noise instead of decreasing it. It is also commonly used in deep learning to reduce deep neural networks’ overreliance on temporal patterns in favor of a more holistic approach. Each pixel in the output file was computed using Equation (2):
x = x · ( 1 + n ) , n N ( 0 , σ )
Here, x is the current value of the filtered image, and σ is the standard deviation of the distribution. The values of variance used in this research were 0.001, 0.01, and 0.1. The results of applying this technique are presented in Figure 4. The processing data pipeline involving data augmentation and lymphocyte obfuscation is presented in Figure 5.

2.3. Image Vectorization

In this section, the preprocessed images are converted into vectors of features ready to be interpreted by the statistical and machine learning models. In this research, background informativeness was measured in regard to one image channel at a time. As a result, each experiment conducted started with selection of the aforementioned channel. The available channels were red (R), green (G), blue (B), hue (H), saturation (S), and value (V).
The first three channels (red, green, and blue) are natural components of RGB-encoded images. However, the hue, saturation, and value metrics are the result of HSV image encoding. HSV is a common type of color expression in an image more akin to the recognition process performed by the human eye.
The hue channel contains information on what color is present in the image. It represents a 360° coordinate of rotation around the circle of colors. Typical representation of the hue coordinates associate the color red with a value of 0°, yellow with 60°, green with 120°, aqua with 180°, blue with 240°, and purple with 300°. It is important to note that the hue represents a rotation angle, so the difference between two hue values is represented by the measurement of the shortest arc connecting two points on the hue circle. For example, the colors red (0°) and purple (300°) are 60° degrees apart instead of 300°. In this research, the OpenCV library was used for image processing, which encoded the hue channels with values from 0 to 180. This behavior was kept for both conducting experiments and presenting the results.
The saturation contains information regarding the color’s intensity. A saturation value of 0 results in the color gray, and the maximum value of 255 provides the most intense version of the color. On the other hand, the value channel contains information on the brightness of the color. The numeric value of 0 results in the color black, and the maximum value of 255 generates the brightest version of the color.
Equations (3)–(6) describe the process of image conversion from RGB encoding to HSV encoding:
M = m a x { R , G , B } , m = m i n { R , G , B } ,
V = M / 255
S = 1 m M M > 0 0 M = 0
H = c o s 1 ( R 1 2 G 1 2 B R 2 + B 2 + G 2 R B R G G B ) G B 360 ° c o s 1 R 1 2 G 1 2 B R 2 + B 2 + G 2 R B R G G B B > G
In the next step, the selected channel is vectorized by grouping its individual values into 30 equally spaced numeric bins and computing their distribution in the whole image. Such encoding provides information on what value occurs most frequently in the image. Pixels encoding black rectangles for obfuscation purposes were not included in the computations of color density distribution. An example of such calculations is presented in Figure 6.

2.4. Distribution Difference Measurement

In this section, images vectorized into a single channel’s values density distributions are compared to measure the amount of information stored in every channel’s histogram. Distributions from both the obfuscated and unmodified images are examined to establish the amount of information stored in the color distributions of both the entire images and the backgrounds only. The quality of image classification measures the information stored in color distributions regarding the representation of healthy and sick individuals.
In the first step of the experiment, the set of vectorized images and the corresponding set of labels are randomly split into training and test sets. Then, the distribution vectors from the training set were split into subsets representing each class. Next, the averaged distributions were computed from the subsets, and each computed average distribution represented the class from which the samples were computed. In another step, the divergence metric was chosen to measure how different the samples were from each representative distribution. The divergence metrics chosen in this research were the cross-entropy and mean squared error, and they are explained in detail in Section 2.4.1 and Section 2.4.2. The divergence metric was used to compute the divergence of each sample from the training set to each representative distribution, and their values were stored in the respective sets. Next, from each class divergence set, the mean and standard deviation were computed. Then, the statistical model training was finished and ready for performing test evaluation.
During the evaluation, the samples from the test set were subjected to divergence computation for the averaged distributions representing each class. The similarities to each distribution were normalized by subtracting the corresponding mean and dividing by the standard deviation. The normalized similarities of each sample to each distribution were compared. The sample was associated with the class whose representative distribution had the smallest normalized divergence. The evaluation of classification quality measures the amount of information contained in the value distribution. The process is graphically presented in Figure 7 and described in detail in the pseudo-code in Section 2.4.3.

2.4.1. Cross-Entropy

Cross-entropy is a technique from information theory for computing the divergence between two probability distributions. Equation (7) describes the process of metric calculation. The value density distribution can be interpreted as the probability that an individual pixel has a certain value. Such an interpretation allows the usage of cross-entropy in value density distribution classification:
H ( P , Q ) = i = 0 | P | P i log Q i , | P | = | Q |
where P and Q are the density distributions subjected to the comparison.

2.4.2. Mean Sqaured Error

The mean squared error (MSE) is a commonly used technique for comparing two vectors. It is described by Equation (8). Due to powering the differences, the MSE is prone to disregarding multiple closely matched dimensions of two vectors in favor of penalizing a few outlying ones. This feature is useful in vector comparison because it forces the algorithms to even out their match functions instead of attuning easily matchable parts of the vector and disregarding the difficult ones:
M S E ( P , Q ) = i = 0 | P | ( P i Q i ) 2 , | P | = | Q |
where P and Q are the vectors subjected to the comparison.

2.4.3. Algorithm

The graphical representation of the experiment process is depicted in Figure 7. The pseudo-code describing this process in detail is presented below Algorithm 1:
Algorithm 1: The mathematical formulation of experimental procedure examining distribution difference measurements
  • Input the set of samples X;
  • Input the set of sample labels Y;
  • Input divergence function f n ;
  • Randomly shuffle the set of indices I = { i | i N + i ( 0 ; | X | + 1 ) } ;
  • s = 2 3 | X | ;
  • I t r a i n = I [ : s ] ;
  • I t e s t = I [ s : ] ;
  • X t r a i n = { X i | i I t r a i n } ;
  • Y t r a i n = { Y i | i I t r a i n } ;
  • X t e s t = { X i | i I t e s t } ;
  • Y t e s t = { Y i | i I t e s t } ;
  • X 0 = [ { X t r a i n [ i ] | i ( 0 , | X t r a i n | + 1 ) Y t r a i n [ i ] = 0 } ] ;
  • X 1 = [ { X t r a i n [ i ] | i ( 0 , | X t r a i n | + 1 ) Y t r a i n [ i ] = 1 } ] ;
  • M 0 = [ { 1 | X 0 | j X 0 X 0 [ j , i ] | i N + i | X 0 [ 0 ] | } ] ;
  • M 1 = [ { 1 | X 1 | j X 1 X 1 [ j , i ] | i N + i | X 1 [ 0 ] | } ] ;
  • D 0 = { f n ( X t r a i n [ i ] , M 0 ) | i N + i ( 0 ; | X t r a i n | + 1 } ;
  • D 1 = { f n ( X t r a i n [ i ] , M 1 ) | i N + i ( 0 ; | X t r a i n | + 1 } ;
  • m 0 , s 0 , m 1 , s 1 = m e a n ( D 0 ) , s t d ( D 0 ) , m e a n ( D 1 ) , s t d ( D 1 ) ;
  • D t 0 = { 1 s 0 ( f n ( X t e s t [ i ] , M 0 ) m 0 ) | i N + i ( 0 , | X t e s t | + 1 ) } ;
  • D t 1 = { 1 s 1 ( f n ( X t e s t [ i ] , M 1 ) m 1 ) | i N + i ( 0 , | X t e s t | + 1 ) } ;
  • P = [ { D t 1 [ i ] < D t 0 [ i ] | i N + i ( 0 ; | D t 0 | + 1 ) } ] ;
  • Compare the prediction vector P with the label vector Y t e s t .

2.5. Shannon Entropy

The Shannon entropy is a mathematical tool from the field of information theory that allows measuring the amount of uncertainty the probability distribution contains. The more evenly spaced the probability among the distribution states, the higher the value of the Shannon entropy. The computation of the Shannon entropy is performed using Equation (9):
H ( P ) = i = 0 | P | P i log P i
where P is the vectorized density distribution subjected to the Shannon entropy computation.
In this research, the Shannon entropy was used to quantify the uncertainty associated with each channel’s value distribution and evaluate whether there was a significant difference between the Shannon entropy of the samples from the healthy class and the entropy of the sick class of samples. The significance of the difference in entropy measurements was established by using the Shannon entropy as a single-value determinant in the classification of whether a patient was healthy or sick. The classification was performed by fitting the logistic regression model on randomly split training data and evaluating it on the test data.

2.6. Machine Learning Algorithms

The last experiment aimed to apply machine learning algorithms directly to the channel’s value distribution to attempt to perform the classification. The algorithms used for this task were XGBoost and logistic regression. The former is one of the most robust, state-of-the-art machine learning algorithms capable of extracting complex, multi-dimensional patterns. The latter is one of the simplest classification algorithms, providing a basis for comparison.

2.7. Metrics

The metrics used for classification quality measurement were accuracy, precision, recall, and F 1 score. The metrics are described by Equations (10)–(13). The following abbreviations are used to simplify these equations:
  • T P = true positive;
  • T N = true negative;
  • F P = false positive;
  • F N = false negative.
A c c u r a c y = T P + T N T P + F P + T N + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 T P 2 T P + F N + F P
The accuracy is a global evaluation metric, and it assesses the model compared to all the data. On the other hand, the precision, recall, and F 1 score are local metrics that evaluate performance regarding the classification of one class vs. all of them.

3. Results

The results of the experiments described in Section 2.4, Section 2.5 and Section 2.6 are presented in Table 2, Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8. Each experiment was repeated 30 times with randomly selected training and test sets. The tables present the averaged metric values from 30 trials. Section 3.1 presents the values acquired from running experiments on images with and without obfuscation without adding data augmentation. Section 3.3 examines the influence of data augmentation. The experiments were run on the images with and without obfuscation, and data augmentation was applied. The results were presented for the channel, and the experimental results presented in Section 3.1 contained the most information.

3.1. Background Information Measurement

This subsection presents the information measurements in the image’s background in the form of the classification quality and compares it against the information in the whole image. Table 2 contains an evaluation of the distribution difference measurement described in Section 2.4. Table 3 presents the results from evaluating the Shannon entropy of the value distribution as a sole healthy or sick determinant. The experiment procedure is described in Section 2.5. Table 4 contains information on the machine learning algorithm’s performance for the value distribution.
Each table contains the column “Lymphocytes Obfuscated”. The rows with the value “True” in this column contain results from experiments with parts of the image representing lymphocytes covered by a black rectangle. As a result, these rows depict the informativeness of the image background. The ones with “False” in the first column contain results from experiments that used whole images and provide the informativeness measurements of all images. These serve as a basis of comparison for the experiments with obfuscated lymphocytes.
Each table contains the column “Channel” as well. This column presents the information regarding channel selection for the experiment. Each experiment evaluated the informativeness of only one channel at a time to establish whether the background in any channel contained unwanted information.
Figure 8 represents the averaged value distribution obtained from each channel for each of four states: images of healthy patients without lymphocytes obfuscated, images of healthy patients with lymphocytes obfuscated, images of sick patients without lymphocytes obfuscated, and images of sick patients with lymphocytes obfuscated. These four states are represented in their respective columns. The rows of the chart grid represent each of the six channels: red, green, blue, hue, saturation, and value.

3.2. Comparison with the Literature

A comparison of the most promising models obtained in this research with other works is presented in Table 5. The results were compared against the outcomes of our previous work and work of Rodrigues et al. [33,35], which according to our literature review obtained the best results on the ALL-IDB2 dataset.

3.3. Influence of Data Augmentation

Table 2, Table 3 and Table 4 prove that the channel containing the most informative image background was the hue channel, with the classification accuracy obtained from merely a background hue value distribution ranging from 82% to 93%. As a result, this channel was subjected to further investigation to determine the data augmentation application’s influence on obfuscated and unmodified images.
Table 6 contains the results of applying Gaussian blur with kernel sizes of 3, 9, 21, and 51. Table 7 presents the results of median blur application with kernel sizes of 3, 9, 21, and 51. Table 8 presents influence of Gaussian noise application on the images with variances of 0.0001, 0.01, and 0.1. The experiments were conducted with and without lymphocyte obfuscation. Each experiment was repeated 30 times, and its values were averaged.

4. Discussion

The experimental results presented in Table 2, Table 3 and Table 4 determined that the hue channel contained the highest amount of image background information. All methods (except for logistic regression) achieved averaged test accuracies above 80%, with the XGBoost model having 93% accuracy. Such scores were obtained merely for the hue distribution of the image background, with the informativeness unconfirmed by academic knowledge.
The methods for background information measurement described in Section 2.4, Section 2.5 and Section 2.6 proved to be efficient in determining whether the background contained classification-sensitive information. These methods can be used as training dataset evaluation techniques. Suppose that a supposedly neutral classification-wise background contains the required information. In this case, artificial intelligence models such as deep convolutional neural networks may learn to recognize some unwanted, dataset-related temporal pattern in the background instead of true generalization to real-world scenarios. This method can help in the evaluation of datasets containing high-quality data.
The logistic regression model trained on raw distributions achieved the worst accuracy, and the XGBoost model achieved the best accuracy. This suggests that, although the information contained in the distribution of the values is substantial, it is not yet obvious. Such background information may not be caught during exploratory data analysis and interfere with machine learning algorithms’ training quality. For these reasons, the background informativeness evaluation may prove beneficial in fool-proofing artificial intelligence systems.
Data augmentation techniques reduced the background informativeness extracted by the application of cross-entropy. It had little effect on the Shannon entropy-based classification and did not affect the XGBoost or MSE classification quality. This suggests that information in the background of ALL-IDB images may be more complex than just random class-specific noise. The authors plan to investigate this phenomenon further.
The research indicates that information is contained in the hue distribution of ALL-IDB image backgrounds. The XGBoost model achieved 93% accuracy on merely the hue distribution in the background. Such high classification quality has been achieved by just studying the background, which is supposed to be classification-neutral. According to our literature review, the primary indication of acute lymphoblastic leukemia is an examination of the lymphocytes. Academics do not unanimously recognize the lymphocytes’ surroundings’ informativeness. However, this research proves that this is not the case in this dataset. It is possible that the suspected “classification-neutral background” contains information allowing for healthy and sick discrimination. The authors plan to investigate this phenomenon further.
In our previous work [35], the best combination of artificial neural networks for raw image encoding, classification heads, and image augmentation resulted in an average classification quality of 94.8%. The neural network used for this task was MobileNet v2, the state-of-the-art neural network for numerous image-processing tasks containing 3.4 million parameters. In this work, the XGBoost algorithm alone, which interpreted the green color value distribution, achieved a classification accuracy of 96.0%. A similar result was obtained by the XGBoost algorithm interpreting the hue distributions of images with lymphocytes obfuscated, achieving a classification quality of 93.0%. Much simpler machine learning models operating on limited data obtained results comparable to the state-of-the-art deep learning method. This suggests that the neural networks experienced overfitting during training despite the application of data augmentation techniques. This also suggests that the task of leukemia classification may be performed using much more straightforward and cost-effective methods that also benefit from human interpretability. The authors plan to investigate this phenomenon further. According to our literature review, at the moment of writing this article, the best result was obtained by Rodrigues et al. [33], with an accuracy of 98.5%. This result is 2.5% percentage points higher than our best model. However, we obtained our results using around 3000 times fewer input data and almost 4000 times fewer parameters. As a result, our best model obtained comparable results, requiring much less computational power while remaining interpretable by humans.
The results indicate that hue distribution of a lymphocyte’s surroundings contains information supporting leukemia classification. However, a distribution is, by definition, an aggregation of the information stripped from temporal patterns akin to hidden Markov chains. It is possible that more detailed information interpretable by humans can be found during image examination in the hue channel. The authors plan to investigate this claim further.

5. Conclusions

The proposed background informativeness measurement proved its efficiency in dataset quality evaluation. This method based on the Shannon entropy, cross-entropy, and machine learning algorithms provides a comprehensive estimation of value distribution patterns in the background that may cause artificial intelligence models to overfit them instead of finding generalized solutions applicable to real-world problems.
The conducted research on background informativeness on the ALL-IDB dataset found a substantial amount of information in the hue distribution of the image background. The hue distributions of healthy people and patients who had acute lymphoblastic leukemia differed vastly from each other and by their Shannon entropy measurements. In this research, the lymphocytes were obfuscated with a black rectangle, so this information was contained within the supposedly classification-neutral background. The authors plan to investigate this phenomenon further.
The highest quality of classification was achieved while examining the green channel distribution using the XGBoost model. On average, it achieved 96.0% accuracy. This is a result comparable with deep neural networks while requiring much less computational power and providing a more human-interpretable decision process.
The background hue distribution differences between the images of healthy and sick patients require further investigation. It is unknown whether the differentiable factor is spread uniformly over the whole picture or is concentrated around semantically separable entities. Medical professionals must examine the nature of these changes to understand the features’ origins and extrapolate the applicability of this knowledge. The authors plan to investigate this phenomenon further.

Author Contributions

Conceptualization, K.P. and D.L.; methodology, K.P. and D.L., software K.P.; validation K.P. and D.L.; formal analysis, K.P., D.L. and T.A.; investigation K.P., D.L. and T.A.; resources, K.P., D.L. and T.A.; data curation, T.A.; writing—original draft preparation, K.P. and D.L.; writing—review and editing, K.P. and T.A.; visualization, K.P and D.L.; project administration, K.P. and T.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ALLAcute lymphoblastic leukemia
AMLAcute myleoid leukemia
CLLChronic lymphocytic leukemia
CLMChronic myleoid leukemia
CNNConvolutional neural network

References

  1. Andrade, A.R.; Vogado, L.H.; Veras, R.D.M.S.; Silva, R.R.; Araujo, F.H.; Medeiros, F.N. Recent computational methods for white blood cell nuclei segmentation: A comparative study. Comput. Methods Programs Biomed. 2019, 173, 1–14. [Google Scholar] [CrossRef] [PubMed]
  2. Labati, R.D.; Piuri, V.; Scotti, F. The Acute Lymphoblastic Leukemia Image Database for Image Processing. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011. [Google Scholar]
  3. Mohamed, M.; Far, B.; Guaily, A. An efficient technique for white blood cells nuclei automatic segmentation. In Proceedings of the 2012 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Seoul, Korea, 14–17 October 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 220–225. [Google Scholar]
  4. Sarrafzadeh, O.; Rabbani, H.; Talebi, A.; Banaem, H.U. Selection of the best features for leukocytes classification in blood smear microscopic images. In Proceedings of the Medical Imaging 2014: Digital Pathology, San Diego, CA, USA, 15–20 February 2014; SPIE: Bellingham, WA, USA, 2014; Volume 9041, pp. 159–166. [Google Scholar]
  5. Zheng, X.; Wang, Y.; Wang, G.; Liu, J. Fast and robust segmentation of white blood cell images by self-supervised learning. Micron 2018, 107, 55–71. [Google Scholar] [CrossRef] [PubMed]
  6. Madhloom, H.; Kareem, S.; Ariffin, H.; Zaidan, A.; Alanazi, H.; Zaidan, B. An automated white blood cell nucleus localization and segmentation using image arithmetic and automatic threshold. J. Appl. Sci. 2010, 10, 959–966. [Google Scholar] [CrossRef] [Green Version]
  7. Arslan, S.; Ozyurek, E.; Gunduz-Demir, C. A color and shape based algorithm for segmentation of white blood cells in peripheral blood and bone marrow images. Cytom. Part A 2014, 85, 480–490. [Google Scholar] [CrossRef]
  8. Nazlibilek, S.; Karacor, D.; Ercan, T.; Sazli, M.H.; Kalender, O.; Ege, Y. Automatic segmentation, counting, size determination and classification of white blood cells. Measurement 2014, 55, 58–65. [Google Scholar] [CrossRef]
  9. Prinyakupt, J.; Pluempitiwiriyawej, C. Segmentation of white blood cells and comparison of cell morphology by linear and naïve Bayes classifiers. Biomed. Eng. Online 2015, 14, 63. [Google Scholar] [CrossRef] [Green Version]
  10. Nasir, A.A.; Mashor, M.; Rosline, H. Unsupervised colour segmentation of white blood cell for acute leukaemia images. In Proceedings of the 2011 IEEE International Conference on Imaging Systems and Techniques, Penang, Malaysia, 17–18 May 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 142–145. [Google Scholar]
  11. Mohapatra, S.; Samanta, S.S.; Patra, D.; Satpathi, S. Fuzzy based blood image segmentation for automated leukemia detection. In Proceedings of the 2011 International Conference on Devices and Communications (ICDeCom), Ranchi, India, 24–25 February 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1–5. [Google Scholar]
  12. Madhukar, M.; Agaian, S.; Chronopoulos, A.T. New decision support tool for acute lymphoblastic leukemia classification. In Proceedings of the Image Processing: Algorithms and Systems X; and Parallel Processing for Imaging Applications II, Burlingame, CA, USA, 23–25 January 2012; SPIE: Bellingham, WA, USA, 2012; Volume 8295, pp. 367–378. [Google Scholar]
  13. Amin, M.M.; Kermani, S.; Talebi, A.; Oghli, M.G. Recognition of acute lymphoblastic leukemia cells in microscopic images using k-means clustering and support vector machine classifier. J. Med. Signals Sens. 2015, 5, 49. [Google Scholar]
  14. Sarrafzadeh, O.; Dehnavi, A.M.; Rabbani, H.; Talebi, A. A simple and accurate method for white blood cells segmentation using K-means algorithm. In Proceedings of the 2015 IEEE Workshop on Signal Processing Systems (SiPS), Hangzhou, China, 14–16 October 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–6. [Google Scholar]
  15. Vincent, I.; Kwon, K.R.; Lee, S.H.; Moon, K.S. Acute lymphoid leukemia classification using two-step neural network classifier. In Proceedings of the 2015 21st Korea-Japan Joint Workshop on Frontiers of Computer Vision (FCV), Mokpo, Korea, 28–30 January 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–4. [Google Scholar]
  16. Vogado, L.H.; Veras, R.D.M.S.; Andrade, A.R.; e Silva, R.R.; De Araujo, F.H.; De Medeiros, F.N. Unsupervised leukemia cells segmentation based on multi-space color channels. In Proceedings of the 2016 IEEE International Symposium on Multimedia (ISM), San Jose, CA, USA, 11–13 December 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 451–456. [Google Scholar]
  17. Kumar, P.; Vasuki, S. Automated diagnosis of acute lymphocytic leukemia and acute myeloid leukemia using multi-SV. J. Biomed. Imaging Bioeng. 2017, 1, 20–24. [Google Scholar]
  18. Mohammed, E.A.; Mohamed, M.M.; Naugler, C.; Far, B.H. Chronic lymphocytic leukemia cell segmentation from microscopic blood images using watershed algorithm and optimal thresholding. In Proceedings of the 2013 26th IEEE Canadian Conference on Electrical and Computer Engineering (CCECE), Regina, SK, Canada, 5–8 May 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 1–5. [Google Scholar]
  19. Abdeldaim, A.M.; Sahlol, A.T.; Elhoseny, M.; Hassanien, A.E. Computer-aided acute lymphoblastic leukemia diagnosis system based on image analysis. In Advances in Soft Computing and Machine Learning in Image Processing; Springer: Berlin/Heidelberg, Germany, 2018; pp. 131–147. [Google Scholar]
  20. Rehman, A.; Abbas, N.; Saba, T.; Rahman, S.I.U.; Mehmood, Z.; Kolivand, H. Classification of acute lymphoblastic leukemia using deep learning. Microsc. Res. Tech. 2018, 81, 1310–1317. [Google Scholar] [CrossRef]
  21. Prellberg, J.; Kramer, O. Acute lymphoblastic leukemia classification from microscopic images using convolutional neural networks. In ISBI 2019 C-NMC Challenge: Classification in Cancer Cell Imaging; Springer: Berlin/Heidelberg, Germany, 2019; pp. 53–61. [Google Scholar]
  22. Ahmed, N.; Yigit, A.; Isik, Z.; Alpkocak, A. Identification of leukemia subtypes from microscopic images using convolutional neural network. Diagnostics 2019, 9, 104. [Google Scholar] [CrossRef] [Green Version]
  23. Guo, Z.; Wang, Y.; Liu, L.; Sun, S.; Feng, B.; Zhao, X. Siamese Network-Based Few-Shot Learning for Classification of Human Peripheral Blood Leukocyte. In Proceedings of the 2021 IEEE 4th International Conference on Electronic Information and Communication Technology (ICEICT), Xi’an, China, 18–20 August 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 818–822. [Google Scholar]
  24. Abhishek, A.; Jha, R.K.; Sinha, R.; Jha, K. Automated classification of acute leukemia on a heterogeneous dataset using machine learning and deep learning techniques. Biomed. Signal Process. Control 2022, 72, 103341. [Google Scholar] [CrossRef]
  25. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  26. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  27. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. arXiv 2016, arXiv:1611.05431. [Google Scholar]
  28. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef] [Green Version]
  29. Koch, G.; Zemel, R.; Salakhutdinov, R. Siamese neural networks for one-shot image recognition. In Proceedings of the ICML deep learning workshop, Lille, France, 6–11 July 2015; Volume 2. [Google Scholar]
  30. Wang, Y.; Yao, Q.; Kwok, J.T.; Ni, L.M. Generalizing from a few examples: A survey on few-shot learning. ACM Comput. Surv. 2020, 53, 63. [Google Scholar] [CrossRef]
  31. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  32. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; IEEE: Piscataway, NJ, USA, 2005; Volume 1, pp. 886–893. [Google Scholar]
  33. Rodrigues, L.F.; Backes, A.R.; Travençolo, B.A.N.; de Oliveira, G.M.B. Optimizing a Deep Residual Neural Network with Genetic Algorithm for Acute Lymphoblastic Leukemia Classification. J. Digit. Imaging 2022, 35, 623–637. [Google Scholar] [CrossRef]
  34. Hinton, G.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  35. Pałczyński, K.; Śmigiel, S.; Gackowska, M.; Ledziński, D.; Bujnowski, S.; Lutowski, Z. IoT Application of Transfer Learning in Hybrid Artificial Intelligence Systems for Acute Lymphoblastic Leukemia Classification. Sensors 2021, 21, 8025. [Google Scholar] [CrossRef]
  36. Golik, P.; Doetsch, P.; Ney, H. Cross-Entropy vs. Squared Error Training: A Theoretical and Experimental Comparison. In Proceedings of the International Sport and Culture Association, Lyon, France, 25–29 August 2013; pp. 1756–1760. [Google Scholar] [CrossRef]
  37. Peng, J.; Lee, K.; Ingersoll, G. An Introduction to Logistic Regression Analysis and Reporting. J. Educ. Res. 2002, 96, 3–14. [Google Scholar] [CrossRef]
  38. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. arXiv 2016, arXiv:1603.02754. [Google Scholar]
Figure 1. A black rectangle covers an example of lymphocytes in the images in the second column. The first row presents versions of the image of a healthy patient, and the second row depicts versions of the photo taken for a patient sick with leukemia.
Figure 1. A black rectangle covers an example of lymphocytes in the images in the second column. The first row presents versions of the image of a healthy patient, and the second row depicts versions of the photo taken for a patient sick with leukemia.
Entropy 24 01560 g001
Figure 2. Presentation of the Gaussian blurring application on the images. The blurring procedure did not affect the first image in the row, and the rest were convolved with kernel of sizes 3, 9, 21, and 51 pixels, respectively.
Figure 2. Presentation of the Gaussian blurring application on the images. The blurring procedure did not affect the first image in the row, and the rest were convolved with kernel of sizes 3, 9, 21, and 51 pixels, respectively.
Entropy 24 01560 g002
Figure 3. Presentation of the median blurring application on the images. The blurring procedure did not affect the first image in the row, and the rest were convolved with kernel of sizes 3, 9, 21, and 51 pixels, respectively. The images were 257 × 257 pixels in size.
Figure 3. Presentation of the median blurring application on the images. The blurring procedure did not affect the first image in the row, and the rest were convolved with kernel of sizes 3, 9, 21, and 51 pixels, respectively. The images were 257 × 257 pixels in size.
Entropy 24 01560 g003
Figure 4. Presentation of the Gaussian noise application on the images. The noise procedure did not affect the first image in the row, and the rest were subjected to the multiplicative noise sampled from the Gaussian distribution, parametrized by the mean equal to zero and variance containing values of 0.001, 0.01, and 0.1, respectively.
Figure 4. Presentation of the Gaussian noise application on the images. The noise procedure did not affect the first image in the row, and the rest were subjected to the multiplicative noise sampled from the Gaussian distribution, parametrized by the mean equal to zero and variance containing values of 0.001, 0.01, and 0.1, respectively.
Entropy 24 01560 g004
Figure 5. Graphical representation of image processing procedure.
Figure 5. Graphical representation of image processing procedure.
Entropy 24 01560 g005
Figure 6. A presentation of images taken from healthy and sick patients (first column) and their corresponding hue distributions. HSV encoding was applied to both of the images.
Figure 6. A presentation of images taken from healthy and sick patients (first column) and their corresponding hue distributions. HSV encoding was applied to both of the images.
Entropy 24 01560 g006
Figure 7. A graphical representation of the classification process involving images’ color distribution comparison, using cross-entropy and mean squared error (MSE) as divergence metrics.
Figure 7. A graphical representation of the classification process involving images’ color distribution comparison, using cross-entropy and mean squared error (MSE) as divergence metrics.
Entropy 24 01560 g007
Figure 8. Presentation of averaged distributions of each channel’s values for each combination of images. The columns represent distributions taken from images of healthy patients without lymphocytes covered, images of healthy patients with lymphocytes covered, images of sick patients without lymphocytes covered, and images of sick patients with lymphocytes covered. Each row represents one channel. The presented channels are red (R), green (G), blue (B), hue (H), saturation (S), and value (V).
Figure 8. Presentation of averaged distributions of each channel’s values for each combination of images. The columns represent distributions taken from images of healthy patients without lymphocytes covered, images of healthy patients with lymphocytes covered, images of sick patients without lymphocytes covered, and images of sick patients with lymphocytes covered. Each row represents one channel. The presented channels are red (R), green (G), blue (B), hue (H), saturation (S), and value (V).
Entropy 24 01560 g008
Table 1. A summary of investigated research works.
Table 1. A summary of investigated research works.
Ref.Image EncodingMethods
[3]GrayscaleArithmetical operations
[6]GrayscaleOtsu threshold, arithmetical operations
[7]RGBOtsu threshold, region growing
[8]GrayscaleOtsu threshold
[9]RGBOtsu threshold, arithmetical operations
[10]HSIK-means, region growing
[11]L*a*b*K-means
[12]L*a*b*K-means
[13]HSVK-means
[14]L*a*b*K-means, arithmetical operations
[15]L*a*b*K-means
[16]CMYK + L*a*b*K-means
[17]HSVK-means
[18]GrayscaleRegion growing, edge detectors
[19]CMYKZack’s algorithm
[20]RGBAlexNet
[21]RGBResNeXt50
[22]RGBCNN
[23]RGBSiamese networks
[24]RGBCNN
[33]RGBResNet50 v2
Table 2. The classification results by the images’ color distribution comparison to the averaged distributions representing their respective classes. The cross-entropy and mean squared error (MSE) metrics were applied. The images were unmodified and had their lymphocytes covered (first column). No image augmentation was applied in this experiment.
Table 2. The classification results by the images’ color distribution comparison to the averaged distributions representing their respective classes. The cross-entropy and mean squared error (MSE) metrics were applied. The images were unmodified and had their lymphocytes covered (first column). No image augmentation was applied in this experiment.
Lymphocytes
Obfuscated
ChannelCross
Entropy Acc.
Cross
Entropy F 1
(Healthy)
Cross
Entropy F 1
(Sick)
MSE Acc.MSE F 1
(Healthy)
MSE F 1
(Sick)
FalseB0.500.560.410.680.680.68
TrueB0.550.590.490.690.680.70
FalseG0.810.810.810.620.630.61
TrueG0.470.530.380.600.620.58
FalseR0.530.510.540.460.490.43
TrueR0.370.490.180.450.480.42
FalseHue0.850.860.830.770.790.73
TrueHue0.830.850.810.820.840.80
FalseSaturation0.790.800.770.790.830.72
TrueSaturation0.730.720.740.790.830.73
FalseValue0.450.510.360.500.510.48
TrueValue0.380.450.280.520.530.49
Table 3. The classification results by the images’ color distributions’ Shannon entropy measurements. The images were unmodified and had their lymphocytes covered (first column). No image augmentation was applied in this experiment.
Table 3. The classification results by the images’ color distributions’ Shannon entropy measurements. The images were unmodified and had their lymphocytes covered (first column). No image augmentation was applied in this experiment.
Lymphocytes
Obfuscated
ChannelAvg. Shannon
Entropy
(Healthy)
Std. Shannon
Entropy
(Healthy)
Avg. Shannon
Entropy
(Sick)
Std. Shannon
Entropy
(Sick)
Acc F 1
(Healthy)
F 1
(Sick)
FalseB0.430.010.450.020.510.310.51
TrueB0.430.020.440.020.510.300.51
FalseG0.480.020.500.020.550.370.55
TrueG0.430.030.460.030.540.390.54
FalseR0.440.020.470.020.540.380.54
TrueR0.390.030.420.040.540.390.54
FalseHue0.630.030.710.070.680.700.65
TrueHue0.570.030.690.100.810.840.78
FalseSaturation0.490.030.510.020.550.380.55
TrueSaturation0.440.030.470.040.510.360.51
FalseValue0.430.020.440.020.510.300.51
TrueValue0.390.030.410.030.520.340.53
Table 4. The classification results by images’ color distribution interpretation by machine learning algorithms. The XGBoost and logistic regression models were used in this experiment. The images were unmodified and had their lymphocytes covered (first column). No image augmentation was applied in this experiment.
Table 4. The classification results by images’ color distribution interpretation by machine learning algorithms. The XGBoost and logistic regression models were used in this experiment. The images were unmodified and had their lymphocytes covered (first column). No image augmentation was applied in this experiment.
Lymphocytes
Obfuscated
ChannelXGBoost AccXGBoost F 1
(Healthy)
XGBoost F 1
(Sick)
Logistic Regression Acc.Logistic
Regression
F 1 (Healthy)
Logistic
Regression
F 1 (Sick)
FalseB0.830.830.830.530.320.53
TrueB0.820.820.820.550.370.55
FalseG0.960.960.960.500.280.51
TrueG0.860.870.850.500.280.51
FalseR0.870.870.860.470.230.49
TrueR0.800.800.790.480.280.49
FalseHue0.940.950.940.570.410.57
TrueHue0.930.940.930.600.460.59
FalseSaturation0.940.940.940.540.320.54
TrueSaturation0.880.890.880.550.360.55
FalseValue0.840.840.840.480.250.49
TrueValue0.800.810.800.500.300.50
Table 5. Comparison of the best models with and without lymphocytes obfuscated against other works. Shortcut “dist” stands for “distribution”.
Table 5. Comparison of the best models with and without lymphocytes obfuscated against other works. Shortcut “dist” stands for “distribution”.
ArticleInput DataInput SizeObfuscationParametersAccuracyPrecisionRecall F 1
This workGreen dist.51False6.4 K0.9600.9600.9600.959
This workHue dist.36True6.4 K0.9350.9350.9350.934
[35]RGB image150 KFalse3.4 M0.9480.9500.9510.948
[33]RGB image150 KFalse25 M0.9850.9860.9850.984
Table 6. Results of Gaussian blur application on the quality of classification.
Table 6. Results of Gaussian blur application on the quality of classification.
Lymphocytes ObfuscatedKernelXGBoost Acc.Logistic Regression Acc.Cross-Entropy Acc.kMSE Acc.Shannon Acc.
False510.930.600.790.740.64
False210.930.580.790.770.69
False00.940.570.850.770.68
False90.950.580.830.780.69
False30.950.570.830.770.69
True510.910.650.820.830.79
True210.930.630.780.820.82
True90.930.620.830.810.82
True30.930.610.820.820.81
True00.930.600.830.820.81
Table 7. Results of median blur application on the quality of classification.
Table 7. Results of median blur application on the quality of classification.
Lymphocytes ObfuscatedKernelXGBoost Acc.Logistic Regression Acc.Cross-Entropy Acc.kMSE Acc.Shannon Acc.
False510.940.590.730.770.67
False00.940.570.850.770.68
False90.950.580.810.790.69
False210.950.590.810.790.69
False30.950.570.830.770.69
True510.910.630.720.800.80
True30.930.610.820.820.81
True90.930.630.810.810.82
True00.930.600.830.820.81
True210.940.650.780.810.81
Table 8. Results of multiplicative Gaussian noise application on the quality of classification.
Table 8. Results of multiplicative Gaussian noise application on the quality of classification.
Lymphocytes ObfuscatedKernelXGBoost Acc.Logistic Regression Acc.Cross-Entropy Acc.kMSE Acc.Shannon Acc.
False00.940.570.840.760.68
False0.00010.950.570.840.760.68
False0.010.950.560.840.760.68
False0.10.950.560.820.760.68
True0.010.930.590.820.820.80
True0.00010.930.600.830.820.81
True00.930.600.820.820.81
True0.10.950.590.820.820.81
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Pałczyński, K.; Ledziński, D.; Andrysiak, T. Entropy Measurements for Leukocytes’ Surrounding Informativeness Evaluation for Acute Lymphoblastic Leukemia Classification. Entropy 2022, 24, 1560. https://doi.org/10.3390/e24111560

AMA Style

Pałczyński K, Ledziński D, Andrysiak T. Entropy Measurements for Leukocytes’ Surrounding Informativeness Evaluation for Acute Lymphoblastic Leukemia Classification. Entropy. 2022; 24(11):1560. https://doi.org/10.3390/e24111560

Chicago/Turabian Style

Pałczyński, Krzysztof, Damian Ledziński, and Tomasz Andrysiak. 2022. "Entropy Measurements for Leukocytes’ Surrounding Informativeness Evaluation for Acute Lymphoblastic Leukemia Classification" Entropy 24, no. 11: 1560. https://doi.org/10.3390/e24111560

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop