Next Article in Journal
Development and Implementation of an Internal Quality Control Sample to Standardize Oligomer-Based Diagnostics of Alzheimer’s Disease
Next Article in Special Issue
Generalist Vision Foundation Models for Medical Imaging: A Case Study of Segment Anything Model on Zero-Shot Medical Segmentation
Previous Article in Journal
Clustered Occurrence of Osteitis Condensans Ilii in Patients with Symptomatic Hip Dysplasia
Previous Article in Special Issue
Deep-EEG: An Optimized and Robust Framework and Method for EEG-Based Diagnosis of Epileptic Seizure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection and Classification of Histopathological Breast Images Using a Fusion of CNN Frameworks

by
Ahsan Rafiq
1,
Alexander Chursin
2,
Wejdan Awad Alrefaei
3,
Tahani Rashed Alsenani
4,
Ghadah Aldehim
5,*,
Nagwan Abdel Samee
6 and
Leila Jamel Menzli
5
1
School of Automation, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
2
Higher School of Industrial Policy and Entrepreneurship, RUDN University, 6 Miklukho-Maklaya St, Moscow 117198, Russia
3
Department of Programming and Computer Sciences, Applied College in Al-Kharj, Prince Sattam Bin Abdulaziz University, Al-Kharj 16245, Saudi Arabia
4
Department of Biology, College of Sciences in Yanbu, Taibah University, Yanbu 46522, Saudi Arabia
5
Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
6
Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(10), 1700; https://doi.org/10.3390/diagnostics13101700
Submission received: 27 February 2023 / Revised: 7 April 2023 / Accepted: 20 April 2023 / Published: 11 May 2023

Abstract

:
Breast cancer is responsible for the deaths of thousands of women each year. The diagnosis of breast cancer (BC) frequently makes the use of several imaging techniques. On the other hand, incorrect identification might occasionally result in unnecessary therapy and diagnosis. Therefore, the accurate identification of breast cancer can save a significant number of patients from undergoing unnecessary surgery and biopsy procedures. As a result of recent developments in the field, the performance of deep learning systems used for medical image processing has showed significant benefits. Deep learning (DL) models have found widespread use for the aim of extracting important features from histopathologic BC images. This has helped to improve the classification performance and has assisted in the automation of the process. In recent times, both convolutional neural networks (CNNs) and hybrid models of deep learning-based approaches have demonstrated impressive performance. In this research, three different types of CNN models are proposed: a straightforward CNN model (1-CNN), a fusion CNN model (2-CNN), and a three CNN model (3-CNN). The findings of the experiment demonstrate that the techniques based on the 3-CNN algorithm performed the best in terms of accuracy (90.10%), recall (89.90%), precision (89.80%), and f1-Score (89.90%). In conclusion, the CNN-based approaches that have been developed are contrasted with more modern machine learning and deep learning models. The application of CNN-based methods has resulted in a significant increase in the accuracy of the BC classification.

1. Introduction

Mammary cancer, commonly known as breast cancer (BC), is a form of life-threatening cancer that primarily affects women. When compared to other forms of cancer, the mortality rate associated with BC for female patients is the second highest [1]. Breast cancer is a collection of cells that begin in a woman’s breast and have the ability to rapidly spread to any other organ in the body [2]. These cells have grown in an uncontrolled manner. There are many subtypes of cancer, but the cancers that most commonly strike humans are lung cancer, breast cancer, and skin cancer. The World Health Organization (WHO) mentioned that the mortality ratio from cancer is up to 7.2 million for lung cancer, 1.3 million fatalities for skin cancer, and 512,030 deaths from breast cancer [3,4]. This statistic refers to the number of people who have died from lung cancer. Breast cancer patients who have tumors that are even less than 10 mm in size have a 98% chance of surviving the disease [5]. Because of this, the likelihood of successfully surviving breast cancer is strongly connected with the volume of the tumor. The extent of a breast tumor is one of the primary factors that determines whether a patient will make it through their cancer treatment. There are different ways of thinking about BC that can be taken into consideration in order to recognize and categorize it. Imaging techniques such as ultrasound [6], X-rays [7], digital mammograms [8,9,10,11,12,13,14], and CT scans [15,16] are more useful. Researchers are able to identify the many types of cancer utilizing a number of techniques, such as early-phase screening (EPS), etc. In addition to this, they have developed original strategies for the diagnosis and treatment of cancer at earlier stages, which has improved patient outcomes. Large volumes of information relating to cancer have been accumulated as a result of the development of breakthrough technologies in medicine, and this information has been made available in the field of bioinformatics as well as the scientific community for the purposes of investigation and analysis. Predicting whether a patient will get breast cancer is, nevertheless, one of the most challenging and tough tasks in the healthcare profession. This includes the danger of incorrect classification when utilizing the diagnostic procedures that have been investigated previously. In addition, a diagnosis of breast cancer in a clinic is dependent on the visual examination of patients by pathologists, who are highly trained specialists in the field of cancer. Pathologists study cancer extensively and have extensive clinical experience. On the other hand, this procedure is performed manually, which is not only labor-intensive but also time-consuming and prone to human mistakes. This is because the vast majority of cells have arbitrary shapes, differing portions, and uncontrolled visual angles. This is the reason why this occurs. Before beginning early therapy, it is critical to determine if a tumor is cancerous (also known as malignant) or non-cancerous (also known as benign).
The use of machine learning (ML) [17,18,19,20] and deep learning (DL) [21,22] methodologies for the automatic categorization of BC has assisted in decreasing the risk of getting cancer and recurrence, and survival prediction might increase the accuracy by 20% to 25% more than it did in the previous year [23]. In the identification of invasive breast cancer, ML/DL is a technique that sees widespread application [8,9,10,11,12,13,14,24]. The results that have been obtained through DL-based models have been superior to those that have been obtained through the more conventional approaches to machine learning [10]. In most cases, the application of DL approaches [21,22,25,26,27,28] is founded on the use of an only one deep learning algorithm including RNN, LSTM, or CNN. Even though these DL-based models have obtained good performance, the employment of hybrid deep learning models [29,30,31,32,33,34,35,36,37,38] has assisted in improving the achieved classification performance. This was possible owing to the assistance provided by hybrid deep learning models. Because of this, we decided to propose a fusion-based CNN model to categorize breast cancer histology.
The following is a list of the most important contributions that may be summarized as follows:
  • A fusion-based CNN model is presented here for the purpose of classifying the histology of breast cancer.
  • The presented models are applied in clinical and biomedical research with the goal of developing a successful identification method for breast cancer in breast tissue.
  • The models that have been built are predicated respectively on the convolutional neural network (CNN) model, the fusion of two CNNs (often referred to as 2-CNN), and three CNNs (3-CNN).
When compared to state-of-the-art approaches, the proposed model performed significantly better when classifying breast cancer histology. The suggested 3-CNN model’s accuracy has outperformed, by a percentage point of 30%, that of the conventional ML approaches, ANN, for the binary classification (benign/malignant) of breast tissue. While the current CNN-based model for the binary classification of breast tissues is quite accurate, the suggested 3-CNN model introduces an improvement, 1%, in the model’s accuracy. The paper is arranged as follows: the first section includes the experimental dataset, and its preprocessing. In Section 2, we explore the various techniques utilized using CNN-based fusion models. In Section 3, we offer the experimental study of the recommended methodologies, which covers ML and DL approaches, and in Section 4, we state the results obtained and we discuss the findings. In Section 5, the conclusion and recommendations for additional research are given.

Related Work

Despite the numerous research that has been undertaken and published, as well as the substantial breakthroughs in the automatic identification of BC features, a number of issues have led to major difficulties in image categorization [39]. These factors include the inconsistency in tissue and cell shape, the phenomenon of cell overlapping, the visual heterogeneity of stained histopathological analysis, and inconsistent color distribution. Despite these factors, significant advances have been made in the automatic detection of BC characteristics [39]. These issues pose significant challenges in the accurate automatic detection of BC [40]. The problem is due to high resolution of pathologist images (PI), making it difficult to automatically transplant some of the effective techniques from natural images to PI. The majority of the early classification methods for BC-PI were based on a binary system that divided BC lesions into two categories: cancerous and non-cancerous [10,41,42,43,44,45] or a more sophisticated tri-classification as invasive carcinoma, situ carcinoma, and normal [46,47]. Most of the work was done using whole images or on extracted nuclei relying on architectural, morphological, and textural aspects using the standard ML approach. It is in-depth mentioning that most of the classification methods were performed on small resolution images at various magnifications. Moreover, these approaches used artificial-rely on the extraction of feature methods, which not only require a significant amount of effort and professional domain expertise but also face certain problems in extracting distinguishing high-quality features. Consequently, employing traditional ML approaches is severely limited when it comes to the classification of BC. Finally, DL algorithms [17,48] achieved impressive success of wide range of biomedical applications [11,12,13,14].
The ML/DL method is one that is utilized frequently in the process of determining whether a patient has breast cancer [8,9,10,11,12,13,14,24]. The results that have been acquired through the use of DL-based models have been demonstrated to be superior to those that have been produced through the use of more conventional approaches to machine learning [10]. The application of DL techniques in healthcare [21,22,25,26,27,28] is typically predicated on the deployment of a unique deep learning model such as CNN, long short-term memory (LSTM), or recurrent neural network (RNN). This is the case in the vast majority of the situations. Despite the fact that these DL-based models have achieved good performance, the utilization of hybrid deep learning models [29,30,31,32,33,34,35,36,37,38] has contributed to the improvement of the classification performance that was accomplished.
Recently, a small group of researchers has argued that AI should be used to automatically detect and diagnose histopathological abnormalities in breast lesions. Recent studies have employed ML/DL algorithms to detect breast tumors; Table 1 highlights the current status of the field and the shortcomings of each technique. The most noteworthy aspects of the new system are highlighted in the table.
Among DL techniques, the CNN is the most important. CNNs are widely used in the classification of BC. Bayramoglu et al. [48] employed the magnification DL algorithm used BreaKHis dataset as well, with an 83% accuracy rate. Arajo et al. [49] conducted an initial investigation on multiple classifications for BC-PI. They were able to recover features by employing a CNN that was analogous to AlexNet, and after that, they used SVM to classify the characteristics that were gathered. On the other hand, a recurrent neural network (RNN) is only occasionally used in applications that include the classification of Histology Images (HI). In contrast to the RNN, the CNN is able to process data by using its own internal state. A multitude of remarkable CNN-based algorithms for the automated and precise categorization of BC have been recently developed [50] in preparation for the ICIAR2018 competition. These methodologies have been extremely significant in propelling the state of the art forward. These techniques’ core concepts are similar. After the high-resolution image (HI) has been pre-processed and data-enhanced, they are then split into patches of a similar size, and the features recovered using a CNN are then acquired from each patch individually. Vesal et al. [51] presented a transfer learning (TL) approach. They categorized each patch of one picture using the simple majority approach of ResNet and GoogLeNet. Vang et al. [52] were the first to suggest utilizing Google Inception accomplish patch-wise categorization. In order to arrive at an image-wise prediction, the patch-wise detection was put through an ensemble fusion architecture that includes logistic regression, gradient boosting, and majority voting. Deep convolutional feature representation was suggested by Rakhlin et al. [53]. The features have been extracted using a generic CNN. At the end, for the final classification result, they employed gradient enhanced trees. ResNet was used by Awan et al. [54] to generate 12-dimensional vector features representing twelve non-overlapped patches of 512 * 512. The final image-wise classification result was based on the majority vote result on the categorization of 1024 * 1024 pixel-overlapping blocks of patches.
In comparison to the present state of the art in utilizing machine and deep learning (ML/DL) to classify BC histology, the network demonstrates improved performance in properly and precisely classifying breast tissues, suggesting its increased application in this area. Moreover, it is unclear how the use of hybrid DL models aids in achieving a high degree of precision when classifying BC samples. The foregoing research demonstrates definitively that hybrid DL models outperform their single deep learning (DL) counterparts when it comes to BC detection and classification. To further enhance BCs’ ability to classify data, in view of the aforementioned information, we came up with a complex deep learning approach to classify BCs from HI pictures using a hybrid CNN learning model.

2. Material and Methods

This section describes the framework of histopathological BC cancer detection, including data collection and data augmentation methods. Furthermore, this section has been extended including the proposed methodology.

2.1. Dataset and Image Augmentation Methods

The dataset related to BC is obtained from Kaggle [55]. This dataset possesses originally PCam dataset, including no repeated samples. This collection has a total of 22,090 BC-HI with patches extracted from 200 scan slide images. These images were obtained from 162 women who had breast cancer and had it identified and tested at the University Medical Center in the Netherlands. The experimental training dataset, contains 170 WSIs BC-HI samples, and the training dataset includes 100 WSIs, are considered training datasets. The images in this collection have a dimension of 2040 pixels wide by 1536 pixels high. For the purpose of maintaining uniformity, each slide was scanned using the same scanner at a resolution of 0.25 micro/pr. Lower-resolution images were sampled to a more manageable 50 pixels by 50 pixels. Malignant and non-metastatic tumors (benign) are the two primary classes of BC in the Kaggle dataset. It includes a total of 220,025 scan images of BC, consisting of 130,893 non-cancers (benign) and 89,117 images of malignant cases (cancer images). The Kaggle data set is split into a training data set (80%) and a testing data set (20%). To avoid overfitting, the best results may be produced, according to empirical research, if just 20–30% of the data are used for testing, with the rest 70–80% of the data being used for training. Both classes’ samples in the RGB color space and the PNG file format are utilized for the image files and have label 0 for benign and 1 for malignant as shown in Figure 1.
As there were sufficient photos of the benign class in the data set utilized for this study, the augmented method was not applied to these images. The number of images that belonged to the malignant class, which was 89,117, was raised to 900,000. The accuracy of CNN architectures in the classification process can be improved through the application of data augmentation techniques [22,31,34] as shown in Figure 2.
The images that were used in this investigation were rotated in various ways before being examined. In this particular investigation, the rotation rate is set at 25. The original images as well as the augmented versions were used to compile the data set for this investigation. The only time the augmented images were shown was when the model was being trained. In order to prevent overfitting, the test makes use of the original data.

2.2. Proposed Model

For detecting and classifying tasks, the CNN based classifier (1-CNN) is adopted. Moreover, we presented further classifiers that fused two CNN models and three CNN models.
A.
The 1-CNN model
We have developed a basic 1-CNN as a method for distinguishing between malignant and benign cancer as shown in Figure 3. This 1-CNN model is put together using the following components:
  • Two convolutional layers, each having 32 filters and a kernel size of (3, 3), with the ReLU serving as the activation function. One of the convolutional layers has 32 filters with a kernel size of (3, 3).
  • A maximum pooling layer with the pool size set equal to (2, 2)
  • A layer with dropouts refers to the process of ignoring neurons during the training step of a randomly selected neurons set; the rate that we use is 0.25.
  • A flattened layer.
  • A dense layer where all one-layer inputs are connected to each succeeding layer’s activation functions. The dense layer has 64 levels with the ReLU serving as an activation function.
  • Another dropout layer having a rate of 0.5.
  • A dropout layer that has an activation function that is sigmoid.
  • The number of epochs is a hyperparameter that controls how many times the training process will iterate. When we talk about an epoch, we imply that every single training dataset has been given the chance to bring the parameters of the internal model up to date. In the course of our research, we utilized a total of thirty epochs.
B.
The 2-CNN model
For BC classification, we suggest a hybrid approach that combines two CNN models. Combining the results of many CNN architectures, each of which may be beneficial in extracting a certain feature from histopathological images of BC, may allow us to develop general features that are useful in the classification of BC.
The 2-CNN model is made up of two CNN models that are very close to one another and have the same architecture as the previous model. Figure 4 shows how the 1-CNN and 2-CNN models are combined; i.e., the two models are firstly concatenated; then, a dense layer employing a SoftMax activation function (AF) is included
C.
The 3-CNN model
The 3-CNN model is constructed using three CNN models that are quite close to one another; as shown in Figure 5, we combined the 1-CNN, 2-CNN, and 3-CNN models. In fact, once the first two CNN models are combined, a second combination with the third model is processed. At the end, a dense layer and a SoftMax AF are added.
D.
Optimizer
We use an optimizer through its paces in order to decrease the amount of error in the predictions (the loss function), as well as to improve the accuracy of the predictions and get them as close to optimal as possible. We employed an optimizer known as ADAM (Adaptive Moment Estimation), which is based on the first-order gradient of stochastic objective functions.

2.3. Experimental Setup

In this experimental study, we use a graphics processing unit (GPU) from NVIDIA and a processor with an Intel Core i6 architecture. In addition, the proposed model was trained using Python 3.8, Keras framework and other DL libraries, as shown in Table 2.

3. Results

This section describes the proposed CNN-based model performance based on key performance metrics (Accuracy, Recall, Precision, F1-score) and key metrics (such confusion matrix, ROC, and time complexity (ms)) in addition to more deeply explained comparative studies.

3.1. Performance Metrics

In order to evaluate the performance of a classifier, it is necessary to differentiate between four distinct types that have been assigned to the target category: true positives, false positives, true negatives, and false negatives.
  • TP: These are the positive values that have been accurately predicted, which indicates that the value of the real class is yes and that the value of the predicted class is also yes.
  • TN: These are the negative values that have been successfully predicted, which indicates that the value of the real class is “no,” and that the value of the predicted class is also “no.”
  • Both false positives and false negatives can arise in situations where the actual class does not match the class that was expected.
  • FP: This situation occurs when the actual class is not yes but yes for the predicted class.
  • FN: This situation occurs when the actual class is yes but the projected class is no.
  • With the help of this matrix, it is easier to understand why the categorization model generates inaccurate predictions. This gives you an idea of the mistakes that are being made, as well as the categories of errors.
The following are the evaluation metrics that we used:
A c c   % = T P + T N T P + T N + F P + F N
P r e c   % = T P T P + F P
R e c a l l   % = T P o s T P + F N
F 1 S c o r e   % = 2 × ( R e c a l l × P r e c R e c a l l + P r e c )

3.2. Confusion Matrix

Confusion matrix is implemented for displaying the outcome of the classifier model. A thorough experiment of the confusion matrix reveals that the fusion of 2-CNN and 3-CNN correctly recognizes the classes. The confusion matrix for all three models is presented in Figure 6, which shows that the proposed approach RNN-BiLSTM accurately detects the classes and performs better than the other two methods; i.e., RNN-LSTM and RNN-GRU.

3.3. ROC

Receiver-operating characteristic, often known as the sensitivity-to-specificity ratio, is an experimental method used to evaluate the efficiency of diagnostic tests. Figure 7 illustrates the ROC of the proposed 1-CNN, 2-CNN, and 3-CNN models for the binary class recognition problem of BC diagnosis. In comparison of different CNN models, the ROC results show that the 3-CNN model successfully classifies malignant and benign cancer.

3.4. Accuracy, Recall, Precision, and F1-Score

In this study, three models of CNN (1-CNN, 2-CNN, and 3-CNN) have been introduced to compare how well they classify cancer images. There are always two classes in a model, labeling 0 (benign) and 1 (malignant). The performance and effectiveness of a classifier can be evaluated based on its accuracy. It demonstrates the number of samples for which the model can adequately account for their characteristics. Figure 8 illustrates how accurate our proposed fusion CNN models are (i.e., 1-CNN, 2-CNN, and 3-CNN approaches). The accuracy, precision, recall, and f1-scores of the 1-CNN model were respectively 90.90%, 89.90%, 89.80%, and 90.20% while those of the 2-CNN model were 94.90%, 93.90%, 93.80%, and 93.20%, respectively. For the 3-CNN model, they were respectively 97.90%, 96.90%, 96.80% and 97.20%. The fusion 3-CNN model had the 3% best accuracy, precision, recall, and f1-scores than the 1-CNN and 2-CNN models, while the 1-CNN model had a poorer performance for BC detection and classification.

3.5. MCC Analysis

The Matthews correlation coefficient (MCC) represents the coefficient ratio between TP and TN. The value of MCC is equal to one when the classifier is perfect (FP = FN = 0), which indicates that there is a perfect positive correlation. On the other hand, when the classifier constantly makes an incorrect classification (TP = TN = 0), we obtain a value of −1, which represents a perfect negative. In this study, the MCC value of the 3-CNN model is close to 1, which showed a perfect positive correlation as compared to the 1-CNN and 2-CNN models, as shown in Figure 9.

3.6. Time Complexity

This section provides the time complexity of the various CNN model combinations, depending on the amount of time required for testing as well as training for classification of breast histology into benign and malignant tissues. The time complexity gives an indication of how effective the 3-CNN model is, which means that it requires a shorter amount of time for training and testing for BC detection, as shown in Figure 10.

4. Discussion

This study proposed different combinations of the CNN model (a simple convolutional neural network (CNN) model), as well as the fusion of two CNNs (also known as 2-CNN) and three CNNs (3-CNN) for BC detection. The features were extracted based on the CNN model. The proposed approach might be broken down into two primary categories: pre-processing (picture scaling and labelling), segmentation of the BC image, and the proposed CNN fusion model. In the data augmentation process, the number of images of malignant cancer was increased to improve the classification of the CNN model. We created and evaluated different fusion models that were trained and used to differentiate between benign and malignant classes.
The results of the experiments were derived from an analysis of our contribution to determine how the primary parameters affect the classification accuracy, specificity, and sensitivity; area under the receiver operating characteristic curve (ROC); and time complexity. In addition, the fusion of the CNN model is compared based on key performance metrics (i.e., accuracy (%), precision (%), recall (%), f1-score (%, ROC, confusion matrix), MCC, and computational time (ms)).The results, as shown in Figure 6, Figure 7 and Figure 8 reveal that when the fusion increased, the classification performance increased too; this means that the 3-CNN model performed better than the 2-CNN and 1-CNN models, in less computational time too.
Using the PCam database, we compared the accuracy of the introduced method to that of the state-of-the-art classification frameworks for BC histology. Because of this, we were able to gauge the significance of the hybrid-CNN method. Table 3 shows the differences between the proposed approach and the current SOTA methods. The AUC has been used as a performance metric for the comparison. Kandel and Castelli [56] have investigated how block-wise fine-tuning affected the performance of VGG16, VGG19, and InceptionV3, three popular Pretrained CNNs. The primary goal of their research is to determine if fine-tuning is an appropriate method for examining the PatchCamelyon histopathology dataset. In their study, they employed a batch size of 64 and three different learning rates of 10 3 , 10 4 , and 10 5 . For the VGG16, VGG19, and InceptionV3, the achieved AUC values were 0.94, 0.93, and 0.92, respectively. Lower values for the learning rate have been reported to yield the best results when training those pretrained CNNs. An ensemble model for the classification of the PCam histopathology dataset has been presented by Kassani et al. [57] that integrates three pre-trained CNNs (VGG19, MobileNet, and DenseNet). The feature representation and extraction processes have both made use of the ensemble model. The classification operation is executed by feeding the retrieved features into a multi-layer perceptron classifier. Normalization, data augmentation, and fine-tuning are among the various pre-processing and CNN tuning approaches used to train the model. Using PatchCamelyon histology datasets, the multi-model ensemble method achieves higher accuracy (94.64%) in its predictions compared to single classifiers and machine learning algorithms. The dense-attention network (DAN) is a method that has been presented for the classification of malignant patches by Liu et al. [58]. This method involves the attention mechanism being further enhanced in order to include prior knowledge about the surrounding tissue. In addition to this, the usefulness of data augmentation in the Inference stage has been verified even further. Using the PatchCamelyon dataset, where images with cancerous tissue in the middle are classified positive while those from the surrounding regions will not affect the label, the proposed approach is tested and evaluated. With the PCam dataset, they achieved an AUC of 0.976 successfully. An ensemble-based paradigm for the classification of lymph node metastases has been proposed by Rane et al. [59] DenseNet201, InceptionV3, and ResNeXt-50 are some of the pre-trained CNN models that are included in the aforementioned ensemble framework proposal. An attention fusion network has been implemented inside the framework that has been proposed in order to aggregate the predictions that have been generated using the different models. The introduced model obtained an AUC-ROC score of 0.9816 when used with the PCam benchmark dataset. Bonne [60] conducted a study in which she investigated various network architectures of CNN for the purpose of detecting breast cancer based on microscopic images of sentinel lymph node tissue. With the PatchCamelyon dataset of lymph node tissue, convolutional models of increasing depth are trained and assessed. It has been determined how much of an influence transfer learning, data augmentation, and fine-tuning of hyperparameters have. They found that increasing the depth of CNN had the greatest impact on performance, but data augmentation and transfer learning did not. With the help of the InceptionV3, they were able to get a value of 0.95 for the AUC performance parameter. Lafarge et al. [61] have proposed a new histopathological patch-classification model called P4M-DenseNet. This model improved upon the competitive standard CNN by mandating that rotation and reflection equivariance be maintained. The densely linked convolutional network serves as the foundation for the patch-classification model that was suggested. The performance of the P4M-DenseNet model was measured and analyzed using the PCam benchmarking dataset. The area under the curve (AUC) that was achieved for the classification of the microscopic images that were included in the PCam dataset was 96.3. By comparing the findings discussed earlier in this section to our proposed framework, we found the following: the performance level attained by the newly designed framework is higher than that achieved by Kandel and Castelli [56], Kassani et al. [57], Liu et al. [58], Bonnet [60], Rane et al. [59], and Lafarge [61].
While the proposed deep learning method showed impressive performance in distinguishing between benign and malignant breast tissue in histological images, it is handicapped in terms of the interpretability and explainability. Real-world applications, such as medical imaging, have gained great benefits from the use of deep neural networks. From a research perspective, however, black-box AI systems that lack or have limited explainability, and interpretability remain obstacles. Because of their non-linear and complex nature, deep neural networks lack interpretability compared to simpler and self-explaining models (such as linear regression). Black-box solutions in medical imaging and other fields are not easily deployed for mission-critical decision making, despite the deeper models’ ability to recognize and model complicated patterns and enable much improved performance. To address this issue, explainable artificial intelligence (XAI) techniques may offer explanations for the DL model black box, judgments that humans can understand. By assessing and reviewing individual predictions through analyzing their reasons, XAI approaches may assist physicians in further evaluating their models beyond typical performance metrics [62]. Furthermore, these techniques may expose biases in the training data, multiple labels, multi-input, and other inaccurate or fictitious correlations discovered using a model. Other forms of XAI include the technique known as Grad-CAM, which stands for Gradient-weighted Class Activation Mapping [63]. This method was first presented by Selvaraju et al. [64] in their study. In order to provide a post hoc local explanation, Grad-CAM can work with any sort of CNN, whereas CAM requires global average pooling in order to function properly. In addition to this, the authors presented guided Grad-CAM, which may be thought of as an element-wise multiplication between guided backpropagation and Grad-CAM. In the field of medical image analysis, both Grad-CAM and Guided Grad-CAM have been utilized. For example, Ji [65] utilized Grad-CAM to show which portions of the lymph node segments could be used as a classifier for the histology of metastatic tissue, and Kowsari et al. [66] used it to detect small bowel enteropathies on histopathology. Both groups of researchers found that Grad-CAM was an effective tool.
In order to improve the explainability of AI models, many applications of Scene graphs are being explored in the healthcare industry. They annotate the data, arrange it, and show it in a comprehensible way. Moreover, they add semantic labels to the healthcare datasets. An XAI model takes advantage of the structure of Scene graphs, which is built on subject–predicate–object relationships, in order to describe concepts and the connections between them [67]. It is possible to apply multi-modal machine translation methods in order to train a shared multilingual visual-semantic embedding space [68]. This may be useful for the purpose of improving the CNN models’ existing multi-modal capabilities. Furthermore, the transfer learning strategy for temporal activity detection [69] might be helpful for boosting the performance of CNN models for temporal BC histological images. This is because transfer learning is used to identify patterns of activity over time.
Another limitation of deep learning models in the diagnosis of BC histopathology is the requirement of extensive annotated datasets provided by physicians or other specialists in order to train the model. These datasets are needed to train the model. Zero-Shot Learning, also known as ZSL, is a relatively new method that has recently been developed to address this issue [70]. ZSL is presented for the purpose of having little human interaction by relying exclusively on previously known or trained concepts in addition to auxiliary information that is now available. Generative adversarial networks have made amazing strides in ZSL in recent years, which has led to some exciting new developments (GAN). An explosion of GAN designs has been constructed by human experts through trial-and-error testing in an effort to compensate for the dearth of training data available in ZSL [71].

5. Conclusions

In this study, we proposed a total of three models: the 1-CNN fusion of the two CNNs (2-CNN) and three CNNs (3-CNN) for BC pathological image classification. Initially, different pre-processing methods were applied in order to obtain the highest possible classification performance. Additionally, the experimental study was conducted by using the well-known Kaggle dataset (PCam). The results of the experiments indicated that the performance metrics of accuracy, recall, precision, and F1-score of the 1-CNN model were 90.10%, 89.90%, 89.80%, and 89.90%, respectively; the values for 2-CNN model were 94.7%, 93.9%, 93.4%, and 93.4, respectively, while those for the 3-CNN model were 97.90%, 97.5%, 97.60%, and 97.40%, respectively. The fusion 3-CNN model accurately classifies BC pathological images into benign and malignant cancerous lesions. Furthermore, the fusion 3-CNN model overcomes the issue of incorrect classification and time-complexity problem and has significantly better performance than the fusion 2-CNN and 1-CNN models, respectively.
In future studies, the proposed CNN fusion model will be expanded to breast tissue detection and classification in order to help medical professionals better identify diseases. The use of XAI approaches may make it possible to provide explanations for the introduced CNN model, sometimes known as the black box, and evaluations that humans can understand. Further objectives would be to test our approach using a more challenging breast cancer images dataset, which would enable us to demonstrate the model’s durability. We want to extend this method to other areas of healthcare, particularly those in which it has the potential to benefit the whole biomedical community.

Author Contributions

Conceptualization, A.R.; methodology, A.R.; software, A.R.; validation, A.R. and W.A.A.; formal analysis, A.C. and W.A.A.; investigation A.C.; resources, A.R., T.R.A. and G.A.; data curation, A.R. and G.A.; writing—original draft preparation, A.C., W.A.A., N.A.S., L.J.M. and G.A.; writing—review and editing, A.C. and A.R.; visualization, A.R., A.C., T.R.A. and G.A.; supervision A.R., N.A.S. and L.J.M.; project administration, G.A., N.A.S. and L.J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work is funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R387), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. This paper has been supported by the RUDN University Strategic Academic Leadership Program (recipient Alexander Chursin).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be provided upon request.

Acknowledgments

The authors would like to express their gratitude to Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R387), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The author declares that no conflict of interest to report regarding the present study.

Abbreviations

The following are the abbreviation used.
DLDeep learning
AIArtificial intelligence
MLMachine learning
GPUGraphics processing unit
RNNRecurrent neural networks
LSTMLong short-term memory
CNNConvolutional neural network

References

  1. Ou, F.S.; Michiels, S.; Shyr, Y.; Adjei, A.A.; Oberg, A.L. Biomarker Discovery and Validation: Statistical Considerations. J. Thorac. Oncol. 2021, 16, 537–545. [Google Scholar] [CrossRef]
  2. Amethiya, Y.; Pipariya, P.; Patel, S.; Shah, M. Comparative Analysis of Breast Cancer Detection Using Machine Learning and Biosensors. Intell. Med. 2022, 2, 69–81. [Google Scholar] [CrossRef]
  3. Solanki, Y.S.; Chakrabarti, P.; Jasinski, M.; Leonowicz, Z.; Bolshev, V.; Vinogradov, A.; Jasinska, E.; Gono, R.; Nami, M. A Hybrid Supervised Machine Learning Classifier System for Breast Cancer Prognosis Using Feature Selection and Data Imbalance Handling Approaches. Electronics 2021, 10, 699. [Google Scholar] [CrossRef]
  4. Chaurasia, V.; Pal, S.; Tiwari, B.B. Prediction of Benign and Malignant Breast Cancer Using Data Mining Techniques. J. Algorithm Comput. Technol. 2018, 12, 119–126. [Google Scholar] [CrossRef]
  5. Roslidar, R.; Rahman, A.; Muharar, R.; Syahputra, M.R.; Arnia, F.; Syukri, M.; Pradhan, B.; Munadi, K. A Review on Recent Progress in Thermal Imaging and Deep Learning Approaches for Breast Cancer Detection. IEEE Access. 2020, 8, 116176–116194. [Google Scholar] [CrossRef]
  6. Guo, R.; Lu, G.; Qin, B.; Fei, B. Ultrasound Imaging Technologies for Breast Cancer Detection and Management: A Review. Ultrasound Med. Biol. 2018, 44, 37–70. [Google Scholar] [CrossRef] [PubMed]
  7. Singh, S.P.; Urooj, S.; Lay-Ekuakille, A. Breast Cancer Detection Using PCPCET and ADEWNN: A Geometric Invariant Approach to Medical X-Ray Image Sensors. IEEE Sens. J. 2016, 16, 4847–4855. [Google Scholar] [CrossRef]
  8. Mohamed, W.A.; Kadah, Y.M. Computer Aided Diagnosis of Digital Mammograms. In Proceedings of the ICCES’07—2007 International Conference on Computer Engineering and Systems, Cairo, Egypt, 27–29 November 2007; pp. 299–303. [Google Scholar]
  9. Abduh, Z.; Wahed, M.A.; Kadah, Y.M. Robust Computer-Aided Detection of Pulmonary Nodules from Chest Computed Tomography. J. Med. Imaging Health Inf. 2016, 6, 693–699. [Google Scholar] [CrossRef]
  10. Alhussan, A.; Abdel Samee, N.; Ghoneim, V.; Kadah, Y. Evaluating Deep and Statistical Machine Learning Models in the Classification of Breast Cancer from Digital Mammograms. (IJACSA) Int. J. Adv. Comput. Sci. Appl. 2021, 12, 304–313. [Google Scholar] [CrossRef]
  11. Mohamed, W.A.; Alolfe, M.A.; Kadah, Y.M. Microcalcifications Enhancement in Digital Mammograms Using Fractal Modeling. In Proceedings of the 2008 Cairo International Biomedical Engineering Conference, CIBEC 2008, Cairo, Egypt, 18–20 December 2008. [Google Scholar]
  12. Alolfe, M.A.; Youssef, A.B.M.; Kadah, Y.M.; Mohamed, A.S. Development of a Computer-Aided Classification System for Cancer Detection from Digital Mammograms. In Proceedings of the National Radio Science Conference, NRSC, Tanta, Egypt, 18–20 March 2008. [Google Scholar]
  13. Alolfe, M.A.; Mohamed, W.A.; Youssef, A.B.M.; Mohamed, A.S.; Kadah, Y.M. Computer Aided Diagnosis in Digital Mammography Using Combined Support Vector Machine and Linear Discriminant Analyasis Classification. In Proceedings of the Proceedings—International Conference on Image Processing, Cairo, Egypt, 7–10 November 2009; IEEE Computer Society: Washington, DC, USA, 2009; pp. 2609–2612. [Google Scholar]
  14. Al-antari, M.A.; Al-masni, M.A.; Park, S.U.; Park, J.H.; Metwally, M.K.; Kadah, Y.M.; Han, S.M.; Kim, T.S. An Automatic Computer-Aided Diagnosis System for Breast Cancer in Digital Mammograms via Deep Belief Network. J. Med. Biol. Eng. 2018, 38, 443–456. [Google Scholar] [CrossRef]
  15. Basurto-Hurtado, J.A.; Cruz-Albarran, I.A.; Toledano-Ayala, M.; Ibarra-Manzano, M.A.; Morales-Hernandez, L.A.; Perez-Ramirez, C.A. Diagnostic Strategies for Breast Cancer Detection: From Image Generation to Classification Strategies Using Artificial Intelligence Algorithms. Cancers 2022, 14, 3442. [Google Scholar] [CrossRef] [PubMed]
  16. Brooks, A.D.; Nover, A.B.; Jagtap, S.; Anjum, W.; Yegingil, H.; Shih, W.Y.; Shih, W.H. Modern Breast Cancer Detection: A Technological Review. Int. J. Biomed. Imaging 2009, 2009, 1–14. [Google Scholar] [CrossRef]
  17. Abdel Samee, N.M. Classical and Deep Learning Paradigms for Detection and Validation of Key Genes of Risky Outcomes of HCV. Algorithms 2020, 13, 73. [Google Scholar] [CrossRef]
  18. Samee, N.A.; Atteia, G.; Alkanhel, R.; Alhussan, A.A.; AlEisa, H.N. Hybrid Feature Reduction Using PCC-Stacked Autoencoders for Gold/Oil Prices Forecasting under COVID-19 Pandemic. Electronics 2022, 11, 991. [Google Scholar] [CrossRef]
  19. Abdel Samee, N.M.; Solouma, N.H.; Kadah, Y.M. Detection of Biomarkers for Hepatocellular Carcinoma Using a Hybrid Univariate Gene Selection Methods. Theor. Biol. Med. Model. 2012, 9, 34. [Google Scholar] [CrossRef]
  20. Nawaz, A.; Abbas, Y.; Ahmad, T.; Mahmoud, N.F.; Rizwan, A.; Samee, N.A. A Healthcare Paradigm for Deriving Knowledge Using Online Consumers’ Feedback. Healthcare 2022, 10, 1592. [Google Scholar] [CrossRef]
  21. Xu, J.; Li, C.; Zhou, Y.; Mou, L.; Zheng, H.; Wang, S. Classifying Mammographic Breast Density by Residual Learning. arXiv 2018, arXiv:1809.10241. [Google Scholar]
  22. Ismail, N.; Sovuthy, C. Breast Cancer Detection Based on Deep. Learning Technique. In Proceedings of the 2019 International UNIMAS STEM 12th Engineering Conference (EnCon), Kuching, Malaysia, 28–29 August 2019; pp. 89–92. [Google Scholar]
  23. Kourou, K.; Exarchos, T.P.; Exarchos, K.P.; Karamouzis, M.V.; Fotiadis, D.I. Machine Learning Applications in Cancer Prognosis and Prediction. Comput. Struct. Biotechnol. J. 2015, 13, 8–17. [Google Scholar] [CrossRef] [PubMed]
  24. Kadah, Y.M.; Farag, A.A.; Zurada, J.M.; Badawi, A.M.; Youssef, A.B.M. Classification Algorithms for Quantitative Tissue Characterization of Diffuse Liver Disease from Ultrasound Images. IEEE Trans. Med. Imaging 1996, 15, 466–478. [Google Scholar] [CrossRef]
  25. Zhang, H.; Han, L.; Chen, K.; Peng, Y.; Lin, J. Diagnostic Efficiency of the Breast Ultrasound Computer-Aided Prediction Model Based on Convolutional Neural Network in Breast Cancer. J. Digit. Imaging 2020, 33, 1218–1223. [Google Scholar] [CrossRef]
  26. Fathy, W.E.; Ghoneim, A.S. A Deep Learning Approach for Breast Cancer Mass Detection. Int. J. Adv. Comput. Sci. Appl. 2019, 10, 175–182. [Google Scholar] [CrossRef]
  27. Lévy, D.; Jain, A. Breast Mass Classification from Mammograms Using Deep Convolutional Neural Networks. arXiv 2016, arXiv:1612.00542. [Google Scholar]
  28. Tan, Y.J.; Sim, K.S.; Ting, F.F. Breast Cancer Detection Using Convolutional Neural Networks for Mammogram Imaging System. In Proceedings of the 2017 International Conference on Robotics, Automation and Sciences, ICORAS 2017, Virtual, 20–22 November 2020; Institute of Electrical and Electronics Engineers Inc.: New York, NY, USA, 2018; pp. 1–5. [Google Scholar]
  29. Samee, N.A.; Atteia, G.; Meshoul, S.; Al-Antari, M.A.; Kadah, Y.M. Deep Learning Cascaded Feature Selection Framework for Breast Cancer Classification: Hybrid CNN with Univariate-Based Approach. Mathematics 2022, 10, 3631. [Google Scholar] [CrossRef]
  30. Samee, N.A.; Alhussan, A.A.; Ghoneim, V.F.; Atteia, G.; Alkanhel, R.; Al-antari, M.A.; Kadah, Y.M. A Hybrid Deep Transfer Learning of CNN-Based LR-PCA for Breast Lesion Diagnosis via Medical Breast Mammograms. Sensors 2022, 22, 4938. [Google Scholar] [CrossRef]
  31. Falconi, L.G.; Perez, M.; Aguilar, W.G. Transfer Learning in Breast Mammogram Abnormalities Classification with Mobilenet and Nasnet. In Proceedings of the International Conference on Systems, Signals, and Image Processing, Osijek, Croatia, 5–7 June 2019; IEEE Computer Society: Washington, DC, USA, 2019; pp. 109–114. [Google Scholar]
  32. Khamparia, A.; Bharati, S.; Podder, P.; Gupta, D.; Khanna, A.; Phung, T.K.; Thanh, D.N.H. Diagnosis of Breast Cancer Based on Modern Mammography Using Hybrid Transfer Learning. Multidimens. Syst. Signal. Process. 2021, 32, 747–765. [Google Scholar] [CrossRef] [PubMed]
  33. Hassan, S.A.; Sayed, M.S.; Abdalla, M.I.; Rashwan, M.A. Breast Cancer Masses Classification Using Deep Convolutional Neural Networks and Transfer Learning. Multimed. Tools Appl. 2020, 79, 30735–30768. [Google Scholar] [CrossRef]
  34. Wang, Z.; Li, M.; Wang, H.; Jiang, H.; Yao, Y.; Zhang, H.; Xin, J. Breast Cancer Detection Using Extreme Learning Machine Based on Feature Fusion with CNN Deep Features. IEEE Access. 2019, 7, 105146–105158. [Google Scholar] [CrossRef]
  35. Patil, R.S.; Biradar, N. Automated Mammogram Breast Cancer Detection Using the Optimized Combination of Convolutional and Recurrent Neural Network. Evol. Intell. 2020, 2020, 1459–1474. [Google Scholar] [CrossRef]
  36. Hasan, M.K.; Aleef, T.A.; Roy, S. Automatic Mass Classification in Breast Using Transfer Learning of Deep Convolutional Neural Network and Support Vector Machine. In Proceedings of the 2020 IEEE Region 10 Symposium, TENSYMP 2020, Dhaka, Bangladesh, 5–7 June 2020; Institute of Electrical and Electronics Engineers Inc.: Interlaken, Switzerland, 2020; pp. 110–113. [Google Scholar]
  37. Al-masni, M.A.; Al-antari, M.A.; Park, J.M.; Gi, G.; Kim, T.Y.; Rivera, P.; Valarezo, E.; Choi, M.T.; Han, S.M.; Kim, T.S. Simultaneous Detection and Classification of Breast Masses in Digital Mammograms via a Deep Learning YOLO-Based CAD System. Comput. Methods Programs Biomed. 2018, 157, 85–94. [Google Scholar] [CrossRef]
  38. Alzubaidi, L.; Al-Shamma, O.; Fadhel, M.A.; Farhan, L.; Zhang, J.; Duan, Y. Optimizing the Performance of Breast Cancer Classification by Employing the Same Domain Transfer Learning from Hybrid Deep Convolutional Neural Network Model. Electronics 2020, 9, 445. [Google Scholar] [CrossRef]
  39. Wu, J.; Hicks, C. Breast Cancer Type Classification Using Machine Learning. J. Pers. Med. 2021, 11, 61. [Google Scholar] [CrossRef]
  40. Tang, J.; Rangayyan, R.M.; Xu, J.; El Naqa, I.; Yang, Y. Computer-Aided Detection and Diagnosis of Breast Cancer with Mammography: Recent Advances. IEEE Trans. Inf. Technol. Biomed. 2009, 13, 236–251. [Google Scholar] [CrossRef] [PubMed]
  41. Belsare, A.D.; Mushrif, M.M.; Pangarkar, M.A.; Meshram, N. Classification of Breast Cancer Histopathology Images Using Texture Feature Analysis. In Proceedings of the TENCON 2015–2015 IEEE Region 10 Conference 2015, Macao, 1–4 November 2015; pp. 1–5. [Google Scholar] [CrossRef]
  42. Zhang, Y.; Zhang, B.; Coenen, F.; Lu, W. Breast Cancer Diagnosis from Biopsy Images with Highly Reliable Random Subspace Classifier Ensembles. Mach. Vision. Appl. 2012, 24, 1405–1420. [Google Scholar] [CrossRef]
  43. Kowal, M.; Filipczuk, P.; Obuchowicz, A.; Korbicz, J.; Monczak, R. Computer-Aided Diagnosis of Breast Cancer Based on Fine Needle Biopsy Microscopic Images. Comput. Biol. Med. 2013, 43, 1563–1572. [Google Scholar] [CrossRef] [PubMed]
  44. Doyle, S.; Agner, S.; Madabhushi, A.; Feldman, M.; Tomaszewski, J. Automated Grading of Breast Cancer Histopathology Using. Spectral Clustering with Textural and Architectural Image Features. In Proceedings of the 2008 5th IEEE International Symposium on Biomedical Imaging: From Nano to Macro, Paris, France, 14–17 May 2008; pp. 496–499. [Google Scholar] [CrossRef]
  45. Bilgin, C.; Demir, C.; Nagi, C.; Yener, B. Cell-Graph Mining for Breast Tissue Modeling and Classification. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. 2007, 2007, 5311–5314. [Google Scholar] [CrossRef] [PubMed]
  46. Cruz-Roa, A.; Basavanhally, A.; González, F.; Gilmore, H.; Feldman, M.; Ganesan, S.; Shih, N.; Tomaszewski, J.; Madabhushi, A. Automatic Detection of Invasive Ductal Carcinoma in Whole Slide Images with Convolutional Neural Networks. SPIE 2014, 9041, 904103. [Google Scholar] [CrossRef]
  47. Zhang, B. Breast Cancer Diagnosis from Biopsy Images by Serial Fusion of Random Subspace Ensembles. In Proceedings of the 2011 4th International Conference on Biomedical Engineering and Informatics, BMEI 2011, Shanghai, China, 15–17 October 2011; Volume 1, pp. 180–186. [Google Scholar] [CrossRef]
  48. Bayramoglu, N.; Kannala, J.; Heikkila, J. Deep Learning for Magnification Independent Breast Cancer Histopathology Image Classification. Proc. Int. Conf. Pattern Recognit. 2016, 2440–2445. [Google Scholar] [CrossRef]
  49. Araujo, T.; Aresta, G.; Castro, E.; Rouco, J.; Aguiar, P.; Eloy, C.; Polonia, A.; Campilho, A. Classification of Breast Cancer Histology Images Using Convolutional Neural Networks. PLoS ONE 2017, 12, e0177544. [Google Scholar] [CrossRef]
  50. Aresta, G.; Araújo, T.; Kwok, S.; Chennamsetty, S.S.; Safwan, M.; Alex, V.; Marami, B.; Prastawa, M.; Chan, M.; Donovan, M.; et al. BACH: Grand Challenge on Breast Cancer Histology Images. Med. Image Anal. 2019, 56, 122–139. [Google Scholar] [CrossRef]
  51. Ahmad, H.M.; Ghuffar, S.; Khurshid, K. Classification of Breast Cancer Histology Images Using Transfer Learning. In Proceedings of the 2019 16th International Bhurban Conference on Applied Sciences and Technology, IBCAST 2019, Islamabad, Pakistan, 8–12 January 2019; pp. 328–332. [Google Scholar] [CrossRef]
  52. Vang, Y.S.; Chen, Z.; Xie, X. Deep Learning Framework for Multi-Class Breast Cancer Histology Image Classification. In Image Analysis and Recognition: 15th International Conference, ICIAR 2018, Póvoa de Varzim, Portugal, 27–29 June 2018; Springer International Publishing: Zurich, Switwerland, 2018; pp. 914–922. [Google Scholar] [CrossRef]
  53. Rakhlin, A.; Shvets, A.; Iglovikov, V.; Kalinin, A.A. Deep Convolutional Neural Networks for Breast Cancer Histology Image Analysis. In Image Analysis and Recognition: 15th International Conference, ICIAR 2018, Póvoa de Varzim, Portugal, 27–29 June 2018; Springer International Publishing: Zurich, Switwerland, 2018; pp. 737–744. [Google Scholar] [CrossRef]
  54. Awan, R.; Koohbanani, N.A.; Shaban, M.; Lisowska, A.; Rajpoot, N. Context-Aware Learning Using Transferable Features for Classification of Breast Cancer Histology Images. In Image Analysis and Recognition: 15th International Conference, ICIAR 2018, Póvoa de Varzim, Portugal, 27–29 June 2018; Springer International Publishing: Zurich, Switwerland, 2018; pp. 788–795. [Google Scholar] [CrossRef]
  55. Histopathologic Cancer Detection|Kaggle. Available online: https://www.kaggle.com/c/histopathologic-cancer-detection (accessed on 17 October 2022).
  56. Kandel, I.; Castelli, M. How Deeply to Fine-Tune a Convolutional Neural Network: A Case Study Using a Histopathology Dataset. Appl. Sci. 2020, 10, 3359. [Google Scholar] [CrossRef]
  57. Kassani, S.H.; Kassani, P.H.; Wesolowski, M.J.; Schneider, K.A.; Deters, R. Classification of Histopathological Biopsy Images Using Ensemble of Deep Learning Networks. In Proceedings of the Conference of the Centre for Advanced Studies on Collaborative Research—Proceedings of the 29th Annual International Conference on Computer Science and Software Engineering 2019, Markham, ON, Canada, 4–6 November 2019; pp. 92–99. [Google Scholar]
  58. Liu, M.; Yu, Y.; Liao, Q.; Zhang, J. Histopathologic Cancer Detection by Dense-Attention Network with Incorporation of Prior Knowledge. In Proceedings of the International Symposium on Biomedical Imaging 2020, Iowa City, IA, USA, 3–7 April 2020; pp. 466–470. [Google Scholar] [CrossRef]
  59. Rane, C.; Mehrotra, R.; Bhattacharyya, S.; Sharma, M.; Bhattacharya, M. A Novel Attention Fusion Network-Based Framework to Ensemble the Predictions of CNNs for Lymph Node Metastasis Detection. J. Supercomput. 2021, 77, 4201–4220. [Google Scholar] [CrossRef]
  60. Bonnet, E. Using Convolutional Neural Networks for the Classification of Breast Cancer Images. arXiv 2021, arXiv:2108.13661. [Google Scholar]
  61. Lafarge, M.W.; Bekkers, E.J.; Pluim, J.P.W.; Duits, R.; Veta, M. Roto-Translation Equivariant Convolutional Networks: Application to Histopathology Image Analysis. Med. Image Anal. 2021, 68, 101849. [Google Scholar] [CrossRef]
  62. Kakogeorgiou, I.; Karantzalos, K. Evaluating Explainable Artificial Intelligence Methods for Multi-Label Deep Learning Classification Tasks in Remote Sensing. Int. J. Appl. Earth Obs. Geoinf. 2021, 103, 102520. [Google Scholar] [CrossRef]
  63. Van Der Velden, B.H.M.; Kuijf, H.J.; Gilhuijs, K.G.A.; Viergever, M.A.; Van Der Velden, B. Explainable Artificial Intelligence (XAI) in Deep Learning-Based Medical Image Analysis. Med. Image Anal. 2022, 4, 102470. [Google Scholar] [CrossRef]
  64. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar] [CrossRef]
  65. Ji, J. Gradient-Based Interpretation on Convolutional Neural Network for Classification of Pathological Images. In Proceedings of the 2019 International Conference on Information Technology and Computer Application, ITCA 2019, Guangzhou, China, 20–22 December 2019; pp. 83–86. [Google Scholar] [CrossRef]
  66. Kowsari, K.; Sali, R.; Ehsan, L.; Adorno, W.; Ali, A.; Moore, S.; Amadi, B.; Kelly, P.; Syed, S.; Brown, D. HMIC: Hierarchical Medical Image Classification, A Deep Learning Approach. Information 2020, 11, 318. [Google Scholar] [CrossRef]
  67. Chang, X.; Ren, P.; Xu, P.; Li, Z.; Chen, X.; Hauptmann, A. A Comprehensive Survey of Scene Graphs: Generation and Application. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 1–26. [Google Scholar] [CrossRef]
  68. Li, M.; Huang, P.Y.; Chang, X.; Hu, J.; Yang, Y.; Hauptmann, A. Video Pivoting Unsupervised Multi-Modal Machine Translation. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 3918–3932. [Google Scholar] [CrossRef]
  69. Zhang, L.; Chang, X.; Liu, J.; Luo, M.; Li, Z.; Yao, L.; Hauptmann, A. TN-ZSTAD: Transferable Network for Zero-Shot Temporal Activity Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 3848–3861. [Google Scholar] [CrossRef]
  70. Rezaei, M.; Shahidi, M. Zero-Shot Learning and Its Applications from Autonomous Vehicles to COVID-19 Diagnosis: A Review. Intell. Based Med. 2020, 3–4, 100005. [Google Scholar] [CrossRef]
  71. Yan, C.; Chang, X.; Li, Z.; Guan, W.; Ge, Z.; Zhu, L.; Zheng, Q. ZeroNAS: Differentiable Generative Adversarial Networks Search for Zero-Shot Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 9733–9740. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Examples of benign and malignant tissues. (a) Examples of Malignant tissues; (b) Examples of benign tissues.
Figure 1. Examples of benign and malignant tissues. (a) Examples of Malignant tissues; (b) Examples of benign tissues.
Diagnostics 13 01700 g001
Figure 2. Examples of augmented images via rotation and resizing.
Figure 2. Examples of augmented images via rotation and resizing.
Diagnostics 13 01700 g002
Figure 3. Block diagram of the 1-CNN model for the classification of breast histological images into benign/malignant tissues.
Figure 3. Block diagram of the 1-CNN model for the classification of breast histological images into benign/malignant tissues.
Diagnostics 13 01700 g003
Figure 4. Block diagram of the 2-CNN model for the classification of breast histological images into benign/malignant tissues.
Figure 4. Block diagram of the 2-CNN model for the classification of breast histological images into benign/malignant tissues.
Diagnostics 13 01700 g004
Figure 5. The proposed 3-CNN model block diagram for the classification of breast histological images into benign/malignant tissues.
Figure 5. The proposed 3-CNN model block diagram for the classification of breast histological images into benign/malignant tissues.
Diagnostics 13 01700 g005
Figure 6. The retrieved confusion matrices of the introduced models (1-CNN, 2-CNN, and 3-CNN) for the diagnosis of BC histology.
Figure 6. The retrieved confusion matrices of the introduced models (1-CNN, 2-CNN, and 3-CNN) for the diagnosis of BC histology.
Diagnostics 13 01700 g006
Figure 7. The retrieved AUC-ROC curves of the introduced models (1-CNN, 2-CNN, and 3-CNN) model for the diagnosis of BC detection and histology.
Figure 7. The retrieved AUC-ROC curves of the introduced models (1-CNN, 2-CNN, and 3-CNN) model for the diagnosis of BC detection and histology.
Diagnostics 13 01700 g007
Figure 8. The retrieved accuracies, recalls, precisions, and f1-scores of the introduced models (1-CNN, 2-CNN, and 3-CNN) for the diagnosis of BC histology.
Figure 8. The retrieved accuracies, recalls, precisions, and f1-scores of the introduced models (1-CNN, 2-CNN, and 3-CNN) for the diagnosis of BC histology.
Diagnostics 13 01700 g008
Figure 9. The retrieved MCC values of the introduced models (1-CNN, 2-CNN, and 3-CNN) for the diagnosis of BC histology.
Figure 9. The retrieved MCC values of the introduced models (1-CNN, 2-CNN, and 3-CNN) for the diagnosis of BC histology.
Diagnostics 13 01700 g009
Figure 10. Time complexity (milliseconds) of the introduced diagnosing models (1-CNN, 2-CNN, and 3-CNN) for BC histology.
Figure 10. Time complexity (milliseconds) of the introduced diagnosing models (1-CNN, 2-CNN, and 3-CNN) for BC histology.
Diagnostics 13 01700 g010
Table 1. An overview of the State of The Art, SOTA, methods and their limitations for the automatic diagnosis of the abnormalities in breast lesions’ histopathology.
Table 1. An overview of the State of The Art, SOTA, methods and their limitations for the automatic diagnosis of the abnormalities in breast lesions’ histopathology.
AuthorApproachesDatasetStrong PointsWeak Points
Bayramoglu et al. [48]CNNsBreaKHis datasetMalignancy prediction using a single-task CNN and simultaneous prediction of malignancy and image magnification level using a multi-task CNN are the two architectures proposed.They were unable to achieve better performance than the state-of-the-art results.
Arajo et al. [49]
  • CNN
  • CNN + SVM
Breast histology classification challenge H&E stain samples from Bioimaging 2015.They have completed image-wise classification into four medically relevant classifications, as opposed to the two classes used via earlier methods: (i) normal tissue, (ii) benign lesion, (iii) in situ carcinoma, and (iv) invasive carcinoma.Patch classification performance may be decreased if some training/testing patches lack significant information.
Aresta et al. [50]CNN-based algorithmsBACH dataset from the ICIAR2018 competitionEffective methods for the Automatic classification systems using hematoxylin–eosin-stained histopathology imagesThe study findings are not yet satisfactory for clinical evaluation.
Ahmed et al. [51]Transfer learning (AlexNet/ResNet/GoogLeNet)Breast Histology datasetThe implemented strategy has brought about an accuracy rate of 85% in the instance of ResNet, which is the highest among others.The design that was presented can be improved to allow better diagnosing accuracy.
Vang et al. [52] Google Inception V3 NetworkBACH dataset from the ICIAR2018 competitionThe introduced framework has resulted in a 12.5% enhancement over the SOTA modelsThe findings of the study do not yet meet the criteria necessary for clinical assessment.
Rakhlin et al. [53].VGG-16+ Gradient boosted trees classifierBACH dataset from the ICIAR2018 competitionThey reported an accuracy rate of 87.2% when doing the 4-class categorization challenge. They reported an accuracy of 93.8% when asked to do a 2-class classification test to identify carcinomas.Lack of comparison with other pretrained CNNs
Awan et al. [54]ResNetBACH dataset from the ICIAR2018 competitionA novel framework has been presented. ResNet is introduced for feature extraction, PCA for feature reduction, SVM for patch classification, and a majority voting classifier for overall input image classification.When compared to the accuracy retrieved using SOTA models, the achieved accuracy of 90% is lower.
Proposed methodFusion of CNNs (1-CNN, 2-CNN, 3-CNN) PCam datasetThe performance of the suggested model, which utilizes hybrid CNNs, is shown to be significantly enhanced when it comes to correctly and precisely identifying abnormalities in breast histology, which suggests that its application in this field should be expanded.
Table 2. Experimental setup.
Table 2. Experimental setup.
GPU1060 6 GB, Nvidia GeForce
RAM16 GB
LanguagesPython 3.7
CPUCore-i6, 6th Generation, and 2.80 GHz processor
Operating systemWindows 64 bit
Table 3. Comparing the attained classification performance to the state-of-the-art for the diagnosis of BC histology using the PCam dataset.
Table 3. Comparing the attained classification performance to the state-of-the-art for the diagnosis of BC histology using the PCam dataset.
AuthorApproachAUCAccuracy
Kandel and Castelli [56]Transfer learning0.94 (VGG16)
0.93 (VGG19)
0.92 (InceptionV3)
-
Kassani et al. [57]Ensemble model (Pre-trained CNNs + Multi-layer perceptron classifier)-94.64%
Liu et al. [58]Dense-Attention Network0.976-
Bonnet [60]CNN, InceptionV30.95-
Rane et al. [59]Ensemble framework (DenseNet201, InceptionV3 and ResNeXt-50)0.9816-
Lafarge [61]P4M-DenseNet96.389.8
Proposed model1-CNN, 2-CNN, and 3-CNN based fusion model0.89, 0.94, and 0.9790%, 94%, and 98%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rafiq, A.; Chursin, A.; Awad Alrefaei, W.; Rashed Alsenani, T.; Aldehim, G.; Abdel Samee, N.; Menzli, L.J. Detection and Classification of Histopathological Breast Images Using a Fusion of CNN Frameworks. Diagnostics 2023, 13, 1700. https://doi.org/10.3390/diagnostics13101700

AMA Style

Rafiq A, Chursin A, Awad Alrefaei W, Rashed Alsenani T, Aldehim G, Abdel Samee N, Menzli LJ. Detection and Classification of Histopathological Breast Images Using a Fusion of CNN Frameworks. Diagnostics. 2023; 13(10):1700. https://doi.org/10.3390/diagnostics13101700

Chicago/Turabian Style

Rafiq, Ahsan, Alexander Chursin, Wejdan Awad Alrefaei, Tahani Rashed Alsenani, Ghadah Aldehim, Nagwan Abdel Samee, and Leila Jamel Menzli. 2023. "Detection and Classification of Histopathological Breast Images Using a Fusion of CNN Frameworks" Diagnostics 13, no. 10: 1700. https://doi.org/10.3390/diagnostics13101700

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop