Next Article in Journal
Can Gastric Juice Analysis with EndoFaster® Reduce the Environmental Impact of Upper Endoscopy?
Next Article in Special Issue
Disparities in Breast Cancer Care—How Factors Related to Prevention, Diagnosis, and Treatment Drive Inequity
Previous Article in Journal
The Role of Resilience in the Relationship between Sociodemographic, Clinical Characteristics, and Social Support among Breast Cancer Patients in Serbia
Previous Article in Special Issue
A Clinical Prediction Model for Breast Cancer in Women Having Their First Mammogram
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Federated Learning Approach to Breast Cancer Prediction in a Collaborative Learning Framework

1
Department of Information Systems, College of Computer and Information Sciences, Jouf University, Al Jouf 72311, Saudi Arabia
2
Department of Avionics Engineering, Air University, Islamabad 44000, Pakistan
3
Institute of Information Technology, Quaid-i-Azam University, Islamabad 45320, Pakistan
*
Author to whom correspondence should be addressed.
Healthcare 2023, 11(24), 3185; https://doi.org/10.3390/healthcare11243185
Submission received: 5 October 2023 / Revised: 13 December 2023 / Accepted: 14 December 2023 / Published: 17 December 2023
(This article belongs to the Special Issue Prevention, Diagnosis, and Treatment of Breast Cancer)

Abstract

:
Breast cancer continues to pose a substantial worldwide public health concern, necessitating the use of sophisticated diagnostic methods to enable timely identification and management. The present research utilizes an iterative methodology for collaborative learning, using Deep Neural Networks (DNN) to construct a breast cancer detection model with a high level of accuracy. By leveraging Federated Learning (FL), this collaborative framework effectively utilizes the combined knowledge and data assets of several healthcare organizations while ensuring the protection of patient privacy and data security. The model described in this study showcases significant progress in the field of breast cancer diagnoses, with a maximum accuracy rate of 97.54%, precision of 96.5%, and recall of 98.0%, by using an optimum feature selection technique. Data augmentation approaches play a crucial role in decreasing loss and improving model performance. Significantly, the F1-Score, a comprehensive metric for evaluating performance, turns out to be 97%. This study signifies a notable advancement in the field of breast cancer screening, fostering hope for improved patient outcomes via increased accuracy and reliability. This study highlights the potential impact of collaborative learning, namely, in the field of FL, in transforming breast cancer detection. The incorporation of privacy considerations and the use of diverse data sources contribute to the advancement of early detection and the treatment of breast cancer, hence yielding significant benefits for patients on a global scale.

1. Introduction

Breast cancer has emerged as a dominant global health concern, surpassing even lung cancer in prevalence [1] and ranking second in terms of mortality, making it the second leading cause of death among women worldwide [2]. The World Health Organization (WHO) recognizes breast cancer as the most commonly diagnosed cancer globally, accounting for 12% of all new cancer cases [3,4]. This formidable health challenge primarily affects women between the ages of 35 and 45, characterized by complex biochemical nuances. Detecting invasive breast cancer at its early stages offers the prospect of cure, particularly for women over 40, yet early detection remains a significant challenge, especially when relying on mammography as a diagnostic tool [5,6]. Despite its ubiquity and cost-effectiveness, mammography often falls short of timely diagnoses. Successful breast cancer diagnosis with mammography frequently occurs at later stages, contributing significantly to high mortality rates [7]. Advances in pharmaceutical research have improved treatment outcomes and reduced adverse effects, turning breast cancer into a malignancy with a more favorable prognosis. Consequently, the approach to breast cancer has evolved into a chronic condition, giving rise to a comprehensive strategy encompassing screening, diagnosis, therapy, and post-diagnosis follow-up [8,9]. Therefore, the early detection of breast cancer plays a pivotal role in this comprehensive strategy, potentially reducing mortality rates and increasing the likelihood of successful treatment. Various imaging modalities, including mammography, ultrasonography, Magnetic Resonance Imaging (MRI), and Positron Emission Tomography (PET), have been integral to breast cancer diagnosis [10]. Figure 1 shows the average ratio of increasing breast cancer cases, conventional diagnosis ratio, computer-aided diagnosis ratio, and AI-based diagnosis for the years 2018 to date. The data are acquired from Google Trends https://trends.google.com/trends/ (accessed on 24 September 2023).
Nevertheless, the substantial volume of imaging data generated during this diagnostic process poses a significant burden on radiologists, and diagnostic precision can be hindered by suboptimal or ambiguous images [11]. In addition, computer-aided diagnosis (CAD) is valuable for swiftly detecting and assessing breast cancer cases. This comprehensive approach encompasses three key stages: the initial identification and localization of potential tumors or cancer cells from preprocessed mammography, the extraction of tumor attributes (morphology, dimensions, radiodensity, prospective mass, and consistency), and final categorization into benign or malignant [12]. Traditional X-ray and mammography have given way to more advanced detection systems, such as data mining, neural networks, and Artificial Intelligence (AI) [13]. It is the field of study that focuses on creating intelligent machines that can perform tasks that require human intelligence [14]. AI systems can recognize patterns and adapt to new situations by analyzing data. They can also solve problems, make decisions, perceive their surroundings, and interact with humans and other machines naturally [15]. A creative and powerful application of AI to healthcare is presented in [16]. Timely and accurate breast cancer detection minimizes unnecessary treatments and improves patient survival rates. Machine Learning (ML) techniques have shown promise in this context; however, there is often a gap between the medical knowledge of professionals and their understanding of ML methods and results [17,18]. ML offers resilience in addressing diagnostic limitations stemming from the subjectivity of human practitioners.
Moreover, ML may represent the most economically viable breast cancer screening alternative in resource-constrained developing nations with large populations and limited healthcare resources. Recent advances in Self Supervised Learning (SSL) show promise in performance by leveraging extensive unlabeled datasets, reducing the need for annotations. However, obtaining substantial and diverse medical data remains critical for effectively training robust medical ML algorithms [19,20]. Besides having huge and diverse datasets, processing capacity has emerged as a prominent limitation in training ML algorithms. The discipline of distributed systems has emerged due to the increasing demand for highly efficient computer resources, including processing power, memory, and storage space [21]. Within this specific field, distributed ML has surfaced as a framework in which algorithms are implemented and deployed across multiple nodes, leveraging a larger number of data and processing capabilities and improving performance and efficiency. The core tenet of distributed learning entails the dispersion of algorithms among computer nodes, as expounded in [21]. Nevertheless, it is crucial to acknowledge that these operations are carried out without considering any potential limitations that might have to be enforced by these nodes. For example, the data employed across those nodes could come from disparate distributions. Consequently, many practical applications in collaborative learning must adhere more to the fundamental principle of Independent and Identically Distributed (IID) data among nodes. This encompasses situations when client information from mobile devices or healthcare data from various geographic and demographic parameters are implicated.
Federated Learning (FL) is acknowledged to be a collaborative learning paradigm that adeptly addresses a range of practical difficulties and sets itself apart from traditional distributed learning situations. It is delivering efficient enhancements in the healthcare domain, showcasing significant improvements [22]. It is crucial to consider several factors to ensure the accuracy and reliability of data analysis in a distributed network. Firstly, the statistical heterogeneity of data across nodes needs to be considered. It means acknowledging and accommodating variations in data characteristics and distributions among nodes. Secondly, addressing data imbalance across nodes is essential. It involves mitigating the situation where certain nodes possess significantly more or less data than others. Strategies should be implemented to balance the data distribution across nodes to avoid biased results. Thirdly, the limited communication capabilities within the distributed network must be managed effectively. Challenges such as loss of synchronization and variability in communication capabilities must be addressed to ensure seamless and accurate data transmission and processing. Lastly, handling the potential scenario of many nodes relative to the available data is important. It requires careful consideration of the distributed system’s scalability, privacy, and efficiency, ensuring that the analysis can handle the increased complexity and volume of data generated by numerous nodes [23].
The following are the contributions of this paper:
  • An FL-based framework is presented for the categorization and early detection of breast cancer using DNN to train, classify, and detect breast cancer collaboratively.
  • Careful data augmentation and class balancing is done using the Synthetic Minority Over-sampling Technique (SMOTE) on a substantial and diverse dataset of patients with breast cancer for testing and validation.
  • Benchmark results are presented using a comprehensive set of evaluation measures to demonstrate the performance and efficiency of the proposed model.
The ensuing sections of the article are structured as follows: the literature review is presented in Section 2. The methodology and its mathematical modeling, along with the dataset, are explained in Section 3. Section 4 discusses the experimental results, and, finally, the conclusion is provided in Section 6.

2. Related Work

The current body of literature encompasses a substantial range of studies about breast cancer that hold important relevance. Sakib et al. [24] did a comparative analysis to investigate the application of ML and Deep Learning (DL) approaches in the context of breast cancer detection and diagnosis. The task of classification involved the utilization of five well-established supervised ML methodologies, namely, the Support Vector Machine (SVM), K-Nearest Neighbour (KNN), Decision Tree (DT), Logistic Regression (LR), and Random Forest (RF), in conjunction with a DL methodology. The solutions were assessed by implementing tenfold cross-validation on a dataset owned by the organization, which comprised 699 samples. The investigation conducted by the researchers indicates that the artificial artificial neural network (ANN) exhibits the highest degree of accuracy, achieving a noteworthy 98.57%. In contrast, it is noteworthy that both the RF and LR models demonstrated a commendable accuracy rate of 95.7%. Singh et al. [7] conducted a comparison investigation to evaluate different machine-learning approaches in the context of breast cancer diagnosis. Furthermore, the research paper presents a novel auto-encoder model that performs unsupervised breast cancer identification. The aim is to determine a succinct portrayal of traits that have a notable association with breast cancer. The utilized approaches were implemented using the openly available Breast Cancer Wisconsin (Diagnostic) Dataset, accessed through the Kaggle platform. The auto-encoder performs better than its counterparts, with a precision and recall rate of 98.4%. The assessment was conducted using various performance indicators, including accuracy in classification, recall, specificity, precision, False-Negative Rate (FNR), False-Positive Rate (FPR), F1-Score, and Matthews Correlation Coefficient (MCC). Based on the experimental findings, it can be shown that, after making the required modifications, the RF framework demonstrated superior performance compared to all alternative models. The model attained a rate of accuracy of 96.66% and an F1-Score of 96.30%.
FL has seen significant expansion since its establishment in 2016. This expansion encompasses collaborative learning and knowledge fusion scenarios involving several organizations [25]. FL has three major subcategories: Federated Transfer Learning, Horizontal FL, and Vertical FL. All three subtypes conform to the essential concepts of FL, which entail decentralizing data pooling through weight sharing and aggregation among multiple clients and a global server. These entities can be distinguished by the discrepancies discovered in their data sources. FL has been identified as a secure and efficient approach for addressing data isolation and labeling challenges, as highlighted in [26,27]. This technique enables the development of ML models by enabling them to integrate and consolidate information from multiple entities while ensuring that the data remains limited to its sources. The researchers employed a client–server architecture in their FL strategy, integrating Federated Averaging (FedAvg) as suggested by [23]. The FedAvg algorithm integrates decentralized Stochastic Gradient Descent (SGD) on local nodes with a centralized server that conducts model averaging. However, the input mammograms were drastically downsampled in the study conducted in [28]. While low resolutions may be deemed satisfactory for density classification, the decrease in detail has a detrimental impact on the accurate classification of malignancy. Additionally, no domain adaptation strategies were employed in this investigation to address the domain shift arising from variations in pixel intensity distributions. In the study by Jimenez et al. [29], a distinct methodology was employed, wherein high-resolution mammograms were utilized in conjunction with federated domain adversarial learning. Furthermore, the researchers implemented curriculum learning in the context of FL to enhance the performance of classification tasks. The proposed methodology enhanced the alignment across various domains and effectively resolved the challenge of domain shift by employing federated adversarial adaptation of domains. This work employs three different Full Field Digital Mammography (FFDM) datasets. It presents experimental results that support the effectiveness of the proposed memory-aware curricular method in improving classification performance.
Ma et al. [30] employed a hybrid approach, integrating the FL framework with Convolutional Neural Networks (CNN), in order to create a federated prediction model. This study showcased improvements in overall modeling and simulation conditions for five distinct forms of cancer. The cancer data exhibited a degree of accuracy that was above 90%, thereby outperforming single-model machines, tree models, linear models, and neural networks. Nevertheless, this work only compared several models beyond MLP. Additionally, it failed to address the concern of data imbalance and its corresponding treatment. Tan et al. [31] proposed developing an FL system that focuses on extracting features from individual participant settings instead of a Centralized Learning (CL) system. The research paper presents several innovative contributions. The initial stage employs transfer learning to retrieve features of data from the Region of Interest (RoI) from an image. This strategy aims to facilitate meticulous pre-processing and data augmentation for training. Additionally, the research in [31] employs the Synthetic Minority Oversampling Technique (SMOTE) for data processing. This technique aims to attain a greater level of consistency in data classification and improve the accuracy of diagnostic predictions for diseases. In addition, the research utilizes the FeAvg-CNN + MobileNet model within the FL framework to safeguard client privacy and personal protection. Finally, the work offers empirical findings derived from DL, transfer learning, and FL models. These models were evaluated using balanced and imbalanced datasets in the domain of mammography. Table 1 shows the comprehensive literature review summary.
The authors in [25] have designed an FL framework to classify breast cancer based on histopathology pictures. This study suggests using an FL methodology to tackle security issues effectively. The proposed concept entails the dissemination of patient model parameters in order to facilitate the integration of knowledge. Hence, the application of FL led to the exclusion of data sharing, and the inquiry was carried out employing the BreakHis dataset. The simulation study results indicate that the built FL model exhibited classification performance on par with centralized learning. Abunasser et al. [32] introduced a DL-based Xception model to enhance breast cancer detection and categorization/classification. The dataset was initially acquired using the Kaggle repository and afterward underwent pre-processing to eliminate unwanted noise. Subsequently, the dataset was partitioned into three distinct categories: validation, training, and testing.
The created exception model received training using the provided training dataset. In the training phase, a data augmentation method was employed to address the issue of overfitting. The layers of the exception model are tasked with extracting the most significant features, which are subsequently utilized by the ultimate softmax classification process to categorize breast cancer. The findings from the simulation analysis indicate that the created model exhibited a significant level of proficiency. Nevertheless, it is important to acknowledge that there was a rise in the duration of processing. Ogier et al. [33] focused on predicting the incidence of triple-negative breast cancer within the framework of FL. In order to guarantee the protection of the confidentiality of the data, this study implemented the concept of FL. The utilization of FL has effectively safeguarded the confidentiality and integrity of the patient’s data. The study proves that conventional ML techniques mostly utilized entire slide images to detect Neoadjuvant Chemotherapy (NACT) responses. Nevertheless, the collective training of ML techniques enhances the effectiveness in detecting triple cancer. Moreover, this study provides evidence to support the assertion that the FL paradigm is well-suited for implementation in real-time applications.
This literature review provides a comprehensive overview of multiple studies pertaining to the detection and diagnosis of breast cancer. It specifically emphasizes the utilization of machine learning and federated learning methodologies. These studies exhibit encouraging outcomes in relation to precision and efficacy. Nevertheless, it is common for these methodologies to have limited comparisons with alternative approaches, and they may not adequately consider issues related to scalability, data imbalance, and privacy. Although these methodologies demonstrate promise in augmenting the detection of breast cancer, additional study is needed to remove their inherent constraints and render them more applicable in real-world clinical settings.

3. Proposed Methodology

The dataset utilized in this study is the Breast Cancer Wisconsin (BCW) Diagnostic dataset with dataset details in Table 2. The dataset consists of 569 observations and 32 variables. For instance, the dataset has 569 patients and 32 columns. These columns include a patient ID, 30 attributes derived from 10 morphological qualities, and a final column denoting whether the patient has been diagnosed with malignant or benign tumors (see Table 3 for patient data details). The pathologists have conducted a preliminary analysis of these traits and then recorded them in tabular form. It is imperative to acknowledge the necessity of including the worst-case values for all ten cell nuclei morphological features in each patient [34]. Among the 569 cases, the class distribution is as follows: 357 cases belong to the Benign class, marked by the numerical value 0, whereas 212 cases belong to the Malignant class, denoted by the numerical value 1. Furthermore, it was determined that none of the columns exhibited any missing values. Therefore, it is necessary to consider the presence of 30 input features in order to gain a comprehensive understanding of constructing a classification prediction model. Furthermore, it may be argued that a well-balanced and non-skewed distribution of class labels contributes to the favorable convergence of loss in ML algorithms, ultimately leading to the attainment of appropriate solutions [35].
According to recent reports, evidence suggests that an increased number of independent features contribute to the development of ML models that exhibit a reasonable level of accuracy when fitting the dataset [36]. Nevertheless, the inclusion of a substantial number of characteristics, particularly in cases with inadequate data points, hinders the generalization capability of prediction models [36]. Additionally, multi-collinearity may result in confounding effects and contribute to increased uncertainty when predicting the parameter values of the FL model. The BCW Diagnostic dataset exhibits a notable disparity in the number of features, namely, 30, compared to the diagnosis outcomes of 569 individuals, including benign and malignant classifications. Table 4 represents the complete set of 30 features of the BCW diagnosis dataset along with their description.Hence, a two-step method should be employed, involving feature selection techniques in conjunction with the ML classification model. The conventional approach involves utilizing all available features for training the classification model. In contrast, our proposed framework employs a two-step process. In the first stage, 30 features undergo a feature selection approach named L1 Regularization (L1) to identify the most relevant features. These optimal features are then utilized for training the classification model. Subsequently, the ML model, namely Deep Neural Network (DNN), is trained and assessed using the feature sets obtained from the feature selection technique. The selection procedure is carried out by utilizing essential classification assessment criteria, such as Accuracy, overall Precision, and Recall (particularly for malignant cases), on a distinct set of test data.

3.1. Data Pre-Processing

In this phase, we acquire and preprocess the data for FL, as shown in Equations (1)–(3). Each client i has its local dataset, denoted as D i , where D i = ( X i , y i ) . Here, X i represents the feature matrix containing input data and y i is the label vector containing target values. Data cleaning is an essential step, which involves addressing missing values ( M m i s s i n g ), handling outliers ( O o u t l i e r s ), and identifying and eliminating duplicates. The resulting cleaned dataset is denoted as D c l e a n e d . Additionally, label encoding is applied to convert categorical labels into a numeric format. We use a label encoder L to perform this transformation, producing encoded labels y ^ i from the original labels y i .
D i = ( X i , y i )
D c l e a n e d = C l e a n ( D i , M m i s s i n g , O o u t l i e r s )
y ^ i = L ( y i )
In this study, a comprehensive integration of the following preprocessing procedures has been undertaken in order to enhance the data preparation process. The dataset preprocessing steps involve a systematic approach to refine and enhance the quality of the BCW (Diagnostic) dataset before its utilization. Each step serves a specific purpose in addressing challenges such as outliers, class imbalance, and limited data diversity. Table 5 summarizes the steps and values involved in data preprocessing.
  • Original Dataset Distribution: The initial dataset consists of 569 instances, with 357 instances categorized as Benign (B) and 212 as Malignant (M). This distribution forms the foundation for subsequent preprocessing steps.
  • Removal of Outlier: Identifying and eliminating outliers is a critical step in the preprocessing stage, wherein outliers are methodically identified and addressed based on the statistical characteristics of the dataset. The dataset has been analyzed to determine important statistical measures, including the first quartile (Q1), the third quartile (Q3), and the interquartile range (IQR). The IQR is a statistical metric that offers significant insights into the dispersion of a given dataset. The calculation involves determining the disparity between the Q1 and the Q3. Q1 represents the 25th percentile, and Q3 represents the 75th percentile. The IQR is calculated as Q3–Q1 with a threshold for identifying potential outliers 1.5 × IQR. The data points are considered potential outliers if they fall below the Q1 − Threshold or above the Q3 + Threshold. Table 6 represents the IQR for each parameter. One pivotal element of this methodology entails the establishment of a threshold, which has been delineated as the equivalent of 1.5 times IQR, functioning as a criterion for detecting probable outliers. In statistical analysis, outliers are commonly defined as data points that lie beyond the range of (Q1 − 1.5 times the interquartile range) or (Q3 + 1.5 times the interquartile range). Following the conclusion of the identification phase, the dataset has been effectively purged of any outliers. After removing outliers, the number of observations is 538.
  • After Outliers Removal: To ensure that extreme values do not unduly influence the model, outliers are removed from the dataset. This refinement results in a reduced but more robust dataset comprising 538 instances, with 340 instances labeled as Benign (B) and 198 as Malignant (M).
  • Data Augmentation: Data augmentation has been employed to expand the dataset by generating extra data samples, hence obviating the need to gather fresh real-world data. Gathering large-scale annotated datasets proves to be exceedingly challenging given the scale that is required. Augmentation stands out as a highly effective interface for impacting the training of Neural Networks [37]. Table 7 and Table 8 show the dataset before and after augmentation values. For instance, the values of Min and Max for ‘ s m o o t h n e s s _ s e ’ in Table 7 are 0.0 and 0.03, respectively, whereas, in Table 8, the values for Min and Max are −0.01 and 0.03, which represents instances with smoother textures. This augmentation introduces a new dimension to the feature, allowing the model to recognize cases with even lower smoothness. Similarly for ‘ c o n c a v i t y _ w o r s t ’, Table 7 shows a Min value of 0.0 and a Max of 0.77, whereas Table 8 shows a Min of −0.01 and a Max of 0.78. This indicates a subtle change in the representation of worst-case concavity. This augmentation introduces a nuanced variation in the dataset, potentially improving the model’s ability to generalize to a broader range of scenarios. These differences in minimum and maximum values showcase the impact of data augmentation in capturing more diverse and nuanced patterns within the dataset, which can contribute to improved model robustness. Likewise, several characteristics display changes in their statistical qualities, indicating the impact of the enhancements. Specifically, attributes such as mean, standard deviation, minimum, and maximum values exhibit variations that signify the augmentation effects. The augmentation process entails the deliberate introduction of controlled variations or perturbations to the existing data points, expanding the dataset’s size and diversity. Implementing the data augmentation method involves the introduction of controlled random noise to the feature values of each data point, enhancing its practicality. To maintain the inherent characteristics of data, it is conventional to ensure that the amplitude of the noise remains minimal.
The enhanced data samples are subsequently produced by introducing alterations to the original data points using this noise while ensuring that their target labels remain unchanged. By incorporating these variances into the dataset, the methodology improves the model’s capacity to acquire knowledge from the data and apply it to samples that have yet to be previously encountered. This procedure helps to reduce the likelihood of overfitting and promotes the enhanced performance of the model, especially in scenarios where the initial dataset may have limitations or imbalances. Data augmentation techniques are employed to enrich the dataset and expose the model to a broader range of scenarios. To expand the dataset’s size, we introduced noise ranging from 0 to 0.01 into the data. This resulted in a total of 5742 records. The augmented data contained 3212 Benign and 2530 Malignant records.
  • Adding Noise: Noise, ranging from 0 to 0.01, is introduced into the dataset. This augmentation significantly increases the dataset size to 5742 instances, with 3212 instances representing Benign (B) and 2530 instances representing Malignant (M). The addition of noise contributes to the model’s ability to generalize by exposing it to a more diverse set of data points [38].
  • SMOTE for Class Balance: Synthetic Minority Over-sampling Technique (SMOTE) is applied to address class imbalance [39]. This technique involves generating synthetic samples for the minority class to balance class distributions. After applying SMOTE, the dataset achieves balance with 3212 instances for both Benign (B) and Malignant (M) classes, resulting in a total of 6424 instances. This balance is crucial for training a machine learning model that is not biased toward the majority class.
The dataset preprocessing steps collectively contribute to creating a more reliable, diverse, and balanced dataset. The refined dataset, encompassing various instances of Benign and Malignant cases, is poised for effective utilization in subsequent machine learning models, particularly deep neural networks, where the quality of the input data is paramount to the model’s performance and generalization capabilities.

3.2. Feature Selection

It reduces the number of input features by eliminating those that are not significant. Selecting the appropriate features helps reduce the computational cost, improves model performance, and reduces overfitting [40]. The selection of pertinent characteristics holds significant significance in enhancing the precision and comprehensibility of ML models, specifically in binary classification for breast cancer diagnosis. The primary aim of this phase was to distinguish between cancerous (1) and benign (0) conditions within the Breast Cancer Dataset. It was achieved by carefully choosing and defining a subset of qualities that provided the most relevant and useful data. By carefully choosing a suitable combination of features, we have effectively enhanced the model’s capacity to classify breast cancers while preserving its comprehensibility and simplicity. The primary aim of the feature selection procedure utilized in this study’s methodology was to determine the most relevant features in the binary classification task of cancer identification. The first version of the Dataset included 30 features, each containing unique information related to different aspects of tumor characteristics.
To determine the optimal set of features, we employed a feature selection approach that integrates L1 (Lasso) regularization into a DNN framework. L1 regularization allocates coefficients to individual features, thereby quantifying their impact on the predictive outcomes of the model. Features that had non-zero coefficients were deemed crucial for the classification job. In contrast, features with coefficients lowered to zero were seen as less significant and eliminated from the model. The rigorous feature selection method yielded a total of 25 features that were deemed essential for the proper classification of breast cancers. The parameters mentioned above encompass crucial tumor attributes that substantially differentiate between malignant and benign cases. The 25 features were chosen based on their ability to enhance the predicted accuracy of the model while minimizing its dimensions and complexity. There are numerous advantages associated with feature selection in the given situation. Firstly, implementing this approach ensures that the model prioritizes the most pertinent qualities, enhancing its ability to differentiate between different entities or categories.
Additionally, it aids in mitigating overfitting by eliminating extraneous or repetitive features that impede the model’s ability to generalize to unfamiliar data. Furthermore, the utilization of a limited number of variables contributes to the interpretability of the model, enabling healthcare practitioners to discern the tumor characteristics that hold the greatest significance in identifying the presence of cancer. Here is the mathematical modeling of the feature selection methodology:

3.2.1. Feature Selection Using L1 Regularization for Breast Cancer Classification

L1 (Lasso) regularization in DNN is used to perform feature selection for breast cancer classification based on Equation (4).
Minimize J ( θ ) = 1 m i = 1 m y ( i ) log h ( θ T x ( i ) ) + ( 1 y ( i ) ) log 1 h ( θ T x ( i ) ) + λ θ 1
where m is the number of data samples; x ( i ) represents the feature vector for the ith sample; y ( i ) is the binary targeted variable 1, which signals cancer, and 0, which does not. In addition, θ reflects LR model coefficients weights for each feature, h ( θ T x ( i ) ) is the sigmoid function maps linear feature combinations θ T x ( i ) to a probability between 0 and 1, λ reflects the L1 regularization strength parameter, and | θ | 1 is the L1 norm (sum of absolute values) of the weight vector θ . Reducing the aforementioned objective function promotes the tendency for numerous feature weights to attain a value of zero, efficiently choosing a subset of characteristics that significantly contribute to the classification task. The features that have been chosen offer significant insights into the diagnosis of breast cancer.

3.2.2. Train, Test, and Validate Dataset

In order to implement FL, the global dataset is carefully divided into distinct subsets that are specific to each client. Every individual client is provided with a specific dataset, guaranteeing the utmost protection of data privacy. A 70:15:15 partitioning strategy is utilized to divide the data into training, testing, and validation sets, facilitating localized model training under our federated system. A randomized order of data instances is shuffled and disseminated to all participating clients.

3.3. Model-Building

Using DNNs, client-side training enables each client to train localized models on their datasets. This decentralized method adds to the global federated model’s collective intelligence in addition to protecting data privacy.
  • Deep Neural Network Framework: In this phase, each client i trains its local model, denoted as M i , using a DNN. The DNN takes the feature matrix X i and the encoded labels y ^ i as input. This local model M i is trained independently on the client’s data, allowing each client to capture local patterns and knowledge, as shown in Equation (5).
    M i = DNN ( X i , y ^ i )
    The primary element of the client-side training process is the meticulously designed DNN, specifically tailored to address the binary classification challenge. The architectural architecture illustrated in Figure 2 consists of a solitary input layer and three hidden layers, along with one output layer, with each holding unique and well-defined attributes.
    • Input Layer: The input layer is designed to align with the dimensions of the feature vectors present in the dataset, successfully adapting the unique qualities of the data.
    • Hidden Layers: The DNN is composed of three hidden layers that have been specifically developed to methodically extract and represent increasingly intricate features from the input data. Each hidden layer has a form of connectivity known as dense connectivity. We used Rectified Linear Units (ReLU) as the activation function. It enhances the level of non-linearity, hence facilitating the model in efficiently capturing complex relationships that exist within the data.
    • Fully Connected Layer: This layer, sometimes referred to as a dense layer, serves as an intermediate stage connecting the hidden layers as well as the output layer. The fully connected layer, similar to the hidden layers, employs the ReLU function for activation. The fully connected layer has a carefully selected number of neurons that enable efficient feature transformation while ensuring the model remains manageable.
  • Output Layer: The final layer of our DNN-based FL design, known as the output layer, is of utmost importance in producing predictions for binary classification. This layer comes with the Sigmoid activation function, which is frequently employed in binary classification applications. The application of this activation function leads to the production of probabilistic outputs that cover the numerical interval from 0 to 1, hence facilitating the estimation of class probabilities. The output layer comprises two neurons, each representing one of the two binary classes. The utilization of the Sigmoid activation function enables individual neurons to evaluate the likelihood of membership within their corresponding class autonomously.

3.4. Model Compilation

The model compilation process encompasses the meticulous specification of the optimizer and loss function.
  • Selection of Optimizer: We utilize the Adam optimizer due to its effectiveness in gradient-based optimization [41]. The learning rates are dynamically adjusted in order to accelerate the process of convergence.
  • Loss Function: As seen in our code implementation, the sparsely categorical cross-entropy loss function was selected. This option is suitable for problems involving binary classification since it measures the discrepancy between the expected and actual class labels [42].

3.5. Training Parameters

The selection of training parameters plays a crucial role in attaining both model convergence and generalization. The parameters are carefully configured in order to examine their influence on the training of the model. A variety of epochs (i.e., 5–70) are utilized to investigate different training scenarios.
  • Number of Epochs: We systematically investigate the impact of varying the number of training epochs. This range of exploration enables us to examine the behavior of the model across different stages of training, encompassing both rapid convergence and prolonged periods of refinement.
  • Batch Size: The code implementation uses a batch size of 32. The selection of this particular configuration is made with the intention of achieving a harmonious equilibrium between computing efficiency and the stability of the model. The modification of batch size can significantly affect the rate at which convergence occurs and the amount of memory used, making it a crucial factor that requires careful deliberation in practical applications.
  • Early Stopping Mechanism: The implementation of an early stopping mechanism, with a patience of 10, is crucial in mitigating the issue of overfitting. The patience parameter denotes the duration, measured in epochs, during which training will continue without any improvement in validation loss before it is terminated. The value is chosen carefully to guarantee that the model terminates training when additional iterations are unlikely to result in substantial enhancements.

3.6. Model Training and Saving

The key aspect of the client-side training function lies in the utilization of the backpropagation technique for the iterative optimization of the DNN model. Throughout various periods, the model acquires the ability to establish a correlation between input characteristics and desired outcomes, gradually improving its capacity to make accurate predictions. Simultaneously, the training process diligently monitors the validation loss in order to identify any indications of overfitting and implements early stopping if necessary. After the completion of successful training, the trained DNN model is safely stored in an archive. The saved model files play a crucial role in the succeeding stages of our FL process. Significantly, every client produces and maintains its unique model file, which secures individual data privacy and autonomy while also contributing to the collective intelligence encapsulated inside the federated model. Nomenclature Section presents a quick description of the symbols used in the section.
Algorithm 1 represents the Local Model Training of the ML model M i at a given client. The input of the system is the local dataset D i = ( X i , y ^ i ) , where X i represents the feature matrices and y ^ i represents the encoded labels. The method progresses by undergoing a sequence of iterations, with each iteration comprising numerous epochs. To ensure unpredictability, the dataset is randomized inside each epoch. The technique of batching is utilized to partition the dataset into smaller segments. Within each batch, the local model calculates predictions y ^ p r e d using the present model weights. The loss function L b a t c h is then computed to evaluate the model’s performance on the given batch. Following this, the weights of the model are adjusted by a weight update function, typically employing gradient descent, in order to minimize the loss. The procedure mentioned above is iterated for the designated number of epochs, yielding a trained local model denoted as M i . The model, as mentioned earlier, acquires and incorporates client-specific patterns and information at the local level from its localized dataset, which will subsequently be employed in the context of FL.
Algorithm 1 Local Model Training Algorithm
  1: 
Input: Local dataset D i = ( X i , y ^ i ) , Neural Network architecture, number of epochs, batch size
  2: 
Output: Trained local model M i
  3: 
procedure Local Model Training( D i , Neural Network, epochs, batch size)
  4:
    Initialize local model: M i
  5:
    for each epoch in 1 to e p o c h s  do
  6:
        Shuffle D i randomly
  7:
        for each batch in D i with batch size b a t c h _ s i z e  do
  8:
            Extract batch data: X b a t c h , y ^ b a t c h
  9:
            Compute predictions: y ^ p r e d = M i ( X b a t c h )
10:
           Calculate loss: L b a t c h = L o s s ( y ^ b a t c h , y ^ p r e d )
11:
           Update model weights: M i . u p d a t e _ w e i g h t s ( L b a t c h )
12:
        end for
13:
    end for
14: 
end procedure
15: 
Return Trained local model M i

3.7. Loop across Clients

Within the framework of our FL orchestration, each client autonomously performs the client training function—the procedure mentioned above results in the production of unique model files. The model files serve as repositories for the knowledge and insights acquired by individual clients throughout the training process. The division of these models highlights our firm dedication to protecting the confidentiality and accuracy of client data while also contributing to the overall enhancement of intelligence in our federated model.

3.8. Server-Side Model

  • Model Aggregation Function: The FL system we have developed concludes with the server-side model aggregation function FedAvg algorithm, which plays a crucial role in combining the varied insights obtained from individual clients [43]. In the implemented code, this particular function has been methodically designed and developed to accomplish its intended purpose effectively. The function for model aggregation on the server side is designed to accept client models as input and provide a federated model as output. The code is designed in a manner that explicitly facilitates the processing of models from different clients. An essential element of this function is its ability to produce federated forecasts. The procedure, as mentioned above, is accomplished by successively executing the client models, producing predictions for a cohesive set of inputs, and subsequently merging these forecasts. The strategy described above forms the basis of the collaborative process of decision-making within our federated architecture.
    To create a global model, server-side model aggregation is performed. This function aggregates the local models from all clients into a federated model, denoted as M f e d , as shown in Equation (6). The federated model is a collective representation of knowledge learned from all clients. It leverages the insights captured by individual clients during local training to make global predictions.
    M f e d = Aggregate _ Models ( { M 1 , M 2 , , M n } )
  • Federated Model: After model aggregation, the federated model M f e d is ready for global training, see Equation (7). It is compiled with an optimizer, typically Adam, and a loss function, such as Sparse Categorical Cross-Entropy. The global training phase fine-tunes the federated model using the aggregated knowledge from all clients, improving its overall performance.
    M f e d . c o m p i l e ( o p t i m i z e r ) = A d a m ( ) , l o s s = S p a r s e _ C a t e g o r i c a l _ C r o s s E n t r o p y
    The fundamental element in our FL framework is around the server-side aggregation of client models. The procedure above entails the amalgamation of information derived from distinct client models in order to construct a unified federated model. The aggregation function utilized on the server side performs an iterative procedure that traverses the models provided by several clients, thereby combining their respective contributions. The aggregation approach is the process of adding together the predictions made by separate client models. It allows the federated model to utilize the collective knowledge obtained from diverse data sources. The federated model is built upon the aggregated forecasts, which are derived from the design of each of the local models. The federated model represents a unified embodiment of the combined intellectual capacity originating from multiple participating clients.
    Algorithm 2 illustrates the process of global model training involving the development and optimization on a global scale. The method of global model training involves the aggregation of predictive capabilities exhibited by individual client models within the framework of FL. The process commences by collecting forecasts from individual client models using validation data. The predictions mentioned above are aggregated to provide a unified collection of predictions referred to as “client_predictions”. The outcome, denoted as g l o b a l _ p r e d i c t i o n , is ascertained by selecting the class that possesses the highest probability. The resultant global federated model, referred to as F e d e r a t e d _ M o d e l , encompasses the collective knowledge derived from all clients involved in the process. The utilization of a collaborative approach significantly improves the model’s capacity to generalize to novel data and generate precise predictions while operating in a decentralized and privacy-preserving manner. Consequently, this renders it an influential tool for jobs involving FL.
  • Final Classifier
    The final classifier used for making predictions is the federated model M f e d , shown in Equation (8). This model embodies the collective intelligence learned from all clients during the FL process. It serves as the ultimate decision-maker for tasks like classification and inference.
    F i n a l _ C l a s s i f i e r = M f e d
    In the proposed Algorithm 3, the participation of each client in model training is based on its dataset D i , which may be represented as D i = ( X i , y i ) . This dataset comprises feature matrices X i and label vectors y i . Data preprocessing involves several important stages to assure the quality of the data, including the treatment of missing values ( M m i s s i n g ) and outliers ( O o u t l i e r s ). The process of encoding labels involves the utilization of a label encoder denoted as L, which subsequently produces encoded labels denoted as y ^ i . The local client models, denoted as M i , are trained in a separate manner using DNN with the input data X i and the corresponding predicted output y ^ i . The variable i is in the range from 1 to n, representing each participating client and iterating over all clients in the federated learning setting. The aforementioned local models can collect patterns that are distinct to individual clients. The process of model aggregation involves the integration of individual local models into a federated model, denoted as M f e d , which serves as a representation of the collective knowledge. The global training of M f e d is conducted, and it afterward functions as the ultimate classifier. The key symbols utilized in this study encompass D i , M m i s s i n g , O o u t l i e r s , L, y ^ i , M i , and M f e d .
Algorithm 2 Global Model Training
  1: 
Input: List of client models [ C l i e n t 1 , C l i e n t 2 , , C l i e n t N ]
  2: 
Input: Number of clients N
  3: 
Output: Global federated model F e d e r a t e d _ M o d e l
  4: 
Initialize an empty list p r e d i c t i o n s
  5: 
for  i 1 to N do
  6:
     p r e d i c t i o n s i C l i e n t i . p r e d i c t ( X )            ▹ Make predictions on validation data
  7:
     p r e d i c t i o n s . a p p e n d ( p r e d i c t i o n s i )                  ▹ Aggregate predictions
  8: 
end for
  9: 
c o m b i n e d _ p r e d i c t i o n s s u m ( p r e d i c t i o n s )              ▹ Sum predictions across clients
10: 
f i n a l _ p r e d i c t i o n s a r g m a x ( c o m b i n e d _ p r e d i c t i o n s )  ▹ Select class with the highest probability
11: 
Initialize the global federated model F e d e r a t e d M o d e l
12: 
Return   F e d e r a t e d M o d e l
Algorithm 3 Federated Learning Algorithm
  1: 
Input: Local datasets D i = ( X i , y i ) for clients i, M m i s s i n g , O o u t l i e r s , label encoder L
  2: 
Output: Federated model M f e d
  3: 
procedure  Data Acquisition and Pre-processing
  4:
    for each client i do
  5:
        Load local dataset D i = ( X i , y i )
  6:
        Clean D i using M m i s s i n g , O o u t l i e r s
  7:
        Encode labels: y ^ i = L ( y i )
  8:
    end for
  9: 
end procedure
10: 
procedure  Local Model Training
11:
    for each client i do
12:
        Initialize local model: M i
13:
        Train M i using DNN with X i and y ^ i
14:
    end for
15: 
end procedure
16: 
procedure  Model Aggregation
17:
    Initialize federated model: M f e d
18:
    Aggregate local models: M f e d = Aggregate _ Models ( { M 1 , M 2 , , M n } )
19: 
end procedure
20: 
procedure  Global Model Training
21:
    Compile M f e d with optimizer (e.g., Adam) and loss (e.g., Sparse Categorical Cross-Entropy)
22:
    Train M f e d using global data
23: 
end procedure
24: 
procedure  Final Classifier
25:
    Output: Final classifier M f e d
26: 
end procedure
27: 
Return   M f e d

3.9. Federated Learning Rounds

The FL technique, as seen in Figure 3, is structured into multiple rounds, where each round corresponds to a distinct iteration in the collaborative learning procedure.
The execution of FL rounds occurs iteratively. In the code implementation, the number of rounds is denoted as ‘NUM–ROUNDS’. Every iteration encompasses a sequence of client training and server-side aggregation. During each iteration of FL, each client autonomously trains their models using their respective local datasets. The training approach employed in this study utilizes the client training function, as outlined in the established methodology. After the completion of client training, the server-side model aggregation function is called. The preceding section elucidates a procedure that entails the synchronization of client models as well as the development of an enhanced federated model.

3.10. Evaluation Parameters

  • Accuracy: The performance of the classifier is evaluated using accuracy as a metric. The metric under consideration is the proportion of accurately identified observations in relation to the total number of observations. The value of the variable falls inside the interval of 0 to 1. In the context of measurement or evaluation, a precision value of 1 denotes a high or optimal level of accuracy. In contrast, a number of 0 represents the lowest or poorest level of accuracy. The mathematical computation of a model’s correctness is possible [44]:
    A C C = ( T R p s + T R n g ) ( T R p s + T R p s + F L p s + F L n g )
    The variable A C C denotes the accuracy, whereas T R p s indicates the True Positive value in the label class, indicating instances where the model correctly recognizes a true label as true. The variable T R n g denotes the True Negative metric in the label class, indicating instances where the model correctly identifies incorrect labels. The variable F L p s denotes the False Positive metric, indicating instances where the model wrongly classifies a sample as false, whereas F L n g represents the False Negative metric, indicating instances where the model incorrectly classifies a sample as true. Table 9 represents the description of these variables in our case. It is important to aim for a balance between them as focusing too much on any one metric could lead to an imbalance in the others. For instance, trying to minimize F L n g (missing a cancer diagnosis) might increase F L p s (overdiagnosing cancer), which can lead to unnecessary treatments and stress for patients.
  • Precision: Precision is defined as the ratio of the total number of correct positive predictive values to the total number of positive predictive values [45]. Additionally, it provides a value that is between 0 and 1. A precision of 1 indicates a highly predicted value or the best value, while a precision of 0 indicates the least predictive or the worst value. Accuracy can be measured using mathematical formulas given in Equation (10).
    pr = T R p s ( T R p s + F L p s )
  • Recall: In addition, the recall bears the label Sensitivity. The accuracy rate is defined as the ratio of correctly detected positive values to the total number of samples that are classified as positive. The recall value is somewhere in the range of 0 to 1, where 1 indicates a strong recall and 0 indicates a lesser recall. The mathematical expression for it is shown in Equation (11).
    rc = T R p s ( T R p s + F L n g )
  • F1-Score: The F1-Score accurately recalls the weighted average (also known as F1-Score). This score includes false positives and false negatives. Assigning more weight to precision and less to recall yields 0.75 beta. When recall is weighted greater than precision, the beta value is 2. We choose ‘1’ for beta to balance precision and recall. The F1-Score frequently outperforms accuracy, especially in unequal class distributions. Accuracy is best when false negatives and positives have similar costs. If false positives and false negatives have different costs, Precision and Recall should be considered jointly. Between 0 (lowest) and 1 (best), small precision or recall yields a low value. The harmonic mean of precision and recall determines it. Equation (12) represents its mathematical form.
    F 1 S c o r e = 2 × ( pr × rc ) ( pr + rc )

4. Experimentation Setup and Results

The proposed work is implemented on a system described as a PC with an Intel (R) Core (TM) GPU @3.20 GHz CPU, 8.0 GB RAM, and a Windows 10 operating system. Python is utilized as a programming language in these experiments, and Google Colab is used as an IDE. It is a web-based programming environment with much interactivity. Python is a simple, open-source, and effective programming language with plenty of open-source modules and community support to construct graphs and statistical models. We picked it because it is the most recent and reliable version of Python. A local host is created by Google Colab so that the code may execute on the browser. The details of a dataset and implementation of the proposed framework are explained in the section below.
The dataset undergoes a shuffling process, ensuring a randomized order of data instances. Subsequently, this uniformly shuffled dataset is disseminated to all participating clients. This meticulous approach is essential to maintain consistency and fairness in the initial distribution of data across the federated network. There were two local clients and one global model. Each model is evaluated on specific performance measures. Each model’s performance is evaluated using Accuracy, F1-Score, Precision, Specificity, and Recall. It describes the performance of classifiers on a set for which true values are known. It helps understand errors the classifier makes [46].

4.1. First Iteration Results

The results after the first iteration of our local and global model are given below.

4.1.1. Local Model

  • Accuracy: During the preliminary phase, we conducted training and validation processes on the common augmented dataset, resulting in the development of two local models. The examination of the outcomes of each model is described below.
    The first performance of the first model depicted in Figure 4a exhibits encouraging indications. In the initial stages of training, the model exhibits a swift learning process and demonstrates adaptability to the provided training data, leading to a notable improvement in training accuracy. As anticipated, the model’s progress reaches a point of stabilization, suggesting that it is unable to enhance its performance on the training set further. The validation accuracy has a strong correlation with the training accuracy pattern but is continually trailing behind due to its exclusion from the model’s training procedure. At around 50s (continued to 70), both the training and validation accuracies exhibit stabilization, with the training accuracy converging to approximately 95% and the validation accuracy hovering around 90%. The observed convergence in this preliminary iteration indicates that the initial model successfully mitigates the issue of overfitting on the training data and exhibits a commendable capacity to generalize effectively to the novel, unobserved data. The results given in this study provide a strong foundation for the possible effectiveness of the model as a dependable predictor.
    Similarly, the second model depicted in Figure 4b demonstrates the same beginning trajectory. During the initials, the model demonstrates its ability to acquire knowledge from the training data efficiently, hence displaying its aptitude for capturing the fundamental patterns present within the dataset. As the training advances, it can be observed that both the training and validation accuracies exhibit a consistent pattern of convergence, which is consistent with the behavior observed in the initial model. At around 50 (continued to 70), the training accuracy achieves a level of approximately 95%, whereas the validation accuracy stabilizes at around 90%. It is worth mentioning that there is a short decline in the accuracy of validation at around the 10-epoch milestone. The initial iteration of the training process is attributed to the model’s dynamic learning and adaption. However, despite this initial performance setback, the second model, similar to the first one, exhibits the capability to efficiently generalize to unfamiliar data, indicating a promising start in its advancement.
    The initial iteration of our models demonstrates promising beginnings as they exhibit respectable performance, particularly given the constrained amount of training repetitions. Additional refinement and investigation will yield a more extensive comprehension of their capacities, yet these first findings indicate a promising outlook for these models as dependable prognosticators.
  • Loss: In the initial iteration, we examined the performance of both ML models. The models utilized several training methodologies, with a specific emphasis on the implementation of data augmentation strategies. The initial model, as depicted in Figure 5a, shows loss analysis focused on the surveillance of misclassified cancer cases during the training process. The graph illustrates a persistent decline in loss over time, as observed in both the training and validation datasets. The current trend suggests that the model’s capacity to identify the cancer dataset for detection accurately is continuously improving. Significantly, the validation data demonstrated a marginally greater loss in comparison to the training data, as expected, due to the fact that the validation data were not encountered during the training process. The observed discrepancy in loss values serves as an indication of the model’s practical efficacy when applied to novel data, confirming that it did not excessively adapt to the training data.
    Furthermore, the observed decrease in loss on the validation data over time suggests that there is potential for additional improvement in the model. The marginal disparity seen between the training and validation loss indicates the potential for further training to optimize the model’s performance on the validation dataset. The second Figure 5b illustrates that the model trained via data augmentation demonstrated a reduced loss in comparison to the model that was trained. The notable reduction in loss demonstrated how well data augmentation works to support the model’s ability to generalize to new inputs. Furthermore, the model’s loss exhibited a plateau at a reduced level when data augmentation techniques were employed, suggesting enhanced efficacy in the acquisition of knowledge from the training data. In the preliminary phase, both models had encouraging attributes. Model 1 consistently demonstrated an increase in accuracy and suggested the possibility of future improvements with longer training. In the second model, the utilization of data augmentation was observed to result in a noteworthy decrease in loss, hence underscoring the efficacy of these strategies in the context of cancer diagnosis.

4.1.2. Global Model

Table 10 provides an overview of the performance of the global model for the first iteration. The true positive (sensitivity) rate depicts that 97.68% of cancer patients are identified correctly with our proposed model. The correct prediction for patients without cancer also turns out to be good (i.e., 96.77%). On the other hand, low values of false positive and false negative rates establish that the model misclassification of patients with breast cancer and without breast cancer is low. The presented Table indicates an accuracy rate of 97.26%, indicating that the model successfully identified 96.26% of the observations. The precision metric is currently at 96.43%, signifying that 96.43% of the anticipated positive instances were indeed positive. In the current context, it is noteworthy that the recall rate stands at 97.68%, indicating that 97.68% of the true positive instances were accurately classified as such. Significantly, the model demonstrates a proclivity for producing erroneous positives as opposed to false negatives. It implies an increased probability of erroneously categorizing a negative observation as positive. Several aspects can contribute to this phenomenon, such as the quality of the data and the intricacy of the task. Hence, the performance of the model can be considered satisfactory; however, there is room for improvement.

4.2. Second Iteration Results

Here are the second iteration’s results.

4.2.1. Local Models

  • Accuracy: The graph presented in Figure 6a provides a visual representation of the performance trajectory of the ML model throughout 50 trainings (due to early stopping criteria). Upon close examination of the graph, a notable trend becomes apparent. Initially, it is clear that as the number of trainings increases, the model’s accuracy increases steadily and consistently. This indicates that the model can enhance its performance as its familiarity with the training data increases. A crucial observation emerges around 10 when the training accuracy begins to plateau. It suggests that additional training periods may not result in significant gains during the training phase. Significantly, the validation accuracy continues to improve even after 10, eventually surpassing the training accuracy. The difference between training and validation accuracy is reassuring evidence that the model avoids overfitting and is effective at generalizing its knowledge to new, unseen data. The second model, as depicted in Figure 6b, shows an accuracy score of 0.87, indicating that it correctly classified 97.26% of the data (in 70 epochs), which is a commendable accomplishment. Despite the possibility that additional training periods could result in minor enhancements, it is essential to recognize that an accuracy score of 0.97 already indicates a high level of performance. Analyzing the trend of this graph further emphasizes that the model travels from early rapid learning to a phase of stable performance. The disparity between training and validation accuracy demonstrates the model’s ability to generalize effectively and avoid overfitting. This trend demonstrates the efficiency and suitability of the model for the classification task at hand.
    In both graphs, a similar pattern is observed: an initial phase of rapid accuracy improvement as the models rapidly adapt to the training data. As training continues, however, a transition occurs in which training accuracy stabilizes and reaches a plateau, indicating diminishing returns from additionals. The behavior of validation accuracy distinguishes these models. In both cases, validation accuracy continues to increase even after training accuracy reaches a plateau, demonstrating the models’ ability to generalize to new, unknown information. This disparity between training and validation accuracy highlights the importance of avoiding overfitting, a vital performance indicator. Both models obtained high final accuracy scores, reaching 0.97%, demonstrating their data classification proficiency. These trends collectively illustrate the learning dynamics of the models, demonstrating their ability to balance learning from the training data with effective generalization, thereby making them appropriate for their respective classification tasks.
  • Loss: Upon careful examination of Figure 7a, a discernible trend becomes apparent. Initially, it is notable that as the model undergoes more training, its loss decreases consistently. This trend is extremely favorable, indicating that the model is able to adapt and improve its predictions over time. At 20, when the training loss begins to level off and reaches a plateau, a crucial observation is made. This phenomenon suggests that additional training periods may produce diminishing gains during the phase of training. In contrast, the validation loss continues to decrease until 40, exhibiting a distinct behavior. The difference between training and validation loss is a positive indicator that the model effectively prevents overfitting the training data as it can generalize its acquired knowledge to new, unobserved data with minimal error. The second graph shown in Figure 7b concludes with an ultimate loss score of 0.06, demonstrating the model’s ability to make highly accurate predictions with a small error rate. This outstanding final loss score demonstrates the model’s skill in its predictive assignment.
    These graphs illustrate how both ML models exhibit a distinct pattern: initial phases of rapid loss reduction as they rapidly adapt to training data. Both models reach a plateau in training loss around 10, indicating decreasing returns from additional trainings. However, the validation loss continues to decrease, demonstrating the models’ capacity to generalize knowledge to new, unobserved data without overfitting. The disparity between training loss and validation loss is a defining characteristic that demonstrates their balanced learning dynamics. While the final loss scores for the graphs are 0.06 and 0.5, respectively, they both indicate the models’ ability to make accurate predictions with varying degrees of precision. These tendencies demonstrate the models’ aptitude for data-driven learning, effective adaptation, and sustaining generalization capabilities while avoiding overfitting.

4.2.2. Global Model

Table 11 depicts the effectiveness of the global model with a 98% correctly classified (i.e., true positive rate) cancer patient. The correct prediction for patients without cancer also turns out to be good (i.e., 96.77%). In addition, a 3.23% false positive rate and a 1.89% false negative rate establish that the model misclassification of patients with breast cancer and without breast cancer is low. Upon comparing the present iteration with our findings of the first iteration, it is evident that in the second iteration, the model has shown improvement in its accuracy level, reaching 98%. This observation indicates the model’s consistent and reliable prediction capability. Moreover, the false negative rate decreased by 0.43% in the second iteration. This demonstrates how the model has performed better by showing a decline in the suggestion of a cancer patient as a negative example. Significantly, the precision is 96.49%, suggesting a greater proportion of accurate positive predictions among the overall positive predictions. Based on these data, it is clear that the model’s performance remains reasonably good. However, there is still potential for additional refining and enhancement. In order to further improve the classification capabilities, feasible strategies may involve investigating other classification algorithms or optimizing the parameters of the existing algorithm. These modifications have the potential to provide a more optimal trade-off between precision and recall, hence resulting in a classification model that is more resilient and dependable.

4.3. Third Iteration Results

The results of the third iteration are as follows.

4.3.1. Local Models

  • Accuracy: In the third iteration of our study, we proceed to assess the efficacy of a discussion of the performance of both models as depicted through a graphical depiction. The provided graphs depict the chronological progression of accuracy in the models, as mentioned earlier, offering significant insights into the models’ performance on both the training and validation datasets. The x-axis denotes the number of trainings, while the y-axis illustrates the accuracy, providing a lucid depiction of the model’s learning progression.
    In the first model depicted in Figure 8a, the evaluation of accuracy is determined by calculating the proportion of accurately classified predictions. At the outset, the training accuracy exhibits a notable elevation, although it progressively diminishes due to the model’s tendency to overfit the training data. In contrast, the validation accuracy demonstrates an initial lower value followed by a gradual increase, indicating a significant improvement in the model’s ability to generalize effectively to new and unseen data. Ideally, the convergence of training and validation accuracies indicates a balanced model. In the present scenario, convergence is observed but accompanied by a minor overfitting concern, as indicated by the training accuracy of 94% and the validation accuracy of 97% in 70 epochs.
    In the second graph, as depicted in Figure 8b, the training accuracy demonstrates convergence at approximately 90%. However, the validation accuracy exhibits convergence at approximately 95%. This indicates that the model exhibits good performance in the training and validation data and remains a commendable model in its entirety. The graph additionally illustrates that the accuracy of the model reaches a peak after approximately 40 epochs (which is due to early stopping criteria). This implies that the model’s learning capacity diminishes beyond this juncture with respect to the training data. The potential exists for enhancing the model by augmenting the number of s. Both models exhibit robust performance with high levels of accuracy.
  • Loss: The present findings entail an analysis of the performance of both models in the third iteration through the examination of their loss curves throughout the training process. In the beginning stages of the first model, as shown in Figure 9a, the training loss exhibits a relatively large value, which after that diminishes gradually as the model progressively adapts to represent the training data accurately. Concurrently, the initial validation loss exhibits a higher value, which then diminishes as the model progressively acquires the ability to generalize to novel data. It is crucial to emphasize that the convergence of both training and validation losses indicates that the model effectively mitigates the risk of overfitting the training data. The training loss demonstrates stability at a value of roughly 0.16, whereas the validation loss remains rather constant at around 0.1. The findings presented in this study clearly indicate that the model exhibits remarkable performance, as evidenced by its low loss metric, hence indicating its efficacy.
    In the case of the second model, as shown in Figure 9b, it is observed that the training loss exhibits a significant drop early on, indicating its ability to acquire essential data patterns. Nevertheless, it is worth noting that the loss curve reaches a plateau after approximately 20 epochs, suggesting that there is a decrease in the amount of knowledge gained from the training data. In the interim, it is observed that the validation loss demonstrates an initial decrease, followed by an upward trend after about 20 epochs. In an optimal situation, the training and validation losses would exhibit convergence, indicating a model that avoids both overfitting and underperformance. Nevertheless, in this particular instance, the model demonstrates certain tendencies towards overfitting. Although the model exhibits a robust foundation, the potential exists for enhancement through the exploration of tactics such as the reduction in training epochs or the adoption of alternative regularization techniques. Hence, it is concluded that both models demonstrate a notable reduction in loss values, hence confirming their efficacy.

4.3.2. Global Model

During the concluding stage of our model evaluation, we conduct a comprehensive assessment of its performance depicted in Table 12. It presents the effectiveness of the global model in the third and final iteration. There are 98% correctly classified (i.e., true positive rate) cancer patients. The correct prediction rate for patients without cancer also turns out to be good (i.e., 96.77%), similar to previous evaluations of the model. Similarly, a 3.23% false positive rate and a 1.89% false negative rate provide empirical support for the accurate predictions by our model. These results exemplify the proficiency of the model in accurately recognizing and classifying.
The recall statistic, which quantifies the proportion of accurately predicted positive instances relative to the overall count of real positive instances, is approximately 98.0%. This finding indicates that the model exhibited a significant level of precision in effectively capturing and classifying about 98.0% of the positive cases. Nevertheless, it is important to acknowledge that occurrences of misclassifications exist. The presence of 3.32% false positives suggests that our model exhibited misclassification by identifying the events as positive when, in reality, they were negative. This phenomenon adversely impacts the model’s accuracy by encompassing instances where the model’s predictions for positive outcomes were determined to be erroneous. Furthermore, the occurrence of false negative instances suggests that our model failed to correctly identify 1.89% true positive cases, incorrectly categorizing them as negative. It has a negative impact on the recall metric since it implies that there are instances where positive cases were not detected.
Therefore, our model demonstrates a significant degree of performance that includes a significant percentage of T R p s as well as T R n g , consequently leading to favorable precision (i.e., 96.5%), accuracy (i.e., 97.54%), and recall (i.e., 98.0%). The presence of a finite number of F L p s as well as F L n g cases suggests the possibility of making improvements to reduce these misclassifications, hence enhancing the reliability and precision of the model.

4.4. Overall Results

Figure 10 illustrates the progressive evolution of our global breast cancer detection model’s performance through numerical metrics. It displays the accuracy, precision, recall, and F1-Score across three iterations. Notably, the graph reveals a steady improvement in the model’s capabilities over time. In the first iteration, all performance indicators, such as precision, recall, precision, and F1-Score, exhibit considerably lower values. However, as the model undergoes refinement, which is particularly evident in the third iteration, there is a remarkable enhancement across all these crucial evaluation criteria. Of particular significance is the F1-Score, a comprehensive measure of the model’s overall performance. It exhibits a notable ascent from 97% in the first iteration to a robust 97.24% in the third iteration. This substantial increase indicates that the model has become significantly more accurate, precise, and efficient at capturing relevant instances while minimizing false negatives. Overall, the graph serves as a promising indicator of the model’s potential. It vividly illustrates the model’s progressive improvement, underlining its increasing effectiveness in accurately detecting breast cancer cases based on numerical data. This upward trend instills confidence in the model’s ability to continually enhance its diagnostic capabilities, ultimately benefiting medical applications and patient care.

4.5. Comparison with State of the Art Studies

The following table provides a detailed comparison between our study and other research endeavors in the domain of breast cancer detection and diagnosis. We conducted comparisons with existing works that utilized similar datasets or methodologies like federated and deep learning. While there might be variations in specific datasets, we ensured that the compared works shared similarities in terms of the nature and characteristics of the data. The results presented in Table 13 exhibit exceptional performance in various important criteria, including accuracy, precision, recall, and F1-Score. Significantly, our methodology integrates Federated Learning (FL) and Deep Neural Networks (DNN), resulting in an accuracy rate of 97%, precision of 96%, recall of 98.0%, and an exceptional F1-Score of 97%. The noteworthy outcomes highlight the capacity of our technology to potentially transform breast cancer diagnosis while simultaneously prioritizing data privacy and fostering collaboration among healthcare institutions. Additional research and the possibility of incorporating these findings into clinical applications should be pursued.
The methodology employed in our study offers a novel and structured approach to showcasing the outcomes, successfully portraying the progressive development of our research. By conducting a thorough examination of each version, we aim to clarify the incremental process of improvement, thereby providing insights into the ongoing development of our breast cancer detection model. This approach reveals the hierarchical structure of improvements and emphasizes the creative nature of our methods in communicating the advancements made in our work. The final iteration of our project demonstrates the effectiveness of our approach as it exhibits exceptional levels of precision, precision, and recall scores for the detection of breast cancer. The utilization of this innovative approach to presenting information exhibits considerable potential in furthering the development of medical AI applications [22].

5. Limitations

The data set utilized in this study exhibits a relatively modest size and is confined to a specific dataset. The constrained variability of the dataset has imposed limitations. Furthermore, the findings presented in this study demonstrate encouraging outcomes in terms of accuracy, along with other performance metrics. However, it is imperative to emphasize the significance of clinical validation in real-world testing as an essential stage in the research process.

6. Conclusions

Traditional methods of breast cancer screening often have high false positive rates, leading to unnecessary stress and medical procedures for patients. Additionally, they sometimes fail to detect cancer at its earliest stages when it is most treatable. In this context, AI offers a promising solution. It is crucial to recognize AI’s role in breast cancer diagnosis as an ever-evolving area that requires constant and in-depth exploration to comprehend its potential and limitations fully. New advancements in AI ultimately lead to improved patient outcomes. This study illustrates a significant leap in breast cancer diagnosis by employing an iterative approach to collaborative learning and utilizing DNN-based Federated Learning. It successfully amalgamates the collective knowledge and data resources from various healthcare institutions while ensuring patient privacy and data security measures. It paved the way for more effective early detection and treatment of breast cancer. However, it also highlights the importance of continuous exploration and innovation in this area to fully leverage the potential of these advanced technologies for improving patient care. In the future, we aim to work on the diagnosis of other cancer types, such as skin cancer.

Author Contributions

Conceptualization, M.F.A.; formal analysis, N.T. and B.A.; funding acquisition, M.F.A.; methodology, M.F.A. and N.T.; project administration, M.H.; resources, B.A.; supervision, M.H.; validation, B.A. and N.T.; writing—original draft, N.T.; and writing—review and editing, M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Deputyship for Research and Innovation, Ministry of Education in Saudi Arabia through the project number 223202.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Can be obtained from author 2 using [email protected].

Acknowledgments

The authors extend their appreciation to the Deputyship for Research and Innovation, Ministry of Education in Saudi Arabia for funding this research work through the project number 223202.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

DDataset
Missing value or special character
D ^ Cleaned dataset
CCategorical variable
C ^ Encoded categorical variable
F i i i th feature in the dataset
F i ^ Scaled feature
D 0 Instances of the majority class
D 1 Instances of the minority class
X i ith input feature
YTarget variable
Score ( X i , Y ) Statistical measure for feature X i and target Y
SSubset of features
Performance ( S ) Performance measure of a model with feature subset S
XFeature matrix
X ^ Transformed feature matrix using PCA

References

  1. Giaquinto, A.N.; Sung, H.; Miller, K.D.; Kramer, J.L.; Newman, L.A.; Minihan, A.; Jemal, A.; Siegel, R.L. Breast cancer statistics. CA Cancer J. Clin. 2022, 72, 524–541. [Google Scholar] [CrossRef] [PubMed]
  2. Ritchie, H.; Spooner, F.; Roser, M. Causes of Death; Our World in Data: Oxford, UK, 2018. [Google Scholar]
  3. Siegel, R.L.; Miller, K.D.; Fuchs, H.E.; Jemal, A. Cancer statistics. CA Cancer J. Clin. 2021, 71, 7–33. [Google Scholar] [CrossRef] [PubMed]
  4. Alwakid, G.; Gouda, W.; Humayun, M.; Jhanjhi, N. Diagnosing Melanomas in Dermoscopy Images Using Deep Learning. Diagnostics 2023, 13, 1815. [Google Scholar] [CrossRef] [PubMed]
  5. Humayun, M.; Khalil, M.I.; Almuayqil, S.N.; Jhanjhi, N.Z. Framework for detecting breast cancer risk presence using deep learning. Electronics 2023, 12, 403. [Google Scholar] [CrossRef]
  6. Lerman, C.; Daly, M.; Sands, C.; Balshem, A.; Lustbader, E.; Heggan, T.; Goldstein, L.; James, J.; Engstrom, P. Mammography adherence and psychological distress among women at risk for breast cancer. JNCI J. Natl. Cancer Inst. 1993, 85, 1074–1080. [Google Scholar] [CrossRef] [PubMed]
  7. Singh, S.J.; Rajaraman, R.; Verlekar, T.T. Breast Cancer Prediction Using Auto-Encoders. In Proceedings of the International Conference on Data Management, Analytics & Innovation, Vellore, India, 19–21 January 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 121–132. [Google Scholar]
  8. Marmot, M.G.; Altman, D.; Cameron, D.; Dewar, J.; Thompson, S.; Wilcox, M. The benefits and harms of breast cancer screening: An independent review. Br. J. Cancer 2013, 108, 2205–2240. [Google Scholar] [CrossRef]
  9. Sun, L.; Legood, R.; Sadique, Z.; dos Santos-Silva, I.; Yang, L. Cost–effectiveness of risk-based breast cancer screening programme, China. Bull. World Health Organ. 2018, 96, 568. [Google Scholar] [CrossRef]
  10. Pomerantz, B.J. Imaging and interventional radiology for cancer management. Surg. Clin. 2020, 100, 499–506. [Google Scholar] [CrossRef]
  11. Giger, M.L. Update on the potential of computer-aided diagnosis for breast cancer. Future Oncol. 2010, 6, 1–4. [Google Scholar] [CrossRef]
  12. Pereira, D.C.; Ramos, R.P.; Do Nascimento, M.Z. Segmentation and detection of breast cancer in mammograms combining wavelet analysis and genetic algorithm. Comput. Methods Programs Biomed. 2014, 114, 88–101. [Google Scholar] [CrossRef]
  13. Übeyli, E.D. Implementing automated diagnostic systems for breast cancer detection. Expert Syst. Appl. 2007, 33, 1054–1062. [Google Scholar] [CrossRef]
  14. Khorasani, E.S. Artificial intelligence: Structures and strategies for complex problem solving. In Scalable Computing: Practice and Experience; Pearson Education: Upper Saddle River, NJ, USA, 2008; Volume 9. [Google Scholar]
  15. Russell, S.J.; Norvig, P. Artificial Intelligence a Modern Approach; University of California: Berkeley, CA, USA, 2010. [Google Scholar]
  16. Koczkodaj, W.; Mansournia, M.; Pedrycz, e.W.; Wolny-Dominiak, A.; Zabrodskii, P.; Strzałka, D.; Armstrong, T.; Zolfaghari, A.; Dębski, M.; Mazurek, J. 1,000,000 cases of COVID-19 outside of China: The date predicted by a simple heuristic. Glob. Epidemiol. 2020, 2, 100023. [Google Scholar] [CrossRef] [PubMed]
  17. Farooq, U.; Tariq, N.; Asim, M.; Baker, T.; Al-Shamma’a, A. Machine learning and the Internet of Things security: Solutions and open challenges. J. Parallel Distrib. Comput. 2022, 162, 89–104. [Google Scholar] [CrossRef]
  18. Erickson, B.J.; Korfiatis, P.; Akkus, Z.; Kline, T.L. Machine learning for medical imaging. Radiographics 2017, 37, 505–515. [Google Scholar] [CrossRef] [PubMed]
  19. Chowdhury, A.; Rosenthal, J.; Waring, J.; Umeton, R. Applying self-supervised learning to medicine: Review of the state of the art and medical implementations. Informatics 2021, 8, 59. [Google Scholar] [CrossRef]
  20. Doersch, C.; Zisserman, A. Multi-task self-supervised visual learning. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2051–2060. [Google Scholar]
  21. Van Steen, M. Distributed systems principles and paradigms. Network 2002, 2, 1. [Google Scholar]
  22. Scollen, S.; Page, A.; Wilson, J. From the data on many, precision medicine for “one”: The case for widespread genomic data sharing. Biomed. Hub 2017, 2, 1–7. [Google Scholar] [CrossRef]
  23. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial Intelligence and Statistics (PMLR), Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  24. Sakib, S.; Yasmin, N.; Tanzeem, A.K.; Shorna, F.; Md. Hasib, K.; Alam, S.B. Breast cancer detection and classification: A comparative analysis using machine learning algorithms. In Proceedings of the Third International Conference on Communication, Computing and Electronics Systems (ICCCES 2021), Coimbatore, India, 28–29 October 2021; Springer: Berlin/Heidelberg, Germany, 2022; pp. 703–717. [Google Scholar]
  25. Li, L.; Xie, N.; Yuan, S. A Federated Learning Framework for Breast Cancer Histopathological Image Classification. Electronics 2022, 11, 3767. [Google Scholar] [CrossRef]
  26. Chowdhury, A.; Kassem, H.; Padoy, N.; Umeton, R.; Karargyris, A. A review of medical federated learning: Applications in oncology and cancer research. In Proceedings of the International MICCAI Brainlesion Workshop, Singapore, 27 September 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 3–24. [Google Scholar]
  27. Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated machine learning: Concept and applications. ACM Trans. Intell. Syst. Technol. (TIST) 2019, 10, 1–19. [Google Scholar] [CrossRef]
  28. Roth, H.R.; Chang, K.; Singh, P.; Neumark, N.; Li, W.; Gupta, V.; Gupta, S.; Qu, L.; Ihsani, A.; Bizzo, B.C.; et al. Federated learning for breast density classification: A real-world implementation. In Proceedings of the Domain Adaptation and Representation Transfer, and Distributed and Collaborative Learning: Second MICCAI Workshop, DART 2020, and First MICCAI Workshop, DCL 2020, Held in Conjunction with MICCAI 2020, Lima, Peru, 4–8 October 2020; Proceedings 2. Springer: Berlin/Heidelberg, Germany, 2020; pp. 181–191. [Google Scholar]
  29. Jiménez-Sánchez, A.; Tardy, M.; Ballester, M.A.G.; Mateus, D.; Piella, G. Memory-aware curriculum federated learning for breast cancer classification. Comput. Methods Programs Biomed. 2023, 229, 107318. [Google Scholar] [CrossRef]
  30. Ma, Z.; Zhang, M.; Liu, J.; Yang, A.; Li, H.; Wang, J.; Hua, D.; Li, M. An assisted diagnosis model for cancer patients based on federated learning. Front. Oncol. 2022, 12, 860532. [Google Scholar] [CrossRef] [PubMed]
  31. Tan, Y.N.; Tinh, V.P.; Lam, P.D.; Nam, N.H.; Khoa, T.A. A Transfer Learning Approach to Breast Cancer Classification in a Federated Learning Framework. IEEE Access 2023, 11, 27462–27476. [Google Scholar] [CrossRef]
  32. Abunasser, B.S.; AL-Hiealy, M.R.J.; Zaqout, I.S.; Abu-Naser, S.S. Breast cancer detection and classification using deep learning Xception algorithm. Int. J. Adv. Comput. Sci. Appl. 2022, 13. [Google Scholar] [CrossRef]
  33. Ogier du Terrail, J.; Leopold, A.; Joly, C.; Béguier, C.; Andreux, M.; Maussion, C.; Schmauch, B.; Tramel, E.W.; Bendjebbar, E.; Zaslavskiy, M.; et al. Federated learning for predicting histological response to neoadjuvant chemotherapy in triple-negative breast cancer. Nat. Med. 2023, 29, 135–146. [Google Scholar] [CrossRef]
  34. Dheeru, D.; Karra Taniskidou, E. UCI Machine Learning Repository; School of Information and Computer Sciences, University of California: Irvine, CA, USA, 2017. [Google Scholar]
  35. Khan, N.M.; Abraham, N.; Hon, M.; Guan, L. Machine learning on biomedical images: Interactive learning, transfer learning, class imbalance, and beyond. In Proceedings of the 2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), San Jose, CA, USA, 28–30 March 2019; pp. 85–90. [Google Scholar]
  36. Sahu, P.; Giri, P.; Sunkara, R.; Sundararajan, R. Extraction of key features and enhanced prediction framework of breast cancer occurrence. In Proceedings of the 2022 6th International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 28–30 April 2022; pp. 1679–1685. [Google Scholar]
  37. Connor, S.; Khoshgoftaar, T.M.; Borko, F. Text data augmentation for deep learning. J. Big Data 2021, 8, 101. [Google Scholar]
  38. Goto, M.; Nagayama, Y.; Sakabe, D.; Emoto, T.; Kidoh, M.; Oda, S.; Nakaura, T.; Taguchi, N.; Funama, Y.; Takada, S.; et al. Lung-optimized deep-learning-based reconstruction for ultralow-dose CT. Acad. Radiol. 2023, 30, 431–440. [Google Scholar] [CrossRef] [PubMed]
  39. Sowjanya, A.M.; Mrudula, O. Effective treatment of imbalanced datasets in health care using modified SMOTE coupled with stacked deep learning algorithms. Appl. Nanosci. 2023, 13, 1829–1840. [Google Scholar] [CrossRef]
  40. Kumbhar, P.; Mali, M. A survey on feature selection techniques and classification algorithms for efficient text classification. Int. J. Sci. Res. 2016, 5, 9. [Google Scholar]
  41. Kumar, A.; Kumar, N.; Kuriakose, J.; Kumar, Y. A review of deep learning-based approaches for detection and diagnosis of diverse classes of drugs. Arch. Comput. Methods Eng. 2023, 30, 3867–3889. [Google Scholar] [CrossRef]
  42. Mao, A.; Mohri, M.; Zhong, Y. Cross-entropy loss functions: Theoretical analysis and applications. arXiv 2023, arXiv:2304.07288. [Google Scholar]
  43. Butt, M.; Tariq, N.; Ashraf, M.; Alsagri, H.S.; Moqurrab, S.A.; Alhakbani, H.A.A.; Alduraywish, Y.A. A Fog-Based Privacy-Preserving Federated Learning System for Smart Healthcare Applications. Electronics 2023, 12, 4074. [Google Scholar] [CrossRef]
  44. Ali, A.; Shaukat, S.; Tayyab, M.; Khan, M.A.; Khan, J.S.; Arshad; Ahmad, J. Network intrusion detection leveraging machine learning and feature selection. In Proceedings of the 2020 IEEE 17th International Conference on Smart Communities: Improving Quality of Life Using ICT, IoT and AI (HONET), Charlotte, NC, USA, 14–16 December 2020; pp. 49–53. [Google Scholar]
  45. Shaukat, S.; Ali, A.; Batool, A.; Alqahtani, F.; Khan, J.S.; Ahmad, J. Intrusion detection and attack classification leveraging machine learning technique. In Proceedings of the 2020 14th International Conference on Innovations in Information Technology (IIT), Al Ain, United Arab Emirates, 17–18 November 2020; pp. 198–202. [Google Scholar]
  46. Brownlee, J. How to Calculate Precision, Recall, and F-Measure for Imbalanced Classification; Machine Learning Mastery: Vermont, UK, 2020. [Google Scholar]
  47. Alqhtani, S.M. BreastCNN: A novel layer-based convolutional neural network for breast cancer diagnosis in DMR-thermogram images. Appl. Artif. Intell. 2022, 36, 2067631. [Google Scholar] [CrossRef]
Figure 1. Average ratio among breast cancer cases and detection mechanisms used.
Figure 1. Average ratio among breast cancer cases and detection mechanisms used.
Healthcare 11 03185 g001
Figure 2. A deep neural network.
Figure 2. A deep neural network.
Healthcare 11 03185 g002
Figure 3. Federated learning.
Figure 3. Federated learning.
Healthcare 11 03185 g003
Figure 4. Accuracy comparison in the first iteration. (a) Accuracy for first model; (b) Accuracy for second model.
Figure 4. Accuracy comparison in the first iteration. (a) Accuracy for first model; (b) Accuracy for second model.
Healthcare 11 03185 g004
Figure 5. Loss comparison in the first iteration. (a) Loss for Model 1; (b) loss for Model 2.
Figure 5. Loss comparison in the first iteration. (a) Loss for Model 1; (b) loss for Model 2.
Healthcare 11 03185 g005
Figure 6. Accuracy comparison in the second iteration. (a) Accuracy for Model 1; (b) accuracy for Model 2.
Figure 6. Accuracy comparison in the second iteration. (a) Accuracy for Model 1; (b) accuracy for Model 2.
Healthcare 11 03185 g006
Figure 7. Loss comparison in the second iteration. (a) Loss for Model 1; (b) loss for Model 2.
Figure 7. Loss comparison in the second iteration. (a) Loss for Model 1; (b) loss for Model 2.
Healthcare 11 03185 g007
Figure 8. Accuracy comparison in the third iteration. (a) Accuracy for Model 1; (b) accuracy for Model 2.
Figure 8. Accuracy comparison in the third iteration. (a) Accuracy for Model 1; (b) accuracy for Model 2.
Healthcare 11 03185 g008
Figure 9. Loss comparison in the third iteration. (a) Loss for Model 1; (b) loss for Model 2.
Figure 9. Loss comparison in the third iteration. (a) Loss for Model 1; (b) loss for Model 2.
Healthcare 11 03185 g009
Figure 10. The overall results.
Figure 10. The overall results.
Healthcare 11 03185 g010
Table 1. Comparison of state-of-the-art studies on breast cancer detection and federated learning.
Table 1. Comparison of state-of-the-art studies on breast cancer detection and federated learning.
Ref.MethodologyDataset UsedKey FindingsLimitations
[7]Auto-encoder for unsupervised breast cancer identificationBreast Cancer Wisconsin (Diagnostic) DatasetAuto-encoder achieved 98.4% precision and recallNo comparison with other unsupervised methods, limited discussion on scalability
[30]FL framework with CNNMultiple cancer datasetsAchieved above 90% accuracy for five cancer typesLimited comparison with MLP and data imbalance not addressed
[31]FL with transfer learning, SMOTE, and FeAvg-CNN + MobileNet modelMammography datasetsProposed method showed superior classification performanceNo exploration of federated privacy concerns
[25]FL for histopathology-based breast cancer classificationBreakHis datasetFL model performed on par with centralized learningLimited discussion on scalability to larger datasets
[32]DL-based Xception modelKaggle datasetModel exhibited proficiency in breast cancer detectionIncreased processing time not discussed
[33]FL for predicting triple-negative breast cancerConfidential patient dataFL enhanced the effectiveness of detecting triple cancerSpecific FL implementation details not provided, potential variability in data quality
Table 2. Dataset details.
Table 2. Dataset details.
AttributeDescription
Number of InstitutionsUniversity of Wisconsin Hospital in Madison, Madison, WI, USA
Data TypesDemographic, Clinical, Pathological, Outcome
Data FormatsNumerical, Categorical, Images
Patient Data VariationsDemographic, Clinical, Pathological, Outcome
Table 3. Patient data details.
Table 3. Patient data details.
Data TypeSpecific Data ElementPatient Data Variations
DemographicAgeRange from 25 to 97 years old
SexMale or female
RaceWhite, Black, Asian, or other
EthnicityHispanic, Non-Hispanic
ClinicalMedical HistoryPresence or absence of various medical conditions
MedicationsTypes and dosages of medications
Laboratory ResultsVariations in blood cell counts, hormone levels, and other biomarkers
Imaging StudiesDifferences in tumor size, shape, and texture on mammograms, ultrasounds, or other imaging modalities
PathologicalTumor PathologyType of breast cancer (e.g., ductal carcinoma, lobular carcinoma)
Molecular ProfilingGenetic alterations in tumor cells
OutcomeSurvivalTime from diagnosis to death or censoring
RecurrencePresence or absence of tumor recurrence after treatment
Treatment ResponseTumor shrinkage or growth in response to treatment
Table 4. Features of BCW (diagnostic) dataset.
Table 4. Features of BCW (diagnostic) dataset.
FeatureDescription
idIdentifier
concave_points_worstWorst concave points on the perimeter of nuclei
diagnosisDiagnosis (B = benign, M = malignant)
symmetry_worstWorst symmetry of nuclei
radius_meanMean radius of nuclei
fractal_dimension_worstWorst fractal dimension of nuclei
texture_meanMean texture of nuclei
texture_seStandard error of texture
perimeter_meanMean perimeter of nuclei
perimeter_seStandard error of perimeter
area_meanMean area of nuclei
area_seStandard error of area
smoothness_meanMean smoothness (1–3) of nuclei
smoothness_worstWorst smoothness (1–3) of nuclei
compactness_meanMean compactness of nuclei
compactness_worstWorst compactness of nuclei
concavity_meanMean concavity of nuclei
concavity_worstWorst concavity of nuclei
concave_points_meanMean of concave points on the perimeter of nuclei
concave_points_worstWorst of concave points on the perimeter of nuclei
symmetry_meanMean symmetry of nuclei
symmetry_worstWorst symmetry of nuclei
fractal_dimension_meanMean fractal dimension of nuclei
fractal_dimension_worstWorst fractal dimension of nuclei
radius_seStandard error of radius
Table 5. Dataset preprocessing summary.
Table 5. Dataset preprocessing summary.
Preprocessing StepDataset Distribution
Original dataset distribution569
Benign (B)357
Malignant (M)212
After outliers removal538
Benign (B)340
Malignant (M)198
Data Augmentation
Adding noise (0 to 0.01)Total = 5742
Benign (B)3212
Malignant (M)2530
SMOTE for class balanceTotal = 6424
Benign (B)3212
Malignant (M)3212
Table 6. Interquartile range (IQR).
Table 6. Interquartile range (IQR).
FeatureIQRFeatureIQR
radius_mean5.42texture_mean5.29
perimeter_mean36.72area_mean504.1
smoothness_mean0.02compactness_mean0.07
concavity_mean0.12concave points_mean0.06
symmetry_mean0.03fractal_dimension_mean0.01
radius_se0.3texture_se0.58
perimeter_se2.03area_se39.02
smoothness_se0.0compactness_se0.02
concavity_se0.03concave points_se0.01
symmetry_se0.01fractal_dimension_se0.0
radius_worst7.45texture_worst8.56
perimeter_worst52.31area_worst766.36
smoothness_worst0.03compactness_worst0.21
concavity_worst0.3concave points_worst0.11
symmetry_worst0.07fractal_dimension_worst0.02
Table 7. Dataset before data augmentation.
Table 7. Dataset before data augmentation.
FeaturesMeanStdMin25%50%75%Max
radius_mean14.463.16.9812.1913.7316.7523.27
texture_mean19.073.749.7116.3918.9221.4729.81
perimeter_mean93.921.1243.7978.288.48109.72152.1
area_mean675.59294.56143.5458.02581.65884.781686.0
smoothness_mean0.10.010.060.090.10.10.13
compactness_mean0.10.040.020.070.090.130.23
concavity_mean0.080.060.00.030.060.120.28
concave points_mean0.050.030.00.020.040.080.16
symmetry_mean0.180.020.120.160.180.190.25
fractal_dimension_mean0.060.010.050.060.060.060.08
radius_se0.370.180.110.230.320.470.97
texture_se1.090.380.360.821.041.342.24
perimeter_se2.591.240.761.592.283.226.79
area_se36.3725.46.818.1526.3847.48116.29
smoothness_se0.010.00.00.010.010.010.01
compactness_se0.020.010.00.010.020.030.06
concavity_se0.030.010.00.020.020.040.08
concave points_se0.010.00.00.010.010.010.02
symmetry_se0.020.010.010.010.020.020.03
fractal_dimension_se0.00.00.00.00.00.00.01
radius_worst16.724.277.9313.3515.6419.9328.01
texture_worst25.665.4712.0221.7125.6229.8842.65
perimeter_worst109.7128.9950.4186.77102.4131.42187.33
area_worst911.11472.97185.2549.28749.51227.852405.48
smoothness_worst0.130.020.080.120.130.150.19
compactness_worst0.250.120.030.160.220.320.66
concavity_worst0.270.170.00.130.240.390.77
concave points_worst0.120.060.00.070.110.160.26
symmetry_worst0.290.050.180.260.290.320.43
fractal_dimension_worst0.080.010.060.070.080.090.12
diagnosis0.440.50.00.00.01.01.0
Table 8. Dataset after data augmentation.
Table 8. Dataset after data augmentation.
FeatureMeanStdMin25%50%75%Max
radius_mean14.463.16.9512.1913.7216.7823.28
texture_mean19.063.739.716.3818.9221.4729.82
perimeter_mean93.921.143.7878.1888.48109.73152.12
area_mean675.59294.3143.49457.9581.65886.31686.02
smoothness_mean0.10.010.050.090.10.110.15
compactness_mean0.10.040.00.070.090.130.25
concavity_mean0.080.06−0.010.030.070.120.31
concave points_mean0.050.03−0.010.020.040.080.18
symmetry_mean0.170.030.090.160.170.190.25
fractal_dimension_mean0.060.010.040.060.060.070.11
radius_se0.370.180.10.230.320.470.98
texture_se1.090.380.350.811.041.342.25
perimeter_se2.591.240.741.582.273.226.81
area_se36.3725.386.7818.1526.3947.59116.31
smoothness_se0.010.01−0.010.00.010.010.03
compactness_se0.020.02−0.030.010.020.030.08
concavity_se0.030.02−0.010.010.020.040.1
concave points_se0.010.01−0.010.010.010.020.06
symmetry_se0.020.01−0.010.010.020.020.05
fractal_dimension_se0.00.01−0.02−0.010.00.010.03
radius_worst16.724.277.9113.3515.6419.9328.04
texture_worst25.665.4712.021.725.6229.8942.66
perimeter_worst109.7128.9750.3986.76102.4131.44187.34
area_worst911.11472.56185.19549.1749.51228.02405.5
smoothness_worst0.130.020.070.120.130.150.2
compactness_worst0.250.120.010.160.220.330.67
concavity_worst0.270.17−0.010.130.250.380.78
concave points_worst0.110.06−0.020.070.110.160.28
symmetry_worst0.290.050.160.250.280.320.44
fractal_dimension_worst0.080.020.030.070.080.090.13
diagnosis0.440.50.00.00.01.01.0
Table 9. Classification outcomes for breast cancer prediction.
Table 9. Classification outcomes for breast cancer prediction.
OutcomeDescription
T R p s Model correctly predicts breast cancer in a patient with breast cancer
T R n g Model correctly predicts no breast cancer in a patient without breast cancer
F L p s Model incorrectly predicts breast cancer in a patient without breast cancer
F L n g Model incorrectly predicts no breast cancer in a patient with breast cancer
Table 10. Global model performance metrics for the first iteration.
Table 10. Global model performance metrics for the first iteration.
MetricValue
True Positive Rate (Sensitivity)97.68%
False Positive Rate3.23%
True Negative Rate (Specificity)96.77%
False Negative Rate2.32%
Accuracy97.26%
Precision96.43%
F1-Score97.05%
Table 11. Global model performance metrics for the second iteration.
Table 11. Global model performance metrics for the second iteration.
MetricValue
True Positive Rate (Sensitivity)98.00%
False Positive Rate3.23%
True Negative Rate (Specificity)96.77%
False Negative Rate1.89%
Accuracy97.54%
Precision96.49%
F1-Score97.24%
Table 12. Global model performance metrics for the third iteration.
Table 12. Global model performance metrics for the third iteration.
MetricValue
True Positive Rate (Sensitivity)98.00%
False Positive Rate3.23%
True Negative Rate (Specificity)96.77%
False Negative Rate1.89%
Accuracy97.54%
Precision96.49%
F1-Score97.24%
Table 13. Comparison of breast cancer detection results.
Table 13. Comparison of breast cancer detection results.
StudyModelTechniquesResults
Proposed StudyDNN (FL)Federated LearningAccuracy: 97.5%,
Precision: 96%, Recall: 98.0%,
and F1-Score: 97%
[7]Auto-encoderUnsupervisedPrecision: 98.4%
[30]Hybrid FL with CNNTransfer LearningAccuracy: Above 90%
[36]DTCFeature SelectionAccuracy: 94
[25]FCNMachine LearningAccuracy: 96.4%,
Sensitivity: 97.5%, and
Specificity: 97.8%.
[47]CNNMachine LearningAccuracy 99.7%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Almufareh, M.F.; Tariq, N.; Humayun, M.; Almas, B. A Federated Learning Approach to Breast Cancer Prediction in a Collaborative Learning Framework. Healthcare 2023, 11, 3185. https://doi.org/10.3390/healthcare11243185

AMA Style

Almufareh MF, Tariq N, Humayun M, Almas B. A Federated Learning Approach to Breast Cancer Prediction in a Collaborative Learning Framework. Healthcare. 2023; 11(24):3185. https://doi.org/10.3390/healthcare11243185

Chicago/Turabian Style

Almufareh, Maram Fahaad, Noshina Tariq, Mamoona Humayun, and Bushra Almas. 2023. "A Federated Learning Approach to Breast Cancer Prediction in a Collaborative Learning Framework" Healthcare 11, no. 24: 3185. https://doi.org/10.3390/healthcare11243185

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop