Next Article in Journal
Dual Stream Long Short-Term Memory Feature Fusion Classifier for Surface Electromyography Gesture Recognition
Previous Article in Journal
Autonomous Image-Based Corrosion Detection in Steel Structures Using Deep Learning
Previous Article in Special Issue
Secure Aggregation Protocol Based on DC-Nets and Secret Sharing for Decentralized Federated Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing IoT Healthcare with Federated Learning and Variational Autoencoder

by
Dost Muhammad Saqib Bhatti
and
Bong Jun Choi
*
School of Computer Science and Engineering, Soongsil University, Seoul 06978, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(11), 3632; https://doi.org/10.3390/s24113632
Submission received: 22 April 2024 / Revised: 24 May 2024 / Accepted: 28 May 2024 / Published: 4 June 2024
(This article belongs to the Special Issue Federated and Distributed Learning in IoT)

Abstract

:
The growth of IoT healthcare is aimed at providing efficient services to patients by utilizing data from local hospitals. However, privacy concerns can impede data sharing among third parties. Federated learning offers a solution by enabling the training of neural networks while maintaining the privacy of the data. To integrate federated learning into IoT healthcare, hospitals must be part of the network to jointly train a global central model on the server. Local hospitals can train the global model using their patient datasets and send the trained localized models to the server. These localized models are then aggregated to enhance the global model training process. The aggregation of local models dramatically influences the performance of global training, mainly due to the heterogeneous nature of patient data. Existing solutions to address this issue are iterative, slow, and susceptible to convergence. We propose two novel approaches that form groups efficiently and assign the aggregation weightage considering essential parameters vital for global training. Specifically, our method utilizes an autoencoder to extract features and learn the divergence between the latent representations of patient data to form groups, facilitating more efficient handling of heterogeneity. Additionally, we propose another novel aggregation process that utilizes several factors, including extracted features of patient data, to maximize performance further. Our proposed approaches for group formation and aggregation weighting outperform existing conventional methods. Notably, significant results are obtained, one of which shows that our proposed method achieves 20.8% higher accuracy and 7% lower loss reduction compared to the conventional methods.

1. Introduction

The development of the Internet of Things (IoT) is driving a significant increase in smart and connected devices, opening up opportunities for new applications and services [1]. One such application area is healthcare, where the development of IoT technology aims to provide efficient services to patients [2]. Many complex problems in the healthcare sector require the use of modern technologies in both the detection and the classification processes [3]. The diversity of data generated by different medical centres plays a crucial role in the detection of health problems using neural networks. However, due to privacy concerns, these centres are often reluctant to share their patient data with third parties. Nevertheless, in order to effectively identify patient symptoms, it is essential to make use of the extensive data collected by medical centres.
Federated learning offers a solution for training neural network models while safeguarding patients’ data privacy. This approach efficiently addresses the issue above by enabling intelligence development at the network’s edge [4]. The implementation of Federated Learning aims to safeguard user data, ensure the development of secure and precise models, and mitigate the risks associated with single points of failure in centralized models [5]. This approach prioritizes the protection of user privacy by enabling model training to occur locally on user devices without the need to share raw data with a central server. By distributing the learning process across multiple devices or edge servers, federated learning enhances data security and minimizes the potential impact of data breaches or system vulnerabilities. Additionally, federated learning facilitates the creation of personalized and context-aware models by leveraging locally collected data, ultimately leading to more accurate and tailored predictions or recommendations for individual users. Furthermore, by decentralizing model training and inference, federated learning reduces the risk of system-wide failures or disruptions that may occur in traditional centralized approaches, thereby ensuring greater robustness and reliability in large-scale machine learning systems [5]. With federated learning, medical centers may deploy neural networks to train the local networks by utilizing patient data. Depending on the network’s requirements, these neural networks can take various forms, such as convolutional neural networks (CNNs), deep neural networks (DNNs), or other types [6].
The ultimate goal is to develop a well-trained global model in any central hospital using diverse local patient data from different medical centers without direct access to it [7]. Each medical center may have its unique characteristics, including the volume and variance of the patient data. Consequently, the collective network comprising all centres and hospitals is characterized by a closely diverse structure, where the datasets of individual medical centres/hospitals are assumed to be non-independent and non-identically distributed (non-IID) [8,9].
The goal is to train a high-quality global model at any central hospital using the data of local patients of different hospitals or medical centers. Henceforth, a server refers to any central hospital overseeing the global training process. In contrast, the term client refers to the local hospital or medical center actively participating in the federated learning framework. Furthermore, the term local data refers to the data of patients within a hospital. First, the global model is distributed to all clients on the network, and each client improves it using its local data. The server collects the updated local models in each round, aggregates them to refine the global model, and sends it back to the clients. This iterative process is repeated many times until the global model converges. However, due to the non-IID nature of the local patient data, the clients’ updated local models vary significantly in each round [8]. This heterogeneity in the local patient data sets can lead to significant differences in the local model weights, posing a challenge for convergence to the appropriate global minimum. Hence, it is worth noting that the performance of the global training process is significantly affected by the aggregation of very different local models from clients in a heterogeneous environment.
To address this issue, our proposed approach employs autoencoders. It transforms the data into latent space to extract the features of the patient data and captures the Jensen–Shanon Divergence (JSD) among the extracted features of the patient data, thus enabling a more streamlined federated aggregation of heterogeneous data. In conventional methods, the server aggregates the local models of the clients without considering the characteristics of the client models, which can lead to a degradation of the global training performance if the locally trained models are learned on data that lack essential features for learning. On the other hand, in our proposed approach, each local hospital condenses its patients’ data into a latent representation using an encoder and transmits the signature to the server. The server then captures the JSD between hospitals and groups them for more efficient federated learning. Each group, a combination of local hospitals, is assigned a particular aggregation weight based on the characteristics of its corresponding latent representation. Our contribution lies in using the variational autoencoder (VAE) to learn the features of patient data, creating various combinations, and assigning the aggregation weight to both local medical centers and combined groups of these centers to facilitate efficient network training at the central hospital.
The paper is organized as follows: Section 2 discusses the related work. Section 3 explains the system model. Section 4 presents the proposed model. The simulation results are presented in Section 5. Finally, in Section 6, the paper is concluded.

2. Related Work

This section outlines some pertinent work related to federated learning. The primary goal of federated learning is to attain an enhanced, accurately trained global model at the server with minimized loss when predicting unfamiliar data. The literature initially focuses on enhancing performance through grouping in federated learning. Subsequently, the analysis delves into the repercussions of irregular and heterogeneous data distribution across clients in federated learning and reviews proposed solutions in the literature. Finally, the exploration extends to some aggregation methods introduced by the researchers.

2.1. Grouping in Federated Learning

Various research works have explored grouping methods, making them widely available in the literature. The authors of [10] have utilized fuzzy c-means to form clusters. The method above is commonly applied in pattern recognition, where each data point is assigned a degree of membership about the center of a cluster. The segregation techniques described in [11] utilize iterative splitting and cosine similarity as a measure for model update comparisons. However, their recursive nature can result in high computing and communication overheads, which might pose a bottleneck in settings of considerable scale. In article [12], two iterative methods are proposed for clustering clients and training sub-models for each cluster. Nevertheless, these methods entail multiple communication rounds until the clusters are formed, and the formation requires sending cluster sub-models to all participating clients in each iteration, which is costly in terms of communication.
In [13], a distance-based hierarchical clustering approach is utilized to cluster clients’ models based on the similarity between the local updates. Similarly, in [14], the client and server model distance determines the association between clients. Hierarchical federated learning is proposed in [15,16], where a fixed number of clients is assumed to be in each cluster, and the clients in each cluster report to their respective edge node. FedGroup [17] uses the Euclidean distance of decomposed cosine similarity to identify the similarity between clients and form groups. In [18], a method is proposed where the network learns the global model parameters to shape the clusters.
Another client grouping approach utilizes exogenous information about clients’ data, as proposed in [19,20]. However, this method requires direct access to clients’ data, making it unsuitable for general cases. Moreover, another hierarchical approach, called Hierarchical-Local-QSGD, is introduced in [21], where clients forward quantized model information to a fixed edge node after several steps of local updates. Furthermore, HFEL [22] is another hierarchical-based federated learning method that partially migrates model aggregation to geographically fixed edge nodes. In addition, Astraea [23] proposes a self-balancing framework to alleviate imbalances among fixed-edge nodes. Also, previous research by Bhatti et al. [24] has utilized affinity propagation (AP) machine learning to form clusters. This involves nodes exchanging local messages with their neighbors until a cluster head with the highest class level is chosen, resulting in a corresponding clustering configuration. Dennis et al. [25] introduce a one-shot federated clustering technique that projects client data onto a chosen subspace and employing iterative Lloyd’s k-means clustering [26] to communicate the results to the server for cluster assignment. However, their study needs to explore the impact of their clustering on subsequent federated learning applications, and linear subspace decomposition proves less practical compared to the non-linear client-side autoencoders implemented in our methodology.
In contrast to these approaches, our proposed method exploits the features of patient data to distill it into its distinctive signature encoding for grouping and assigning aggregation weightage, thus avoiding the use of model parameters as an indirect representation of the client data structures. Our proposed VAE-based method on local data interconnection is more efficient than the above methodologies, which extracts and learns the feature of patient data to group them and assign the aggregation weightage accordingly.

2.2. Federated Learning under Heterogeneous Environment

Since federated learning involves different types of clients, it is unrealistic to achieve an even distribution of data between them, i.e., IID [27]. Therefore, it is essential to take into account the heterogeneity of the data and the uneven volume of samples across clients. In ref. [28], it is reported that the participation of clients with varied data distribution in global training influences the accuracy of the overall model. Data heterogeneity can lead to a significant diversity in the weights of local and global models. The issue of statistical heterogeneity in federated learning has been addressed through various approaches in the literature. Personalized federated learning creates customized models for clients by applying multi-task learning techniques or meta-learning [29]. Moreover, Yao et al. [30] propose to create a mini-dataset containing a limited number of data samples from different label classes, which can be shared with all clients in a network. However, this approach compromises the privacy of clients, as their data are used to create the dataset. To address this issue, Zhang et al. [31] propose a technique that refines the global model before consolidating the local models. Alternatively, Tian et al. [32] introduce FedProx, which restricts local updates by adding a proximal term to prevent the local model from deviating significantly from the initial state of the global model. Finally, the authors of [33] investigate the personalized variant of federated learning and focuse on finding an initial shared model that fresh clients can readily adopt.
The effectiveness of federated learning predominantly relies on how the server aggregates the weights of the local models [34]. McMahan et al. [35] propose a well-known aggregation method for federated learning called Federated Averaging (FedAvg). In FedAvg, the global model is updated by taking a weighted average of the local models.
Additionally, to enhance comprehension of the distinctions between conventional methods and our proposed approach and to elucidate the rationale behind the necessity of our algorithm, we summarize the existing work in a Table 1.
Table 1 offers a comparative analysis of various methods in federated learning, assessing each based on framework, clustering capability, heterogeneity handling, aggregation approach, and major contributions. To begin with, FCM introduces a clustering algorithm aimed at minimizing energy wastage in networks. Moreover, Astraea proposes a self-balancing federated learning framework to tackle data imbalances. Furthermore, AP cluster formation presents a machine learning-based approach for energy-efficient clustering. In addition, HFEL migrates model aggregation to edge nodes and introduces a resource scheduling algorithm. Distance-based clustering is proposed in [13], which offers hierarchical clustering for client separation. Furthermore, k-FED addresses communication costs and device failures in federated settings. Additionally, Hier-Local-QSGD tightens convergence bounds and selects aggregation intervals based on features. Lastly, the Hyper-parameter-based method introduces a consumer clustering technique without sharing confidential data. All the aforementioned methods concentrate solely on either clustering, heterogeneity, or aggregation individually. None of the articles put forth a comprehensive model that endeavors to generate improved clusters based on heterogeneity while simultaneously executing robust aggregation. Our proposed method prioritizes the creation of superior clusters by delving into the heterogeneity of the users’ data while also executing aggregation concurrently. It utilizes variational auto-encoder to optimize aggregation based on feature extraction and data heterogeneity.

2.3. Contributions

The primary objective of this paper is to propose an efficient grouping and aggregation approach that effectively incorporates key parameters, including extracted features of local patient data. The key contributions of our approach are as follows:
  • A performance-efficient VAE-based method is proposed that extracts and learns patients’ features for forming groups and accounts for data heterogeneity.
  • The global model is updated while reducing the prediction loss and enhancing classification accuracy. It is evaluated through an extensive set of simulations using a federated learning simulator, with varying scenarios of data heterogeneity, and shows that it outperforms conventional methods.
  • The proposed approach takes into account the characteristics of patient data and performs aggregation in a way that achieves the highest accuracy and the lowest loss, making it robust compared to conventional methods. This is accomplished through accurately aggregating local models trained by medical centers and hospitals, considering various parameters such as their latent spaces, data volume, and data variance.

3. System Model

Let us consider a network with K medical centers deployed with neural networks, communicating with a central hospital, which is called the server, for global training. Each medical center or hospital, which is known as a client, aims to update its local model using the patients’ data available to it, and the server aims to update the global model by aggregating the updated local models from all clients. If we denote the kth client in the network as c l k , it holds D k data comprising n k samples and l k classes of labels. The clients must generate a latent representation of their local patient’s data using autoencoders, denoted by δ k , which is then shared with the server. Upon receiving the latent representations from the clients, the server evaluates the JSD between them to format the client groups. Let us assume that M groups are generated as shown in Figure 1. Once the clients are segregated based on the JSD among them using their latent representations, the server initiates training by distributing the trained network at round t as ω t for further refinement using the end user’s data.
In other words, Figure 1 presents our proposed method of clustering clients based on the divergence of their data, utilizing latent space representations generated by VAE. The pictorial representation is given on the left side and an explanatory flowchart is on the right side. The procedure starts with initializing the VAE model, which is used to encode the data from each client into latent representations. This transformation captures the underlying patterns and features of the data in a lower-dimensional space, which is crucial for efficient analysis. For each client c l k in all clients K, latent representation δ k is generated using the VAE and forwarded to the server. The next step involves computing divergence D j s ( δ k | δ j ) between the latent embeddings of each pair of clients c l k and c l j . This step is essential for understanding the similarities and differences between the data distributions of different clients. The final step is the clustering decision. Based on the computed divergence, a decision is made to group the clients.
After the server shares the global model at round t, the clients commence their local training. As each local model is updated, it is sent to the server as ω k t + 1 . The clients are previously classified based on their latent embeddings. The server performs two types of aggregations: combination aggregation and global aggregation. The combination aggregation combines local models from clients within a specific group, generated based on the client’s latent spaces. The global aggregation aggregates all combination models to update the global model at each round. The proposed method’s novelty lies in grouping local hospitals utilizing the latent representation of patient data and the aggregation weightage assigned to both combination aggregation and global aggregation. The proposed method is discussed in detail in the following Section 4. Moreover, the list of some important symbols and abbreviations is given below.
  • D k : kth client’s dataset;
  • n k : Total number of samples in patients’ data in the kth medical center;
  • l k : Number of classes of labels in the kth medical center’s patients’ data;
  • δ k : Latent representation of the kth medical center’s patients’ data;
  • ω t : Global model at round t;
  • ω k t + 1 : The kth medical center’s local model at round t + 1 ;
  • q ϕ ( δ k | D k ) : Encoder network that maps input data D k to latent variable δ k ;
  • p θ ( δ k | D k ) : The decoder network that maps latent variable δ k ;
  • D j s ( δ k δ m ) : The divergence between latent spaces of client k and m;
  • D j s t h : Divergence threshold;
  • C m : The mth group;
  • C m k : The kth client of the mth group;
  • ω m k t : Local model of the kth client of the mth group at round t;
  • η m k : Learning rate of the kth client of the mth group;
  • ð m k t : Gradients of the kth client of the mth group;
  • ω C m t + 1 : Clustered trained model of the mth group at round t + 1 ;
  • δ m : Latent space of the mth group;
  • δ : Latent space of the whole network;
  • δ m k : Latent space of the kth client of the mth group;
  • K L ( q ϕ ( δ k | D k ) | | p ( δ k ) ) : The Kullback–Leibler divergence between learned latent distribution q ϕ ( δ k | D k ) and prior distribution p ( δ k ) .

4. Proposed Model

4.1. Latent Representation Transformation

The VAE represents a generative model that integrates a deep neural network with a probabilistic framework [36]. It consists of an encoder network responsible for transforming input data into a latent space. The primary objective of the VAE is to acquire the latent representation of the input data, aiming to optimize the likelihood of the data while minimizing the difference between the learned latent distribution and a predefined prior distribution.
The objective function of the kth client for transforming the data into latent representation using VAE can be expressed as
δ k = E δ k q ϕ ( δ k | D k ) [ log p θ ( D k | δ k ) ] + K L ( q ϕ ( δ k | D k ) p ( δ k ) ) ,
where q ϕ ( δ k | D k ) is the encoder network that maps input data D k to latent variable δ k with parameters ϕ , p θ ( δ k | D k ) is the decoder network that maps latent variable δ k back to input data D k with parameters θ , and KL ( q ϕ ( δ k | D k ) | | p ( δ k ) ) is the Kullback–Leibler divergence between learned latent distribution q ϕ ( δ k | D k ) and prior distribution p ( δ k ) . The first term represents the reconstruction loss, i.e., the anticipated negative log-likelihood. This term encourages the model to generate data similar to the observed data. On the other hand, the second term encourages the learned latent distribution to be close to the prior distribution.
To be more specific, the objective is to establish a lower bound on the log likelihood of the data. Maximizing δ k during training effectively optimizes this lower bound on the true log likelihood. By optimizing δ k , the VAE aims to achieve two main objectives simultaneously: accurately reconstructing the input data and regularizing the learned latent space representations. This dual optimization process facilitates the discovery of meaningful latent representations of the input data. These representations capture essential features and patterns in the data, enabling the VAE to learn rich and informative representations. Subsequently, these representations are then utilized for clustering the clients and global model aggregation, contributing to the versatility and effectiveness of the VAE in learning useful representations from complex data.

4.2. Group Formation

Once the latent representation of each client is generated, the divergence between them is computed using JSD to segment the client models into different groups based on their extracted features. Jensen–Shannon divergence is a metric for quantifying the divergence between two probability distributions. Hence, the divergence between two latent representations is given as
D j s ( δ k δ j ) = 1 2 ( K L ( δ k δ k + δ j 2 ) + K L ( δ j δ k + δ j 2 ) ) ,
where K L ( δ k δ k + δ j 2 ) represents the Kullback–Leibler divergence from δ k to the average of both distributions, expressed as
K L ( δ k δ k + δ j 2 ) = i δ k ( i ) × l o g ( δ k ( i ) ( δ k ( i ) + δ j ( i ) ) / 2 ) .
Summation is carried out over all possible values of i, while δ k ( i ) and δ j ( i ) denote the probabilities of i within distributions δ k and δ j of clients k and j, respectively. The latent embeddings of clients, generated through VAE, serve as a basis for feature extraction from their data. Consequently, clients are grouped into distinct categories according to the similarity observed in the features extracted from their patients’ data.

4.3. Client Update

Once the groups are formed, the global model is sent, and local training is initialized. Subsequently, the updated trained models are shared with the server. Furthermore, we provide a pictorial representation of our method, shown in Figure 2. The contribution stems from the equations generated by the central hospital and the medical centers. In addition, the pseudo-code of the proposed method is given in Algorithm 1. The process begins by initializing clients and transforming their data into latent representations using VAE from Lines 1 to 5. Subsequently, from Lines 6 to 15, the divergence between the latent representations of each client is computed using JSD and compared against a predefined threshold, D j s t h . If the divergence between the extracted features of two clients is less than the threshold, they are grouped into the same group, denoted as C m . After forming the groups, the aggregation process commences. The weightage for group aggregation is determined, with aggregation weightage assigned according to the equation given in Line 24 of Algorithm 1. Following the completion of group aggregation, global aggregation proceeds, allocating aggregation weightage according to the equation in Line 26.
Algorithm 1 Proposed VAE based local model segmentation
  1: A. Initialization of Autoencoding:
     c l i e n t s k = 1 , 2 , , K .
  2: for  k = 1 , 2 , , K  do
  3:     Latent representation using VAE of the kth client, δ k
  4:      δ k = E δ k q ϕ ( δ k | D k ) [ log p θ ( D k | δ k ) ] + K L ( q ϕ ( δ k | D k ) p ( δ k ) )
  5: end for
  6: for  k = 1 , 2 , , K  do
  7:     for  j = 1 , 2 , 3 , , K  do
  8:         Divergence evaluation using latent embedding of clients using Jensen–Shannon Divergence (JSD)
  9:          D j s ( δ k δ j ) = 1 2 ( K L ( δ k δ k + δ j 2 ) + K L ( δ j δ k + δ j 2 ) )
10:         with K L ( δ k δ k + δ j 2 ) = i δ k ( i ) × l o g ( δ k ( i ) ( δ k ( i ) + δ j ( i ) ) / 2 )
11:         if  D j s ( δ k δ j ) < D j s t h  then
12:              C m k t h c l i e n t
13:              C m j t h c l i e n t
14:         end if
15:     end for
16: end for
17: B. Initialization of Global Model Training:
      Server execution: initialize ω 0
18: for  r o u n d t = 0 , 1 , 2 , , T  do
19:     for  every combination , C m = 1 , 2 , , M  do
20:         for  e a c h c l i e n t m k C m  do
21:              ω m k t + 1 ω m k t η m k ð m k t
22:         end for 
23:          Combination aggregation :
24:          ω C m t + 1 = C m k C m K ( 1 γ m β m ) n m k n m + ( 1 α m γ m ) l m k l m + ( 1 α m β m ) δ m k δ m ω m k t + 1
25:     end for
26:      Global aggregation :
27:      ω t + 1 = C m C M ( 1 γ β ) n m n + ( 1 α γ ) l m L + ( 1 α β ) δ m δ ω C m t + 1
28: end for
The aim is to leverage patient data to enhance the classification accuracy and reduce the loss function of the global model on the server. The ultimate objective is to attain the lowest loss value when predicting any given sample, ( x i , y i ) D of the patient. The objective is to reduce the loss on the patients’ data, D, using the global model, ω , which can be expressed as
m i n ω R l ( ω , D ) ,
where l ( ω , D ) = 1 n i n l i ( ω , D ) with l i ( ω , D ) is the loss of the prediction. At each round t, global model ω t is shared with all hospitals and medical centers which then use their local data to calculate gradients and update the global model. The gradients of the kth client of mth group at round t are computed as
ð m k t = ω m k t l ( ω m k t , D m k ) ,
where D m k denotes the dataset of the m k th client of mth group, comprising n m k samples ( x m k i , y m k i ) with 1 < i < n m k , as defined earlier. The local objective for that client can be formulated as
m i n ω m k t R l ( ω m k t , D m k ) ,
where l ( ω m k t , D m k ) , is the prediction loss expressed as
l ( ω m k t , D m k ) = 1 n m k ( x m k i , y m k i ) D m k f m k i ( ω m k t ) .
The loss function, f m k i ( ω m k t ) , for making prediction on given samples, ( x m k i , y m k i ) , using model ω m k t is defined as L ( ω m k t , x m k i , y m k i ) . The above Equation (5) is given by
ð m k t = ω m k t ( x m k i , y m k i ) D m k L ( ω m k t ; x m k i , y m k i ) .
Once the gradients are calculated, the kth client of the mth group’s local model is updated, which can be written as
ω m k t + 1 ω m k t η m k ð m k t ,
where η m k is the learning rate of the kth client of the mth group. The updated local model of that client at round t + 1 is given as ω m k t + 1 , which is forwarded to the server.

4.4. Server Aggregation

Upon receiving the locally trained models, the server engages in two aggregations: combination aggregation and global aggregation. In combination aggregation, group aggregation occurs, and the aggregation weighting is determined based on the three parameters: volume of patient data, classes of labels of that data, and the variance in the latent space of each medical center. Each client has previously computed its own latent space using VAE, which has already been shared with the server. The server, having segmented clients based on their respective latent spaces, assigns aggregation weightage to clients during combination aggregation, considering the above three parameters. After the grouping aggregation, multiple grouped local models are obtained. Let us denote the mth combination of clients by C m and assume that it comprises a K number of clients’ models. The combination aggregation of the group is computed by the group head, which is the medical center, with a large variance in the latent space. This combination aggregation is represented as
ω C m t + 1 = C m k C m K ( 1 γ m β m ) n m k n m + ( 1 α m γ m ) l m k l m + ( 1 α m β m ) δ m k δ m ω m k t + 1 ,
where n m k , l m k , and δ m k show the volume, the latent space variance, and the classes of labels of patients data of the kth medical center of the mth group. The n m , δ m , and l m are defined as n m = C m k C m K n m k , δ m = C m k C m K δ m k and l m = C m k C m K l m k . Moreover, ( 1 γ m β m ) represents the weightage assigned to the data volume, ( 1 α m γ m ) denotes the aggregation weightage allocated to data label classes, and ( 1 α m β m ) indicates the aggregation weightage attributed to the variance of the latent space in m—the combination.
To clarify further, the above Equation (10) encapsulates a sophisticated combination aggregation process designed for clustered federated learning. It represents the aggregation of updates from individual medical centers grouped together, aiming to leverage distributed data while preserving privacy and security. At its core, the equation computes the aggregated result for a specific group at a given time. The updated clustered model reflects the collective insights gleaned from various medical centers within the group. Each medical center contributes to the aggregation process based on its unique data characteristics, which include data volume, latent space variance, and data label classes. These components capture essential aspects of the data distribution and characteristics within each medical center, providing valuable insights for the aggregation process.
Moreover, the equation incorporates weighted contributions from each data characteristic, with weightages ( α m , β m , γ m ) determining the relative importance of data label classes, latent space variance, and data volume in the aggregation process. These weightages offer flexibility in adjusting the emphasis placed on different data characteristics based on the specific requirements or priorities of the learning task. Additionally, the equation calculates aggregated metrics ( n m , δ m , l m ) that represent the total volume, label classes, and latent space variance, respectively, for the entire group. These aggregated metrics provide a comprehensive overview of the collective data characteristics within the group of medical centers, enabling informed decision-making during the aggregation process.
The medical center already has information regarding the number of patient data samples and label classes. However, the latent space for patient data needs to be computed using VAE. In essence, the first and second parts of the equation represent the weighting based on the number of samples and label classes, while the third part relies on the patient data’s latent space.
To enhance the understanding of the third part of the equation, we perform some simulations as shown in Figure 3. The algorithm can segment clients into groups based on their JSD between latent embeddings. However, for illustrative purposes, we present a simulation involving three groups of medical centers. In Figure 3a–c, the latent variables of the three groups are displayed. Specifically, Group 2 exhibits the latent variable of five centers with an accumulative probability density, showcasing substantial variation in the latent variable of that group as shown in Figure 3b. A wide range of latent variables implies that diverse patterns in the data are captured, enabling the model to represent a wide variety of information. This indicates that the model has learned to encode different features or aspects of the data in various regions of the latent space. The group characterized by significant variation is assigned a higher weight. Among the three groups illustrated in this example, as evident in Figure 3d, Group 2 receives the highest weightage due to its substantial variance in the combined latent space. Following that, Group 3 is considered, and finally, Group 1 exhibits a comparatively lower variance in the combined latent space.
Subsequently, global aggregation takes place, where the volume of patient data determines the aggregation weightage assigned to each combination, the variance in the latent space, and the classes of labels of that particular group, n m , l m , and δ m . Once the M number of trained groups is updated, the global aggregation is performed at round t + 1 as
ω t + 1 = C m C M ( 1 γ β ) n m n + ( 1 α γ ) l m L + ( 1 α β ) δ m δ ω C m t + 1 ,
where n m , l m , and δ m were defined earlier. The values of n, δ , and l are determined as n = C m C M n m , δ = C m C M δ m , and L = C m C M l m . Additionally, ( 1 γ β ) represents the weightage assigned to the data volume, ( 1 α γ ) denotes the aggregation weightage allocated to data label classes, and ( 1 α β ) indicates the aggregation weightage attributed to the variance of the latent space.
To be more specific, Equation (11) describes the global aggregation process in federated learning, which occurs after individual groups update their models based on local data. At its core, the equation computes the aggregated result for the entire federated learning system at a given round. This aggregated result consolidates the knowledge from all trained groups. Each group contributes uniquely to the global aggregation, leveraging its distinct data characteristics, encompassing data volume, latent space variance, and data label classes. These facets capture critical aspects of the data distribution and properties within each group of medical centers, enriching the aggregation process with valuable insights. Moreover, the equation integrates weighted contributions from each data characteristic, where γ , β , and α dictate the relative significance of the data volume, latent space variance, and data label classes in the aggregation process. These weightages offer adaptability, allowing for adjustments to be made based on specific task requirements or priorities.
Additionally, the equation computes aggregated metrics representing the cumulative volume, label classes, and latent space variance across all trained groups. These metrics furnish a holistic overview of collective data characteristics, facilitating informed decision-making during the aggregation process. In essence, Equation (11) serves as the cornerstone of our proposed method, tailored for scenarios where data are distributed across multiple groups of medical centers. It seamlessly amalgamates insights from diverse sources, acknowledging their individual traits, thereby fostering collaborative learning endeavors without compromising data privacy or security.
Following the update of the global model, it is communicated back to all medical centers, and the global training process is iteratively repeated for multiple rounds until convergence is attained.
We additionally examine the computational complexity of our proposed approach. The complexity of initializing the VAE model involves O ( ε · K · d ) , where ε represents the epochs, K denotes the number of clients, and d signifies the features per client. The encoding process for each client’s data incurs O ( K · d ) complexity, transforming data into latent representations. The calculation of divergence between each pair of clients results in O ( K 2 ) complexity. Decision-making based on pairwise divergences adds O ( K 2 ) complexity.
The total complexity of our algorithm, considering these components, is given by
C V A E = O ( ε · K · d ) + O ( K · d ) + O ( K 2 )
For practical scenarios with a large K, the O ( K 2 ) term usually dominates. Thus, the overall complexity primarily becomes
C V A E = O ( ε · K · d + K 2 ) .
This analysis underscores the feasibility and efficiency of our algorithm in real-world applications.

5. Experiments

5.1. Data Set

To validate our proposed approach, we conduct experiments using real-world datasets, including brain tumor magnetic resonance imaging (MRI) [37] and chest X-ray images for pneumonia detection [38]. Our method is designed to achieve high performance regardless of the characteristics of the data. Consequently, we assess and compare our method on additional datasets such as Fashion-MNIST [39] and CIFAR-10 [40] to demonstrate its performance.
The brain tumor MRI dataset [37] comprises 3264 images categorized into four classes: glioma, meningioma, pituitary, and no tumor. The chest X-ray images for pneumonia detection consist of 5863 images classified into two classes. Fashion-MNIST comprises 60,000 training samples and is evaluated using 10,000 testing images. Each example is a 28 × 28 grayscale image associated with a label of one of 10 classes. The CIFAR-10 dataset consists of 60,000 color images of 32 × 32 with 10 classes. Among 60,000 images, 50,000 are for training and 10,000 are for testing. Each client is randomly assigned data examples in a non-IID manner. Specific clients may possess only a small amount of data, while others may have much larger data samples.

5.2. Network Model

Our method’s implementation involves the use of TensorFlow. We employ two neural network models: CNN for the detection of brain tumors and pneumonia as well as for the CIFAR dataset, and DNN for the Fashion-MNIST dataset. The CNN model configuration consists of three Conv2D layers, each with 32 filters of size 3 × 3 and ReLu activation. This is succeeded by two hidden layers, each containing 500 neurons with the ReLu activation function. The model concludes with an output layer utilizing the ’Softmax’ function. The DNN model comprises an input layer, four hidden layers comprising 500 neurons each, utilizing a ReLu activation function, and an output layer employing a softmax activation function.
The data’s integer labels undergo encoding through the OneHotEncoder method, yielding a binary column for each label and generating a dense array. Every client updates its local model using batch sizes of 32 and completes a single epoch per aggregation round. Local training consists of one epoch, with a learning rate set at 0.01 and a mini-batch size of 32. Simulations span up to 100 rounds. The ultimate goal of the proposed method is to achieve improved accuracy and minimal loss in the prediction of data samples using the global model. This is attained by forming efficient client groups through using autoencoders and assigning aggregation weightage accordingly.

5.3. Results

Referring to the literature, FCM [10], AP cluster formation [24], Atraea [23], Hier Local-QSGD [21], HFEL [22], and Hyperparameter-based [27] are acknowledged as influential approaches for grouping the nodes in federated learning. Hence, we compare our proposed method with the conventional approaches mentioned above.
The ultimate aim of our proposed method is to achieve increased accuracy and reduced loss in predicting patient data using the global model. This improved performance is attained through the effective formation of clusters and appropriate weightage assigned to clients during the aggregation process. The performance of our proposed method with real-world datasets for brain tumor and MRI detection using the CNN network is illustrated in Figure 4 and Figure 5. Specifically, the accuracy for tumor and pneumonia detection is depicted in Figure 4 and Figure 5, respectively, while the loss is presented in Figure 6 and Figure 7. It is evident from the figures that our proposed method achieves the highest accuracy and lowest loss compared to conventional methods in detecting tumors and pneumonia. The primary factor contributing to achieving the highest accuracy and minimizing loss is our method of clustering, which involves leveraging autoencoders to incorporate the extracted features of local hospitals. Additionally, by assigning aggregation weights based on these extracted features as well as the volume and variance of their data, we further enhance the performance of our proposed method.
Our method extends beyond IoT healthcare. By applying our proposed approach of grouping and aggregation weightage assignment, one can achieve optimal performance in terms of both accuracy and loss. To demonstrate this, we perform additional simulations of our method using the Fashion MNIST and CIFAR datasets. The performance comparison with a CNN network using the CIFAR dataset is depicted in Figure 8 and Figure 9. These figures also clearly demonstrate the superior performance in terms of accuracy and loss.
In other words, the results depicted in Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9 indicate that the proposed method attains higher accuracy and lower loss in comparison to conventional methods when employing a CNN network trained on the MRI images, X-Ray Images, and the CIFAR dataset. Additionally, the performance of the proposed method, employing the Fashion-MNIST dataset with a DNN network, is depicted in Figure 10 and Figure 11, showcasing leading results. Figure 10 clearly shows that the proposed method is superior in terms of accuracy across all aggregated rounds. It not only starts strong but also maintains and slightly improves performance, stabilizing at a higher accuracy level compared to the other methods. This consistency and superior performance make it a robust and reliable method for the given task. The proposed method’s rapid rise in accuracy and its ability to maintain a high accuracy level throughout the training process highlight its effectiveness and efficiency in learning and generalizing from the data. On the other hand, Figure 11 also demonstrates that the proposed method excels in reducing loss over a series of training rounds. It starts with a higher loss but quickly and consistently minimizes it, maintaining the lowest loss compared to other methods. This consistent and significant reduction in loss indicates the proposed method’s efficiency in training and optimizing the model. The ability to achieve and maintain a lower loss highlights the proposed method’s robustness and effectiveness in improving model performance and minimizing errors, making it the superior choice among the compared methods.
Moreover, for a better illustration, we present the performance enhancements across all the data sets considered in Table 2. The data in the table reveal that the proposed method achieves the highest performance in every aspect. To assess the performance improvement of our proposed method compared to baseline approaches, we contrast the accuracy and loss attained by each baseline method with those achieved by our proposed method. For each baseline method, we compute the discrepancy between its performance and that of our proposed method, indicating the absolute improvement conferred by our approach over that specific baseline. To standardize the measure of improvement, we then normalize each absolute improvement by the performance of the corresponding baseline method. This computation yields the relative improvement conferred by our proposed method in comparison to each baseline approach. By averaging these relative improvements across all baseline methods, we derive a comprehensive assessment of the overall improvement realized by our proposed method relative to the typical performance of existing approaches. As indicated in Table 2a, when using MRI data, the proposed method achieves the highest accuracy of 0.848. This is a substantial improvement over other methods, with the Hyper-parameter-based method coming in second at 0.799 and Hier-Local-QSGD coming in at 0.758. The proposed method also records the lowest loss at 1.675, which indicates better performance and less error. The next best performing method in terms of loss is Astraea with a loss of 1.791. Overall, with the MRI dataset, the proposed method achieves a 15% improvement in accuracy and a 7% reduction in loss, as shown in Table 2a. Similarly, with X-ray images, the proposed method again outperforms all others with an accuracy of 0.941. The Hyper-parameter-based method achieves 0.909, while the Hier-Local-QSGD method has an accuracy of 0.891. The proposed method’s loss is the lowest at 1.629, indicating better model efficiency. The Hyper-parameter-based method follows with a loss of 1.668. With a 10.5% improvement in accuracy and a 4.8% reduction in loss, the proposed method demonstrates notable superiority in handling the pneumonia dataset, as shown in Table 2b. Furthermore, when employing Fashion-MNIST data, the proposed method reaches the highest accuracy of 0.839. This is significantly higher than that of Hier-Local-QSGD at 0.771 and the Hyper-parameter-based method at 0.799. The lowest loss is also achieved by the proposed method at 1.748. HFEL follows with a loss of 1.801, and the Hyper-parameter-based method has a loss of 1.779. In other words, the proposed method attains a 13.1% improvement in accuracy and a 3.7% reduction in loss, as shown in Table 2c. In contrast, with CIFAR data, the proposed method achieves the highest accuracy of 0.551, outperforming the Hyper-parameter-based method (0.485) and Hier-Local-QSGD (0.457). It also records the lowest loss at 1.960. The next best is Hier-Local-QSGD at 2.025, followed by the Hyper-parameter-based method at 2.026. Specifically, the proposed method achieves an accuracy improvement of 20.8%, along with a 4.8% reduction in loss, as shown in Table 2d.
We also simulate the time complexity of our proposed method and compare it with traditional approaches. Figure 12 above illustrates the time complexity of various methods in relation to the number of hospitals involved. The x-axis represents the number of hospitals, which ranges from 2 to 50. This variable indicates the scale of the data involved and the corresponding number of clients contributing to the distributed learning process. The y-axis denotes the time complexity. This metric captures the computational time required by each method as the number of hospitals increases. All methods show an increasing trend in time complexity as the number of hospitals increases. This trend is expected due to the larger amount of data and higher computational demands with more clients. It is clear from Figure 12 that the proposed method exhibits a relatively lower time complexity compared to other methods, indicating efficient computational performance as the number of hospitals increases. It also outperforms other conventional methods, showing lower computational time as the number of hospitals increases.
Furthermore, Figure 13 presents a bar chart illustrating the computational complexity when considering 50 hospitals. This suggests that the proposed method is well-suited for scalable applications where computational efficiency is critical.
In addition, we include a radar plot in Figure 14 comparing our proposed method with conventional methods, demonstrating that our approach achieves superior performance. The figure illustrates the comparative performance based on three key metrics, complexity, loss, and accuracy. The data for each method are normalized to allow for a clear visual comparison. Specifically, the vertices of the chart correspond to the normalized values of complexity, loss, and accuracy for each method. It is clear from the figure that the proposed method achieves a balance across all three metrics, maintaining low complexity, low loss, and high accuracy.
The simulation results above indicate that our proposed method shows promise in enhancing performance in healthcare and federated learning through the use of autoencoders compared to conventional methods.

6. Conclusions

Our proposed method effectively handles non-IID data challenges and enhances IoT healthcare by leveraging federated learning and variational autoencoders. Our method comprises two proposed approaches: the efficient formation of clusters and the proper assignment of the aggregation weightage to local hospitals and formed clusters. Specifically, by learning and extracting the features of patients’ data and utilizing their latent space representations, we develop a novel approach to forming diverse groups of local hospitals. In addition, we propose a method that assigns aggregation weightage based on several parameters, including the variance of extracted features of patients, the data volume of patients, and the data classes of patients. Our methodology surpasses the performance limitations associated with existing methods by combining these two approaches.

Author Contributions

Conceptualization: D.M.S.B. and B.J.C.; data curation: D.M.S.B.; formal analysis: D.M.S.B.; funding acquisition: B.J.C.; methodology: D.M.S.B. and B.J.C.; project administration: B.J.C.; resources: D.M.S.B.; software: D.M.S.B.; supervision: B.J.C.; validation: D.M.S.B. and B.J.C.; visualization: D.M.S.B.; writing—original draft: D.M.S.B.; writing—review and editing: D.M.S.B. and B.J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This researchwas supported by the MSIT Korea under the NRF Korea (NRF-2022R1A2C4001270, 80%), the Innovative Human Resource Development for Local Intellectualization support program (IITP-2024-RS-2022-00156360, 10%) supervised by the IITP, and the KIAT grant funded by the Korean government (MOTIE) (P0017123, The Competency Development Program for Industry Specialist, 10%).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within this article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
APAffinity propagation
CNNConvolutional neural network
DNNDeep neural network
FedAvgFederated averaging
Non-IIDNon-independent and non-identically distributed
IoTInternet of things
JSDJensen–Shanon divergence
MRIMagnetic resonance imaging
VAEVariational autoencoder

References

  1. Mengistu, T.M.; Kim, T.; Lin, J.W. A Survey on Heterogeneity Taxonomy, Security and Privacy Preservation in the Integration of IoT, Wireless Sensor Networks and Federated Learning. Sensors 2024, 24, 968. [Google Scholar] [CrossRef] [PubMed]
  2. Hyysalo, J.; Dasanayake, S.; Hannu, J.; Schuss, C.; Rajanen, M.; Leppänen, T.; Doermann, D.; Sauvola, J. Smart mask—Wearable IoT solution for improved protection and personal health. IoT 2022, 18, 100511. [Google Scholar] [CrossRef] [PubMed]
  3. Avanzato, R.; Beritelli, F.; Lombardo, A.; Ricci, C. Lung-DT: An AI-Powered Digital Twin Framework for Thoracic Health Monitoring and Diagnosis. Sensors 2024, 24, 958. [Google Scholar] [CrossRef]
  4. Chaves, A.J.; Martín, C.; Díaz, M. Towards flexible data stream collaboration: Federated Learning in Kafka-ML. IoT 2024, 25, 101036. [Google Scholar] [CrossRef]
  5. Sun, P.; Shen, S.; Wan, Y.; Wu, Z.; Fang, Z.; Gao, X.-Z. A survey of iot privacy security: Architecture, technology, challenges, and trends. IEEE IoT J. 2024. [Google Scholar] [CrossRef]
  6. Bhatti, D.M.S.; Nam, H. FedCLS: Class-Aware Federated Learning in a Heterogeneous Environment. IEEE Trans. Netw. Serv. Manag. 2023, 20, 1517–1528. [Google Scholar] [CrossRef]
  7. Hossain, M.B.; Shinde, R.K.; Oh, S.; Kwon, K.C.; Kim, N. A Systematic Review and Identification of the Challenges of Deep Learning Techniques for Undersampled Magnetic Resonance Image Reconstruction. Sensors 2024, 24, 753. [Google Scholar] [CrossRef] [PubMed]
  8. Khan, A.N.; Rizwan, A.; Ahmad, R.; Khan, Q.W.; Lim, S.; Kim, D.H. A precision-centric approach to overcoming data imbalance and non-IIDness in federated learning. IoT 2023, 23, 100890. [Google Scholar] [CrossRef]
  9. Bhatti, D.M.S.; Nam, H. A Performance Efficient Approach of Global Training in Federated Learning. In Proceedings of the 2023 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Bali, Indonesia, 20–23 February 2023; pp. 112–115. [Google Scholar] [CrossRef]
  10. Bhatti, D.M.S.; Saeed, N.; Nam, H. Fuzzy C-Means Clustering and Energy Efficient Cluster Head Selection for Cooperative Sensor Network. Sensors 2016, 16, 1459. [Google Scholar] [CrossRef]
  11. Sattler, F.; Müller, K.R.; Samek, W. Clustered Federated Learning: Model-Agnostic Distributed Multitask Optimization Under Privacy Constraints. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 3710–3722. [Google Scholar] [CrossRef]
  12. Ghosh, A.; Chung, J.; Yin, D.; Ramchandran, K. An Efficient Framework for Clustered Federated Learning. IEEE Trans. Inf. Theory 2022, 68, 8076–8091. [Google Scholar] [CrossRef]
  13. Briggs, C.; Fan, Z.; András, P. Federated learning with hierarchical clustering of local updates to improve training on non-IID data. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; pp. 1–9. [Google Scholar]
  14. Xie, M.K.; Long, G.; Shen, T.; Zhou, T.; Wang, X.; Jiang, J. Multi-Center Federated Learning. arXiv 2020, arXiv:2108.08647. [Google Scholar]
  15. Abad, M.S.H.; Ozfatura, E.; Gündüz, D.; Erçetin, Ö. Hierarchical Federated Learning ACROSS Heterogeneous Cellular Networks. In Proceedings of the ICASSP 2020–2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 8866–8870. [Google Scholar] [CrossRef]
  16. Liu, L.; Zhang, J.; Song, S.; Letaief, K.B. Client-Edge-Cloud Hierarchical Federated Learning. In Proceedings of the 2020 IEEE International Conference on Communications, Dublin, Ireland, 7–11 June 2020; pp. 1–6. [Google Scholar] [CrossRef]
  17. Duan, M.; Liu, D.; Ji, X.; Liu, R.; Liang, L.; Chen, X.; Tan, Y. FedGroup: Ternary Cosine Similarity-based Clustered Federated Learning Framework toward High Accuracy in Heterogeneous Data. arXiv 2020, arXiv:2010.06870. [Google Scholar]
  18. Qiao, C.; Brown, K.N.; Zhang, F.; Tian, Z. Federated Adaptive Asynchronous Clustering Algorithm for Wireless Mesh Networks. IEEE Trans. Knowl. Data Eng. 2021, 35, 2610–2627. [Google Scholar] [CrossRef]
  19. Huang, Y.; Chu, L.; Zhou, Z.; Wang, L.; Liu, J.; Pei, J.; Zhang, Y. Personalized Federated Learning: An Attentive Collaboration Approach. arXiv 2020, arXiv:2007.03797. [Google Scholar]
  20. Qayyum, A.; Ahmad, K.; Ahsan, M.A.; Al-Fuqaha, A.I.; Qadir, J. Collaborative Federated Learning For Healthcare: Multi-Modal COVID-19 Diagnosis at the Edge. arXiv 2021, arXiv:2101.07511. [Google Scholar] [CrossRef]
  21. Liu, L.; Zhang, J.; Song, S.; Letaief, K.B. Hierarchical Federated Learning with Quantization: Convergence Analysis and System Design. IEEE Trans. Wirel. Commun. 2022, 22, 2–18. [Google Scholar] [CrossRef]
  22. Luo, S.; Chen, X.; Wu, Q.; Zhou, Z.; Yu, S. HFEL: Joint Edge Association and Resource Allocation for Cost-Efficient Hierarchical Federated Edge Learning. IEEE Trans. Wirel. Commun. 2020, 19, 6535–6548. [Google Scholar] [CrossRef]
  23. Duan, M.; Liu, D.; Chen, X.; Tan, Y.; Ren, J.; Qiao, L.; Liang, L. Astraea: Self-Balancing Federated Learning for Improving Classification Accuracy of Mobile Deep Learning Applications. In Proceedings of the 2019 IEEE 37th International Conference on Computer Design (ICCD), Abu Dhabi, United Arab Emirates, 17–20 November 2019; pp. 246–254. [Google Scholar] [CrossRef]
  24. Bhatti, D.M.S.; Ahmed, S.; Chan, A.S.; Saleem, K. Clustering formation in cognitive radio networks using machine learning. AEU-Int. J. Electron. Commun. 2020, 114, 152994. [Google Scholar] [CrossRef]
  25. Dennis, D.K.; Li, T.; Smith, V. Heterogeneity for the Win: One-Shot Federated Clustering. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 2611–2620. [Google Scholar]
  26. Lloyd, S. Least squares quantization in PCM. IEEE Trans. Inf. Theory 1982, 28, 129–137. [Google Scholar] [CrossRef]
  27. Gholizadeh, N.; Musilek, P. Federated learning with hyperparameter-based clustering for electrical load forecasting. IoT 2022, 17, 100470. [Google Scholar] [CrossRef]
  28. de Moraes Sarmento, E.M.; Ribeiro, I.F.; Marciano, P.R.N.; Neris, Y.G.; de Oliveira Rocha, H.R.; Mota, V.F.S.; da Silva Villaça, R. Forecasting energy power consumption using federated learning in edge computing devices. IoT 2024, 25, 101050. [Google Scholar] [CrossRef]
  29. Cao, X.; Sun, G.; Yu, H.; Guizani, M. PerFED-GAN: Personalized Federated Learning via Generative Adversarial Networks. IEEE IoT J. 2022, 10, 3749–3762. [Google Scholar] [CrossRef]
  30. Yao, X.; Huang, T.; Zhang, R.; Li, R.; Sun, L. Federated Learning with Unbiased Gradient Aggregation and Controllable Meta Updating. arXiv 2019, arXiv:1910.08234. [Google Scholar]
  31. Zhang, L.; Shen, L.; Ding, L.; Tao, D.; Duan, L.Y. Fine-tuning Global Model via Data-Free Knowledge Distillation for Non-IID Federated Learning. arXiv 2022, arXiv:2203.09249. [Google Scholar] [CrossRef]
  32. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated Optimization in Heterogeneous Networks. arXiv 2018, arXiv:1812.06127. [Google Scholar] [CrossRef]
  33. Fallah, A.; Mokhtari, A.; Ozdaglar, A. Personalized Federated Learning: A Meta-Learning Approach. arXiv 2020, arXiv:2002.07948. [Google Scholar] [CrossRef]
  34. Bhatti, D.M.S.; Nam, H. A Robust Aggregation Approach for Heterogeneous Federated Learning. In Proceedings of the 2023 Fourteenth International Conference on Ubiquitous and Future Networks (ICUFN), Paris, France, 4–7 July 2023; pp. 300–304. [Google Scholar] [CrossRef]
  35. McMahan, H.B.; Moore, E.; Ramage, D.; Arcas, B.A.Y. Federated Learning of Deep Networks using Model Averaging. arXiv 2016, arXiv:1602.05629. [Google Scholar]
  36. Ahmed, T.; Longo, L. Examining the Size of the Latent Space of Convolutional Variational Autoencoders Trained With Spectral Topographic Maps of EEG Frequency Bands. IEEE Access 2022, 10, 107575–107586. [Google Scholar] [CrossRef]
  37. Sartaj, B.A.; Kadam, P.; Bhumkar, S.; Dedge, S.K. Brain Tumor Classification (MRI). 2020. Available online: https://www.kaggle.com/dsv/1183165 (accessed on 25 January 2024).
  38. Kermany, D.; Zhang, K.; Goldbaum, M. Labeled Optical Coherence Tomography and Chest X-ray Images for Classification. 2018. Available online: https://www.kaggle.com/datasets/paultimothymooney/chest-xray-pneumonia/data (accessed on 28 February 2024).
  39. Xiao, H.; Rasul, K.; Vollgraf, R. Fashion-MNIST: A Novel Image Dataset for Benchmarking Machine Learning Algorithms. arXiv 2017, arXiv:1708.07747. [Google Scholar]
  40. Alex, K.; Vinod, N.; Geoffrey, H. The CIFAR-10 Dataset. 2009. Available online: http://www.cs.toronto.edu/~kriz/cifar.html (accessed on 25 November 2023).
Figure 1. Proposed VAE-based segmentation for global training.
Figure 1. Proposed VAE-based segmentation for global training.
Sensors 24 03632 g001
Figure 2. Proposed global training for IoT healthcare integrated with federated learning.
Figure 2. Proposed global training for IoT healthcare integrated with federated learning.
Sensors 24 03632 g002
Figure 3. Example of Latent space variance of each hospital and group.
Figure 3. Example of Latent space variance of each hospital and group.
Sensors 24 03632 g003
Figure 4. Comparison of accuracy with CNN network for brain tumor detection.
Figure 4. Comparison of accuracy with CNN network for brain tumor detection.
Sensors 24 03632 g004
Figure 5. Comparison of accuracy with CNN network for pneumonia detection.
Figure 5. Comparison of accuracy with CNN network for pneumonia detection.
Sensors 24 03632 g005
Figure 6. Comparison of loss with CNN network for brain tumor detection.
Figure 6. Comparison of loss with CNN network for brain tumor detection.
Sensors 24 03632 g006
Figure 7. Comparison of loss with CNN network for pneumonia detection.
Figure 7. Comparison of loss with CNN network for pneumonia detection.
Sensors 24 03632 g007
Figure 8. Comparison of accuracy with CNN network using CIFAR data.
Figure 8. Comparison of accuracy with CNN network using CIFAR data.
Sensors 24 03632 g008
Figure 9. Comparison of loss with CNN network using CIFAR data.
Figure 9. Comparison of loss with CNN network using CIFAR data.
Sensors 24 03632 g009
Figure 10. Comparison of accuracy with DNN network using Fashion-MNIST data.
Figure 10. Comparison of accuracy with DNN network using Fashion-MNIST data.
Sensors 24 03632 g010
Figure 11. Comparison of loss with DNN network using Fashion-MNIST data.
Figure 11. Comparison of loss with DNN network using Fashion-MNIST data.
Sensors 24 03632 g011
Figure 12. Comparison of time complexity of proposed method with conventional approaches.
Figure 12. Comparison of time complexity of proposed method with conventional approaches.
Sensors 24 03632 g012
Figure 13. Comparison of time complexity with fixed number of hospitals.
Figure 13. Comparison of time complexity with fixed number of hospitals.
Sensors 24 03632 g013
Figure 14. Performance comparison of proposed method with conventional approaches.
Figure 14. Performance comparison of proposed method with conventional approaches.
Sensors 24 03632 g014
Table 1. Comparison of existing work with proposed method. The check mark indicates that the listed reference has addressed the specific issue, whereas the cross signifies that the method has not focused on that particular issue.
Table 1. Comparison of existing work with proposed method. The check mark indicates that the listed reference has addressed the specific issue, whereas the cross signifies that the method has not focused on that particular issue.
Existing MethodsFramework UsedClusteringHeterogeneityAggregationMajor Contributions
FCM [10]Fuzzy c-means clustering××
  • A fuzzy c-means-based clustering algorithm is proposed to save the wastage of energy.
  • The algorithm makes the cluster and selects the heads such that very little amount of network energy is consumed.
Astraea [23] Kullback–Leibler divergence (KLD)××
  • A self-balancing federated learning framework is proposed, which alleviates the imbalances by global data distribution-based data augmentation and mediator-based multi-client rescheduling.
AP cluster formation [24] Affinity propagation machine learning××
  • A machine learning-based algorithm is proposed to save the wastage of energy and enhance performance.
  • The algorithm dynamically makes the clusters based on the data received from other users.
HFEL [22]Hierarchical Federated Edge Learning×
  • Model aggregation is partially migrated to geographically fixed edge nodes.
  • A resource scheduling algorithm is proposed in the HFEL framework.
FL+ HC [13]Distance-based clustering×
  • A hierarchical clustering is proposed to separate clusters of clients by the similarity of their local updates to the global joint model.
k-FED [25]Lloyd’s method for k-means clustering×
  • A one-shot communication scheme is proposed that addresses common practical concerns in federated settings, such as high communication costs, stragglers, and device failures.
Hier-Local QSGD [21]Hierarchical Federated Learning with Quantization×
  • A tighter convergence bound for hierarchical federated learning with quantization is presented to formulate the problem of selecting aggregation intervals.
Hyperparameter-Based [27]Hyperparameter-based clustering××
  • A consumer clustering technique is proposed that better reflects the consumer consumption pattern and does not require sharing confidential consumer data.
Proposed methodVariational auto-encoder
  • A novel method is proposed which extracts and learns data features to create groups taking into consideration the heterogeneity of the data.
  • The proposed approach considers the extracted features of data and conducts aggregation in a way that optimizes both accuracy and loss.
Table 2. Detailed performance of the proposed and conventional methods on a variety of data sets.
Table 2. Detailed performance of the proposed and conventional methods on a variety of data sets.
(a) Performance with CNN network using MRI dataset
Comparing methodsAccuracy @MRI dataLoss @MRI data
FCM0.6711.889
AP0.7001.862
Astraea0.7411.791
HFEL0.7591.701
Hier-Local QSGD0.7581.752
Hyper-parameter Based0.7991.752
Proposed method0.8481.675
Performance improvement15%7%
(b) Performance with CNN network using pneumonia dataset
Comparing methodsAccuracy @pneumonia dataLoss @pneumonia data
FCM0.8001.758
AP0.8301.711
Astraea0.8301.738
HFEL0.8491.701
Hier-Local-QSGD0.8911.668
Hyper-parameter-based0.9091.668
Proposed method0.9411.629
Performance improvement10.5%4.8%
(c) Performance with DNN network using Fashion-MNIST
Comparing methodsAccuracy @Fashion MNISTLoss @Fashion MNIST
FCM0.7111.848
AP0.7161.819
Astraea0.7191.847
HFEL0.7411.801
Hier-Local-QSGD0.7711.779
Hyper-parameter-based0.7991.779
Proposed method0.8391.748
Performance improvement13.1%3.7%
(d) Performance with CNN network using CIFAR
Comparing methodsAccuracy @CIFARLoss @CIFAR    
FCM0.4482.079
AP0.4492.061
Astraea0.4502.066
HFEL0.4522.065
Hier-Local-QSGD0.4572.025
Hyper-parameter-based0.4852.026
Proposed method0.5511.960
Performance improvement20.8%4.8%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bhatti, D.M.S.; Choi, B.J. Enhancing IoT Healthcare with Federated Learning and Variational Autoencoder. Sensors 2024, 24, 3632. https://doi.org/10.3390/s24113632

AMA Style

Bhatti DMS, Choi BJ. Enhancing IoT Healthcare with Federated Learning and Variational Autoencoder. Sensors. 2024; 24(11):3632. https://doi.org/10.3390/s24113632

Chicago/Turabian Style

Bhatti, Dost Muhammad Saqib, and Bong Jun Choi. 2024. "Enhancing IoT Healthcare with Federated Learning and Variational Autoencoder" Sensors 24, no. 11: 3632. https://doi.org/10.3390/s24113632

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop