Next Article in Journal
Exploring the Diversity of Nuclear Density through Information Entropy
Previous Article in Journal
Two Levels of Integrated Information Theory: From Autonomous Systems to Conscious Life
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bidirectional Decoupled Distillation for Heterogeneous Federated Learning

School of Computer and Artificial Intelligence, Beijing Technology and Business University, Beijing 100048, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2024, 26(9), 762; https://doi.org/10.3390/e26090762
Submission received: 19 July 2024 / Revised: 3 September 2024 / Accepted: 3 September 2024 / Published: 5 September 2024
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
Federated learning enables multiple devices to collaboratively train a high-performance model on the central server while keeping their data on the devices themselves. However, due to the significant variability in data distribution across devices, the aggregated global model’s optimization direction may differ from that of the local models, making the clients lose their personality. To address this challenge, we propose a Bidirectional Decoupled Distillation For Heterogeneous Federated Learning (BDD-HFL) approach, which incorporates an additional private model within each local client. This design enables mutual knowledge exchange between the private and local models in a bidirectional manner. Specifically, previous one-way federated distillation methods mainly focused on learning features from the target class, which limits their ability to distill features from non-target classes and hinders the convergence of local models. To solve this limitation, we decompose the network output into target and non-target class logits and distill them separately using a joint optimization of cross-entropy and decoupled relative-entropy loss. We evaluate the effectiveness of BDD-HFL through extensive experiments on three benchmarks under IID, Non-IID, and unbalanced data distribution scenarios. Our results show that BDD-HFL outperforms state-of-the-art federated distillation methods across five baselines, achieving at most 3% improvement in average classification accuracy on the CIFAR-10, CIFAR-100, and MNIST datasets. The experiments demonstrate the superiority and generalization capability of BDD-HFL in addressing personalization challenges in federated learning.

1. Introduction

In the age of big data, protecting data privacy is becoming increasingly important. Machine learning (ML) relies on high-quality datasets to create unbiased and generalizable models. However, implementing and maintaining such data is challenging due to technical and regulatory barriers [1]. Additionally, the inability to collect large amounts of data on a centralized server poses significant privacy challenges for deep learning.
To solve the data privacy problem, a decentralized privacy-preserving solution known as federated learning has emerged, which helps us mitigate the “data silo” dilemma by allowing client devices to complete each training procedure independently. Subsequently, the trained models are uploaded to the server, aggregated, and distributed globally [2,3]. Despite its advantages in data privacy, FL faces a significant challenge that impacts its performance: statistical diversity. Specifically, the Heterogeneous Federated Learning (HFL) models are trained on the device using locally available data. However, data across different devices often vary in distribution, encompassing different types and imbalanced quantities. This heterogeneity leads to models optimizing in different directions during training. As a result, when the global federated model is aggregated from these local clients, it often struggles to generalize effectively across individual devices, resulting in insufficient local personalization.
To solve the problem of local personalization, many Personalized Heterogeneous Federated Learning (pHFL) methods have been proposed, which can be divided into three categories: (1) the global fine-tuning model, exemplified by FedRep [4] and Per-FedAvg [5], which uses the global model for fine-tuning to obtain the client models; (2) the local personalized aggregation model, typified by PartialFed [6], APPLE [7], FedALA [8], and FedAF [9], which leverages client relationships to improve personalized model performance; and (3) the personalized private model, including FedMD [10], FML [11], pFedMe [12], and Ditto [13], which increases the level of personalization by introducing additional models in clients. However, the global fine-tuning model solely relies on a single global model for fine-tuning client models as shown in Figure 1a. In practical applications, the model structure of each client differs, which restricts local personalization [14]. The local personalized aggregation model improves personalization capabilities efficiently. However, similarity metrics, such as clustering or unsupervised technologies [15] on clients are challenging, and the associated computational cost is high.
In contrast, the personalized private model achieves local personalization by introducing additional models, which leverages knowledge distillation to overcome the restriction imposed by heterogeneous clients. FedMD, FML, or pFedMe distills knowledge from the private model to the local model using Kullback–Leibler divergence, guiding the local model’s optimization to align with the private model’s objectives. This one-way distillation facilitates local client personalization during FL training. However, as illustrated in Figure 1b, the one-way knowledge transfer limits the private model’s access to global information from other clients, which diminishes its generalization capability.
Inspired by Deep Mutual Learning (DML) [16] and federated mutual learning (FML) [11], we introduce a bidirectional distillation to allow both local and private networks to learn from each other during the training process. Specifically, mutual learning employs Kullback–Leibler divergence to align the probability estimates of the two related networks. This bidirectional distillation can be adjusted by cross-referencing the experience information of local and private models, preventing overfitting due to the excessive smoothing of model training, and thus enhancing the generalization of the model. This strategy makes up for the shortcomings brought by one-way distillation so that the private model can learn the global knowledge contained in the local model while enhancing the personalization of the local client model. The mechanism of bidirectional distillation in FL learning is shown in Figure 1c. However, these approaches [11] transfer knowledge with the target class, not fully utilizing the information of the non-target class, which hinders the convergence speed and personalization of local models [17].
Therefore, to solve this problem, we propose a novel federated learning paradigm called Bidirectional Decoupled Distillation for Heterogeneous Federated Learning (BDD-HFL). Firstly, we facilitate the transfer of private knowledge to the local model by introducing a private model to each local client. Through bidirectional distillation during each training iteration, we ensure that knowledge transfer occurs without compromising model accuracy. Secondly, we optimize both the local and private models with decoupled relative-entropy loss. Specifically, we decouple the logistics output of all models to target class and non-target class information, giving balanced weight to them, and then the Kullback–Leibler loss is used to optimize them respectively. This allows us to balance the effect of knowledge distillation across target and non-target classes, enhancing the overall effectiveness of the process. After that, we integrate our method into classical federated learning methods to demonstrate its effectiveness. Experimental results demonstrate that BDD-HFL outperforms other personalized approaches. It addresses the challenges posed by data heterogeneity while maintaining local personalization.
The main contributions of this paper are as follows:
  • We design a trustworthy federated learning framework with bidirectional distillation for heterogeneous scenarios, allowing two local networks to learn from each other through their decoupled relative-entropy loss. This approach prevents the increasing divergence in optimization direction between the local and private models, a common issue in one-way distillation methods.
  • We introduce decoupled knowledge distillation to balance the effect of knowledge distillation across target and non-target classes. This approach enhances cross-referencing information interaction and improves the convergence speed of both server and client models.
  • We integrate our proposed framework into four classical federated learning methods to verify its superiority on the CIFAR-10, CIFAR-100, and MNIST datasets.

2. Related Work

2.1. Heterogeneous Federated Learning

FL work has emerged as a prominent research area in recent years. McMahan et al. first proposed FedAVG [2], which demonstrated the effectiveness of FL while preserving data privacy. FedAVG achieves this by updating the parameters of multiple clients through weighted average parameters. However, Li et al. [18] pointed out that FL currently faces several challenges, including expensive communication costs between the server and the client, system heterogeneity at the client, data distribution among clients, and privacy issues.
Numerous studies have shown that Non-IID data distribution leads to degraded performance of the aggregated model, drifting of the updated client model, and slow convergence. Various federated learning (FL) approaches have been proposed to address these issues. Some works focus on reducing the variance in client updates to accelerate convergence. FedProx [19] enhances model stability by adding proximal regularization to the client model, where the proximal term reduces client drift by making the updated local model parameters close to the global model. However, this approach does not account for the fact that the optimal point of the local empirical target is different from the global optimal point, leading to lower performance. Scaffold [20] customizes the gradient update of the personalized model to correct the client drift. FedDyn [21] proposes a dynamic regularizer for each client to align the global and local models and also speed up model convergence. FedDC [22] uses the idea of grafting to fill parameter gaps with learned local drift variables as a way to maintain consistency in parameter levels. FedDisco [23] adds the discrepancy between local and global category distributions as a beneficial and complementary indicator to aggregate weights. The main limitation of these approaches is that they ignore differences in the client model and ignore local model personalization, leading to suboptimal performance. Some other approaches (Tuor [24]; Hybrid-FL [25]) use simple strategies such as data sharing, which implies that the client shares part of its data with the server, but data-sharing-based approaches may violate FL privacy requirements [26]. Also for privacy security, shared data can be extracted by distillation [27] or generated by generative adversarial networks (GANs [28]). The other approaches [29,30] optimize the FL training convergence by utilizing a probabilistic selection scheme. Refs. [31,32] presents a task-specific FL platform while reducing the overall communication cost.
In addition to the client drift problem, Non-IID data will also cause personalization problems. The server model, trained using data from all clients, often fails to generalize well to each client’s individual data, resulting in insufficient local personalization.

2.2. Personalized Federated Learning

Personalized federated learning for clients with Non-IID data has attracted a lot of attention. Existing methods can be divided into three categories: (1) global fine-tuning model, (2) local personalized aggregation model, and (3) local personalized private model.
The global fine-tuning model builds personalized models for each client by customizing a trained global model: Wang et al. [33] use each client’s private data to fine-tune the global model to generate a personalized model for the client. The model mixing approach APFL [34] combines a global model with a local model to customize a personalized model for the client. However, these methods do not fully consider the difficulties faced in the Non-IID case. The local personalized aggregation methods focus on learning local models with some aggregation strategy: FedAMP [35] generates an aggregated model for each client using an attention-inducing function and personalized aggregation. PartialFed [6] constructs a new local model for each batch by learning a binary and layer-level aggregation strategy for each client, selecting parameters from either the global model or the local model. FedALA [8] adaptively aggregates the downloaded global model and local model towards the local objective on each client to initialize the local model before training in each iteration. FedAF [9] proposed a novel aggregation-free FL algorithm to let clients collaboratively learn by leveraging peer knowledge. APPLE [7] involves downloading multiple client models, which increases the communication overhead and raises privacy concerns, as data from other clients may be compromised. The local personalized model methods learn local private models to achieve local personalization via information distillation: pFedMe [12] learns an additional personalized model for each client with Moreau envelopes. Specifically, it formulates a new bilevel optimization problem designed for personalized federated learning (pFedMe). By using the Moreau envelope as a regularized loss function, it decouples the cross-entropy optimization of personalized models from the learning of the global model. This approach updates the global model similarly to standard FL algorithms such as FedAvg while simultaneously optimizing the personalized models with low complexity. In Ditto [13], each client learns its own additional personalized model by incorporating a proximal term, which fetches information from the downloaded global model.
The current distillation model [36] uses a large pre-trained network as a teacher to provide additional knowledge to a smaller network of students. Applying this approach in FL [37], although the local knowledge is constantly transmitted to the client, the client’s model changes with the server-side model, causing the optimization direction of the private model with the local model to be larger and larger, and the mutual information interaction will seriously affect the performance of the global model. Unlike one-way knowledge transfer observed in a teacher–student network, Deep Mutual Learning (DML) [16] allows two networks to learn from each other through relative-entropy optimization during the training process. This bidirectional distillation mechanism is well suited for addressing the limitations associated with one-way distillation in the context of federated learning personalization problems, facilitating local personalization while maintaining model performance. However, the approach does not fully consider the transferability of knowledge and does not take full advantage of the model.
Hence, this article primarily integrates the concept of bidirectional distillation into federated learning to address the challenge of inadequate local personalization in Non-IID scenarios. Subsequently, we delve into the distillation process within bidirectional distillation in depth, aiming to enhance the effectiveness of model knowledge transfer.

3. Methods

To tackle the challenges stemming from the catastrophic forgetting of local data, insufficient local personalization, and diminished model effectiveness due to data heterogeneity, we introduce a more adaptable federated learning approach in this section: Bidirectional Decoupled Distillation For Heterogeneous Federated Learning. This approach ensures local personalization while significantly enhancing the performance of global models.

3.1. Problem Definition

In conventional FL, there is a server and N clients. The server randomly initializes the global model and sends the global model to selected clients, and each client copies the model and trains it using its private data. Subsequently, the trained model parameters are uploaded to the server, where they are aggregated and subjected to a weighted average. The server then transmits the aggregated parameters back to the clients, repeating this iterative process until model training converges. The overarching training objectives are delineated below:
w t = a r g min w F w = k = 1 K n k n f k ( w )
The aim is to obtain a server-side model parameter w t . f k ( w ) denotes the local cross-entropy loss function of the client, K represents the total number of clients, n represents the number of samples of all clients, and n k represents the amount of private data of client k. The process of server-side distribution is w w k , where clients copy the server-side parameters as local model parameters. The whole process of server-side model aggregation is as follows:
w t + 1 = k = 1 K n k n w t k
Its w t k represents the model parameters of local client k after the tth round of training. This signifies that the server sequentially aggregates the model parameters from the client side, weighted by the data, to produce the model parameters for the server side in the subsequent round.
Rather than solely addressing the conventional federated learning (FL) problem, we obtain local client-personalized models while the global model converges to the optimal level.

3.2. Framework

The whole framework of Bidirectional Decoupled Distillation For Heterogeneous Federated Learning (BDD-HFL) is shown in Figure 2. The server first delivers the global model to every client as the local model, including C l i e n t 1 to C l i e n t n . The additional private models act as the personalized model. For local training, every client performs training updates of the two local and private models. Taking C l i e n t n in Figure 2 as an example, the logits of the local model are decoupled into the target class (the yellow square) and non-target classes (the blue rectangle) while the logits of the private model are decoupled into the target class (the green square) and non-target classes (the grey rectangle). In bidirectional distillation between private and local models, each model distills information related to both target and non-target classes from the other. After several rounds of local updates, the client uploads the refined local model to the server, contributing to the generation of a new global model.

3.3. Bidirectional Decoupled Knowledge Distillation

One-way distillation from private to local is developed by relative-entropy loss, named Kullback–Leibler (KL) divergence loss. The relative-entropy loss is described as the formula below:
KL ( p p r i p l o c ) = i = 1 M p p r i i log p p r i i p l o c i
where p l o c and p p r i are the logits of the local and private models, while M is the class number. K L ( p p r i p l o c ) is used to distill knowledge from the private model to the local model. However, the traditional one-way distillation process limits the effective transfer of global information and hinders comprehensive knowledge exchange. Additionally, Kullback–Leibler (KL) loss tends to prioritize the target class, often overlooking the significance of non-target classes, which restricts the flexibility and learning capacity of both models. To address these issues, we introduce a novel bidirectional distillation method with decoupled relative-entropy loss for both local and private models. Unlike the conventional approach, our distillation process operates in two directions: from the local model to the private model, and from the private model to the local model. We structure the knowledge exchange through decoupled distillation into two components, target class (TC) and non-target class (NC) knowledge distillation, weighted appropriately to balance their contributions. Here, we define t as the index of the target class. The Formula (3) can be rewritten as:
KL ( p p r i p l o c ) = p p r i t log p p r i t p l o c t + i = 1 , i t M p p r i i log p p r i i p l o c i
Hence, the decoupled relative-entropy loss (DREL) can be described as follows:
DREL ( p p r i p l o c ) = α p p r i t log p p r i t p l o c t T C + β i = 1 , i t M p p r i i log p p r i i p l o c i N C
where the α and β balance the importance of the target class and non-target class distillation for more efficient knowledge transfer.
Then, we add the decoupled knowledge distillation to the federated process and rewrite the joint loss function for the private model as well as the local model as follows:
L l o c = L C l o c + DREL ( p p r i p l o c )
L p r i = L C p r i + DREL ( p l o c p p r i )
where L C p r i and L C l o c denote the cross-entropy loss, D R E L ( p p r i p l o c ) is the decoupled relative-entropy loss from the private model to the local model, and D R E L ( p l o c p p r i ) is the relative-entropy loss from the local to the private model. The direction of our knowledge transfer between the local model and the private model is bidirectional, where the local model transfers the server-side knowledge to the private model and obtains feedback from it, while both models are trained using private data with joint optimization on D R E L ( p p r i p l o c ) and D R E L ( p l o c p p r i ) . The effectiveness of the knowledge transfer can be improved while addressing local personalization, which improves the model’s performance.

3.4. Training Pipeline

In Algorithm 1, we show the algorithm pseudo-code of the process of BDD-HFL. We denote the global model as g l o b a l , total number of clients as K, and communication rounds as T. Firstly, the server initializes the g l o b a l model. In each round t, k clients are randomly selected, obtaining the global model g l o b a l as their local model l o c a l t k , then executing the ClientUpdate process and obtaining the updated local model l o c a l t + 1 k . On the client side, we first decouple the logits of the local model to the target class and non-target class. After that, we implement the bidirectional distillation, optimized with decoupled relative-entropy loss: l o c a l t k p r i v a t e t k and p r i v a t e t k l o c a l t k . Finally, the global is updated by the weighted aggregate l o c a l t + 1 k . Each client pushes its trained local model to the server, which weights and averages these local models to obtain a new global model. The whole process is repeated until the global model converges.
As illustrated in Algorithm 1, our proposed BDD-HFL shares the logits (soft labels) of private samples rather than model parameters. This design allows the federated distillation framework to support heterogeneous local models and reduces the risk of white-box privacy attacks as supported by the findings in [38,39]. The bidirectional distillation process can also be interpreted as introducing noise to the global parameters during uploading and downloading, thereby enhancing privacy protection. Moreover, the distillation occurs dynamically between the local client and the private model without involving the server. This separation breaks the connection between the server and the private model, further increasing the difficulty of attacks that rely on shared model parameters.    
Algorithm 1: BDD-HFL
1 
Server:
2 
Initialize the global model: g l o b a l , total number of clients: K, communication rounds T
3 
for each round t = 1, 2, …, T do
Entropy 26 00762 i001
10 
end (
11 
ClientUpdate:
12 
Initialize the local model l o c a l t k and the private model p r i t k , the local epoch: E,
13 
distributing: l o c a l t k g l o b a l
14 
for each epoch e = 1, 2, …, E do
Entropy 26 00762 i002
19 
end
20 
Get l o c a l t + 1 k and upload it to sever

4. Results

In this section, we focus on verifying the effectiveness of the BDD-HFL method across various existing federated learning methods, as well as testing the performance on different natural datasets. In addition, we compare the experimental results of one-way distillation KD, bidirectional distillation FML, and our method on some baselines, for instance, FedAvg [2], FedProx [19], FedDyn [21], FedDC [22], and FedDisco [23].
In the above, we mainly evaluate the trustworthiness of our proposed method by evaluating its robustness [40] and fairness [41] in different heterogeneous scenarios. To evaluate robustness, we examine the method’s generalization capabilities in Heterogeneous Federated Learning (HFL) under various scenarios, including IID, Non-IID, and unbalanced data distributions, using average classification accuracy as a metric. This evaluation is crucial because each client possesses different local data heterogeneity, and the aggregated global model may struggle to capture individual data patterns effectively. In contrast, we assess fairness in HFL by evaluating client selection with accuracy metrics. Here, fairness is measured by considering scenarios where only a fraction of clients are selected to participate in the training process of each round. This evaluation ensures that the method performs equitably across different clients, regardless of their selection frequency or data distribution.

4.1. Dataset

We conduct comparisons across three datasets, including the natural datasets cafir10, cafir100 [42], and MNIST [43]. We follow the same experimental setting with FedDC [22]. We classify the data into two primary distributions: IID and Non-IID. In the IID setting, the training data are evenly and randomly distributed across all clients, with each category equally distributed among clients. Conversely, in the Non-IID setting, we divide the data into two distributions according to the Dirichlet distribution [44]: D1 = 0.6 and D2 = 0.3. D2 exhibits greater heterogeneity compared to D1. Respectively, we adopt LeNet convolutional neural networks [43] for CIFAR-10 and CIFAR-100 datasets, and a fully connected network (FCN) [45] for the MNIST dataset.

4.2. Baseline and Hyper-Parameter Settings

Furthermore, we evaluate the efficacy of our method by integrating it with various federated learning methods, including FedAvg [2], FedProx [19], FedDyn [21], FedDC [22], and FedDisco [23]. These methods primarily address the issue of client-side model drift caused by data heterogeneity, but they do not explicitly consider the problem of insufficient personalization of local models due to the aggregation-induced catastrophic forgetting of knowledge. Hence, we validate the effectiveness of our approach across these five baselines.
Most parameters continue to extend the settings of the FedDC, with the difference that we add decoupled bidirectional knowledge distillation between the private and local models during the learning process, and the local model interacts with the server model, while the private model is kept locally from the beginning to the end. We initialize the private model, local model, and server-side model to be identical. For the distillation hyperparameters, we set λ = 0.5 and μ = 0.5 for both the local model and private model. Additionally, to emphasize non-target class knowledge, we set the distillation weight of the target class knowledge to α = 1.0 and the distillation weight of the non-target class knowledge to β = 8.0 . We utilize the SGD algorithm as the local optimizer for all methods. To ensure consistency, we set the batch size to 50 during the local training phase, with each round consisting of five local epochs. The initial learning rate is set to 0.1, with a decay rate of 0.998. Respectively, we run federated learning for 600 rounds on MNIST and 1000 rounds on CIFAR-10 and CIFAR-100 datasets. We set the number of clients to 500 and 100 and divide them into full participation and 15% partial participation. We set the penalty coefficient of FedDC and FedDisco as 0.01 on the CIFAR10 and CIFAR100 datasets and 0.1 on the MNIST datasets.

4.3. Results and Analysis

We integrate our BDD-HFL method into both classic federated learning (FL) models and the latest FL methods, conducting numerous comparative experiments to assess improvements in accuracy. Additionally, we design various degrees of data heterogeneity to evaluate the model’s robustness and applicability. By comparing the performance before and after incorporating the BDD-HFL method, we validate its effectiveness in enhancing model accuracy and personalization capabilities, with the primary validation being based on the final model’s performance on the server side. Furthermore, we compare our BDD-HFL method with bidirectional distillation FML and one-way distillation KD.
Better performance of BDD-HFL. We keep the local model personalized by adding a private model to the client and passing local knowledge to the local model during the iteration process. In Table 1 and Table 2, we present a comparison of the experimental results obtained by incorporating one-way distillation KD, bidirectional distillation DML, and our BDD-HFL method into the latest federated learning approaches, FedAvg, FedDyn, FedProx, FedDC, and FedDisco. It is found that among the three datasets, the addition of both the KD and DML methods impair the performance of FedDC, and the KD effect is worse, which is 84.80% and 85.39% for partial participation and full participation in setting one. In contrast, the addition of our BDD-HFL method is 86.83% and 86.65% for partial participation and full participation in setting one. As a result, it significantly enhances the performance of FedDC, particularly in scenarios with multiple heterogeneous effects across different datasets. For the four kinds of data distribution in Setting 1, our method, combined with FedDisco, achieves an accuracy of 86.35% on the setting of CIFAR10-IID, 85.82% on the setting of CIFAR10-D1, 84.88% on the setting of CIFAR10-D2, and 86.20% on the setting of CIFAR10-unbalance in 100 clients’ partial participation. These results maintain a similar accuracy, which proves that our method is robust to different data heterogeneities and data unbalance situations. We analyze that KD and FML do not necessarily enhance heterogeneous FL, and heterogeneous data characteristics need to be considered. Our method considers this issue by decoupling the logits values into target class and non-target class distillation, which are guided by decoupled relative-entropy loss. The relative-entropy loss is measured by assigning greater weight to non-target classes compared to target classes. By focusing on the importance of non-target class knowledge, local and private models in the client can learn multi-class information, enhancing the effectiveness of knowledge transfer. This approach effectively maintains local personalization while preserving and even improving the performance of the global model.
Generalization to client structure. In Table 1 and Table 2, we validate the results of the experiments in two different settings: partial participation and full participation of the clients. For partial participation setting (100 clients), we test three datasets, each with four different ways of distributing the data. On top of this, we conduct a comparative analysis of the efficacy between incorporating bidirectional distillation, one-way distillation, and our BDD-HFL approach, against traditional federated learning methods. The results show that BDD-HFL on cafir10 always shows the best performance and is better suited to traditional federated learning methods, while the bidirectional distillation learning approach shows worse results. Similarly, on cafir100, our method again proves to have better stability and applicability. For MNIST data, our method is still effective in FedDyn and FedAvg, and has no serious impact on the other methods, and it is worthwhile to say that, for the two heterogeneous conditions of MNIST-D1 and MNIST-D2, we can also show the effectiveness of our method in solving the problem of insufficient personalization caused by the heterogeneous data.
For full participation setting (500 clients), the results show that our BDD-HFL method is also effective in large-scale distributed settings, and our method also improves 0.98% on CIFAR10-IID and 3.93% on CIFAR100-IID in the latest federated learning method FedDC. We validate the experimental results in the case of full client participation and find that our method improves the performance of the model regardless of the total number of clients as well as the number of client participants. The experiments show that our method effectively compensates for the lack of local personalization in classical federated learning and improves the performance of the model. The unique feature of BDD-HFL is the introduction of a novel bidirectional distillation decoupling idea, which adds a private model locally so that the local model can learn more extra knowledge while approaching the optimization direction of the private model no matter how many clients there are, thus making up for the poor local performance caused by heterogeneous data on the client.
Robustness to client sampling. From Table 1, we find that, for the case of the full participation of 100 clients, the accuracy of our method added on the FedDC on CIFAR10-IID is 86.65%, but for the case of the partial participation of 100 clients, the accuracy rises to 86.83%. Combined with the results of the other experiments in Table 2, we can see that the decision on the amount of client participation does not have a great impact on the experimental results, and the partial participation of the clients can still maintain a similar level of accuracy to the full participation of the clients. Therefore, based on the above performance, our method does not affect the elasticity of the client sampling, which shows our method performs well on different types and numbers of clients. This result proves the robustness and fairness of our method. It is only through the private model that the network will refer to the experience of the private model network to adjust its learning during the training process, and ultimately be able to converge to a smoother point of the smallest value so that it has better generalization performance, thus improving the performance of the global model, and it will not be interrupted due to the interruption of the training of the model.
Convergence on heterogeneous data. Due to the characteristics of wide data distribution, the model convergence speed is often slow. In Figure 3, we compare the convergence graphs (a and b, and d and e) to illustrate the different convergence speeds and model accuracies for various data distributions. The results indicate that convergence under IID data distribution is faster than under Non-IID conditions, and model accuracy is also higher with IID data.
When applying our method, BDD-HFL, on top of FedDC, we observe significant improvements in different scenarios of data heterogeneity and client numbers. Our method consistently outperforms other methods not utilizing BDD-HFL in terms of both convergence speed and model accuracy. This includes superior performance compared to FML added to FedDC. Even in heterogeneous setups, the convergence graphs clearly visualize the enhancements our method brings, demonstrating its effectiveness in improving both convergence speed and model accuracy.

5. Conclusions

In this work, we address the issue of insufficient local personalization caused by data heterogeneity by proposing Bidirectional Decoupled Distillation For Heterogeneous Federated Learning (BDD-HFL). BDD-HFL leverages the interaction between the local model and a private model through decoupled relative-entropy knowledge distillation, enabling a more flexible learning process. Extensive experiments on various categorized datasets demonstrate that our BDD-HFL approach outperforms other federated learning methods.
Limitations. Our method integrates an additional private model into each client, and knowledge distillation approaches in FL typically assume that both the private and local models have access to the same local training data. This assumption increases the risk of information leakage and adds to the computational cost during the training phase. In contrast, Data-Free Knowledge Distillation (DFKD) transfers knowledge by generating synthetic data, thereby eliminating the need to access the original data. By further analyzing DFKD strategies, the above issues may be tackled, which is left as our future work. Furthermore, heterogeneous scenarios involve heterogeneous multi-task clients or heterogeneous client architectures. In the future, we will explore various federated learning scenarios to evaluate our proposed method.

Author Contributions

L.H. provided Methodology, Resources, Validation; W.S. provided writing, experiments and revision; M.Y. provided data curation, formal analysis, validation; X.L. provided visualization, writing and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (No. 62202015), the University Synergy Innovation Program of Anhui Province (GXXT-2022-052), R&D Program of Beijing Municipal Education Commission (KM202310011013).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

We use public CIFAR datasets to evaluate the performance of proposed method. The CIFAR-10 and CIFAR-100 public datasets can be freely downloaded from the web page at https://www.cs.toronto.edu/~kriz/cifar.html, accessed on 7 October 2022. The MNIST dataset can be freely downloaded from the website https://opendatalab.com/MNIST, accessed on 12 June 2022. The source code of our BDD-HFL approach is available at https://github.com/drafly/BDKFD.git.

Conflicts of Interest

There are no competing interests to declare.

References

  1. Liu, B.; Ding, M.; Shaham, S.; Rahayu, W.; Farokhi, F.; Lin, Z. When machine learning meets privacy: A survey and outlook. ACM Comput. Surv. 2021, 54, 1–36. [Google Scholar] [CrossRef]
  2. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Artificial Intelligence and Statistics; PMLR: Cambridge, MA, USA, 2017; pp. 1273–1282. [Google Scholar]
  3. Kaissis, G.A.; Makowski, M.R.; Rückert, D.; Braren, R.F. Secure, privacy-preserving and federated machine learning in medical imaging. Nat. Mach. Intell. 2020, 2, 305–311. [Google Scholar] [CrossRef]
  4. Husnoo, M.A.; Anwar, A.; Hosseinzadeh, N.; Islam, S.N.; Mahmood, A.N.; Doss, R. Fedrep: Towards horizontal federated load forecasting for retail energy providers. In Proceedings of the 2022 IEEE PES 14th Asia-Pacific Power and Energy Engineering Conference (APPEEC), Melbourne, Australia, 20–23 November 2022; pp. 1–6. [Google Scholar]
  5. Fallah, A.; Mokhtari, A.; Ozdaglar, A. Personalized federated learning with theoretical guarantees: A model-agnostic meta-learning approach. Adv. Neural Inf. Process. Syst. 2020, 33, 3557–3568. [Google Scholar]
  6. Sun, B.; Huo, H.; Yang, Y.; Bai, B. Partialfed: Cross-domain personalized federated learning via partial initialization. Adv. Neural Inf. Process. Syst. 2021, 34, 23309–23320. [Google Scholar]
  7. Luo, J.; Wu, S. Adapt to adaptation: Learning personalization for cross-silo federated learning. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22), Vienna, Austria, 23–29 July 2022; pp. 2166–2173. [Google Scholar]
  8. Zhang, J.; Hua, Y.; Wang, H.; Song, T.; Xue, Z.; Ma, R.; Guan, H. Fedala: Adaptive local aggregation for personalized federated learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 11237–11244. [Google Scholar]
  9. Wang, Y.; Fu, H.; Kanagavelu, R.; Wei, Q.; Liu, Y.; Goh, R.S.M. An aggregation-free federated learning for tackling data heterogeneity. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 26233–26242. [Google Scholar]
  10. Li, D.; Wang, J. Fedmd: Heterogenous federated learning via model distillation. arXiv 2019, arXiv:1910.03581. [Google Scholar]
  11. Shen, T.; Zhang, J.; Jia, X.; Zhang, F.; Lv, Z.; Kuang, K.; Wu, C.; Wu, F. Federated mutual learning: A collaborative machine learning method for heterogeneous data, models, and objectives. Front. Inf. Technol. Electron. Eng. 2023, 24, 1390–1402. [Google Scholar] [CrossRef]
  12. T Dinh, C.; Tran, N.; Nguyen, J. Personalized federated learning with moreau envelopes. Adv. Neural Inf. Process. Syst. 2020, 33, 21394–21405. [Google Scholar]
  13. Li, T.; Hu, S.; Beirami, A.; Smith, V. Ditto: Fair and robust federated learning through personalization. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 6357–6368. [Google Scholar]
  14. Diao, E.; Ding, J.; Tarokh, V. HeteroFL: Computation and Communication Efficient Federated Learning for Heterogeneous Clients. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
  15. An, P.; Wang, Z.; Zhang, C. Ensemble unsupervised autoencoders and Gaussian mixture model for cyberattack detection. Inf. Process. Manag. 2022, 59, 102844. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Xiang, T.; Hospedales, T.M.; Lu, H. Deep mutual learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4320–4328. [Google Scholar]
  17. Zhao, B.; Cui, Q.; Song, R.; Qiu, Y.; Liang, J. Decoupled knowledge distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11953–11962. [Google Scholar]
  18. Li, T.; Sahu, A.K.; Talwalkar, A.; Smith, V. Federated learning: Challenges, methods, and future directions. IEEE Signal Process. Mag. 2020, 37, 50–60. [Google Scholar] [CrossRef]
  19. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated optimization in heterogeneous networks. Proc. Mach. Learn. Syst. 2020, 2, 429–450. [Google Scholar]
  20. Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.; Stich, S.; Suresh, A.T. Scaffold: Stochastic controlled averaging for federated learning. In Proceedings of the International Conference on Machine Learning, Virtual, 12–18 July 2020; pp. 5132–5143. [Google Scholar]
  21. Durmus, A.E.; Yue, Z.; Ramon, M.; Matthew, M.; Paul, W.; Venkatesh, S. Federated learning based on dynamic regularization. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
  22. Gao, L.; Fu, H.; Li, L.; Chen, Y.; Xu, M.; Xu, C.Z. Feddc: Federated learning with non-iid data via local drift decoupling and correction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10112–10121. [Google Scholar]
  23. Ye, R.; Xu, M.; Wang, J.; Xu, C.; Chen, S.; Wang, Y. Feddisco: Federated learning with discrepancy-aware collaboration. In Proceedings of the International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; pp. 39879–39902. [Google Scholar]
  24. Tuor, T.; Wang, S.; Ko, B.J.; Liu, C.; Leung, K.K. Overcoming noisy and irrelevant data in federated learning. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Virtual, 10–15 January 2021; pp. 5020–5027. [Google Scholar]
  25. Yoshida, N.; Nishio, T.; Morikura, M.; Yamamoto, K.; Yonetani, R. Hybrid-FL for wireless networks: Cooperative learning mechanism using non-IID data. In Proceedings of the ICC 2020–2020 IEEE International Conference On Communications (ICC), Dublin, Ireland, 7–11 June 2020; pp. 1–7. [Google Scholar]
  26. Mothukuri, V.; Parizi, R.M.; Pouriyeh, S.; Huang, Y.; Dehghantanha, A.; Srivastava, G. A survey on security and privacy of federated learning. Future Gener. Comput. Syst. 2021, 115, 619–640. [Google Scholar] [CrossRef]
  27. Wang, T.; Zhu, J.Y.; Torralba, A.; Efros, A.A. Dataset distillation. arXiv 2018, arXiv:1811.10959. [Google Scholar]
  28. Chen, H.; Wang, Y.; Xu, C.; Yang, Z.; Liu, C.; Shi, B.; Xu, C.; Xu, C.; Tian, Q. Data-free learning of student networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3514–3522. [Google Scholar]
  29. Chen, M.; Poor, H.V.; Saad, W.; Cui, S. Convergence time optimization for federated learning over wireless networks. IEEE Trans. Wirel. Commun. 2020, 20, 2457–2471. [Google Scholar] [CrossRef]
  30. Yang, Z.; Chen, M.; Saad, W.; Hong, C.S.; Shikh-Bahaei, M. Energy efficient federated learning over wireless communication networks. IEEE Trans. Wirel. Commun. 2020, 20, 1935–1949. [Google Scholar] [CrossRef]
  31. Shinde, S.S.; Tarchi, D. Joint air-ground distributed federated learning for intelligent transportation systems. IEEE Trans. Intell. Transp. Syst. 2023, 24, 9996–10011. [Google Scholar] [CrossRef]
  32. Shinde, S.S.; Bozorgchenani, A.; Tarchi, D.; Ni, Q. On the design of federated learning in latency and energy constrained computation offloading operations in vehicular edge computing systems. IEEE Trans. Veh. Technol. 2021, 71, 2041–2057. [Google Scholar] [CrossRef]
  33. Wang, K.; Mathews, R.; Kiddon, C.; Eichner, H.; Beaufays, F.; Ramage, D. Federated evaluation of on-device personalization. arXiv 2019, arXiv:1910.10252. [Google Scholar]
  34. Deng, Y.; Kamani, M.M.; Mahdavi, M. Adaptive Personalized Federated Learning. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 3–5 May 2021. [Google Scholar]
  35. Huang, Y.; Chu, L.; Zhou, Z.; Wang, L.; Liu, J.; Pei, J.; Zhang, Y. Personalized cross-silo federated learning on non-iid data. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 7865–7873. [Google Scholar]
  36. Hinton, G. Distilling the Knowledge in a Neural Network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  37. Lin, T.; Kong, L.; Stich, S.U.; Jaggi, M. Ensemble distillation for robust model fusion in federated learning. Proc. Adv. Neural Inf. Process. Syst. 2020, 33, 2351–2363. [Google Scholar]
  38. Huang, D.; Ye, X.; Sakurai, T. Knowledge distillation-based privacy-preserving data analysis. In Proceedings of the Conference on Research in Adaptive and Convergent Systems, Virtual, 3–6 October 2022; pp. 15–20. [Google Scholar]
  39. Faisal, F.; Leung, C.K.; Mohammed, N.; Wang, Y. Privacy-Preserving Learning via Data and Knowledge Distillation. In Proceedings of the 2023 IEEE 10th International Conference on Data Science and Advanced Analytics (DSAA), Thessaloniki, Greece, 9–12 October 2023; pp. 1–10. [Google Scholar]
  40. Sánchez, P.M.; Huertas, A.; Xie, N.; Bovet, G.; Martinez Perez, G.; Stiller, B. FederatedTrust: A solution for trustworthy federated learning. Future Gener. Comput. Syst. 2024, 152, 83–98. [Google Scholar] [CrossRef]
  41. Zhang, Y.; Zeng, D.; Luo, J.; Fu, X.; Chen, G.; Xu, Z.; King, I. A Survey of Trustworthy Federated Learning: Issues, Solutions, and Challenges. ACM Trans. Intell. Syst. Technol. 2024, 2157–6904. [Google Scholar] [CrossRef]
  42. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images; Technical Report TR-2009; University of Toronto: Toronto, ON, Canada, 2009. [Google Scholar]
  43. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  44. Lin, J. On the Dirichlet Distribution. Master’s Thesis, Department of Mathematics and Statistics, Queens University, Kingston, ON, Canada, September 2016. [Google Scholar]
  45. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
Figure 1. Changes in optimization direction of models (server, client, and private) for distillation methods in Heterogeneous Federated Learning and normal federated learning. The server is constantly updated toward optimal point x * (red square), the client toward optimal point x 1 * (blue square), and the private model toward optimal point x p 1 * (green square). The server and the client start from the same starting point x of the function f ( x ) , and the private model starts from point x p 1 . (a) In standard federated learning, cross-entropy optimization is applied without any knowledge exchange among clients. This approach results in a loss of personalization due to data heterogeneity across different clients. (b) The one-way distillation method, which utilizes relative-entropy loss, allows the local model to distill knowledge from the private model. However, this approach causes the private model to lose access to global information from the aggregated server model, impairing its generalization capability. (c) In contrast, bidirectional distillation with decoupled relative-entropy optimization enables mutual knowledge transfer between the private and local models. This approach allows both models to learn global information, thereby maintaining generalization while preserving local personalization.
Figure 1. Changes in optimization direction of models (server, client, and private) for distillation methods in Heterogeneous Federated Learning and normal federated learning. The server is constantly updated toward optimal point x * (red square), the client toward optimal point x 1 * (blue square), and the private model toward optimal point x p 1 * (green square). The server and the client start from the same starting point x of the function f ( x ) , and the private model starts from point x p 1 . (a) In standard federated learning, cross-entropy optimization is applied without any knowledge exchange among clients. This approach results in a loss of personalization due to data heterogeneity across different clients. (b) The one-way distillation method, which utilizes relative-entropy loss, allows the local model to distill knowledge from the private model. However, this approach causes the private model to lose access to global information from the aggregated server model, impairing its generalization capability. (c) In contrast, bidirectional distillation with decoupled relative-entropy optimization enables mutual knowledge transfer between the private and local models. This approach allows both models to learn global information, thereby maintaining generalization while preserving local personalization.
Entropy 26 00762 g001
Figure 2. Each client in the BDD-HFL method contains two models: the local model and the private model, where the local model replicates the global model, and the private model is responsible for multiple rounds of decoupled Deep Mutual Learning with the local model. The grey line denotes bidirectional decoupled distillation, blue line denotes local update, and green line denotes server update. In scenarios where clients have heterogeneous data distributions, the logits of both local and private models are decoupled into target and non-target classes during local training. Bidirectional decoupled distillation facilitates mutual learning between the local and private models by allowing them to distill information from each other, focusing separately on target and non-target classes. This process refines both models. After each major round of local training, the refined local models are aggregated and uploaded to the server.
Figure 2. Each client in the BDD-HFL method contains two models: the local model and the private model, where the local model replicates the global model, and the private model is responsible for multiple rounds of decoupled Deep Mutual Learning with the local model. The grey line denotes bidirectional decoupled distillation, blue line denotes local update, and green line denotes server update. In scenarios where clients have heterogeneous data distributions, the logits of both local and private models are decoupled into target and non-target classes during local training. Bidirectional decoupled distillation facilitates mutual learning between the local and private models by allowing them to distill information from each other, focusing separately on target and non-target classes. This process refines both models. After each major round of local training, the refined local models are aggregated and uploaded to the server.
Entropy 26 00762 g002
Figure 3. Convergence plots of BDD-HFL plus on FedDC and DML plus on FedDC and other baselines in different settings for the CIFAR10 and CIFAR100 datasets with IID, D2 Non-IID (Dirichlet-0.3), and imbalanced data with 100 clients participating. (ac) represent partial participation in CIFAR10 training. (df) represent full participation in CIFAR 100 training.
Figure 3. Convergence plots of BDD-HFL plus on FedDC and DML plus on FedDC and other baselines in different settings for the CIFAR10 and CIFAR100 datasets with IID, D2 Non-IID (Dirichlet-0.3), and imbalanced data with 100 clients participating. (ac) represent partial participation in CIFAR10 training. (df) represent full participation in CIFAR 100 training.
Entropy 26 00762 g003
Table 1. Top 1 test accuracy (%) for IID, Non-IID, and unbalanced data with 100 and 500 clients partially engaged (15%). The comparison method is to add KD, DML, and BDD-HFL implemented with FedAvg [2], FedProx [19], FedDyn [21], FedDC [22], and FedDisco [23] five baselines. Bold numbers indicate the best results.
Table 1. Top 1 test accuracy (%) for IID, Non-IID, and unbalanced data with 100 and 500 clients partially engaged (15%). The comparison method is to add KD, DML, and BDD-HFL implemented with FedAvg [2], FedProx [19], FedDyn [21], FedDC [22], and FedDisco [23] five baselines. Bold numbers indicate the best results.
MethodsFedAvg+KD+FML+BDD-HFLFedProx+KD+FML+BDD-HFLFedDC+KD+FML+BDD-HFLFedDyn+KD+FML+BDD-HFLFedDisco+KD+FML+BDD-HFL
Setting 1100 clients partial participation
CIFAR10-IID81.6780.7981.2682.4182.1681.7881.5782.0285.7184.8084.9486.8384.5083.6183.9685.2084.8185.0685.2886.35
CIFAR10-D181.0580.2580.7782.3481.3280.6580.8182.3184.7784.2684.3385.9084.1082.5583.1784.6084.4084.3684.3885.82
CIFAR10-D279.7779.0380.1981.6779.8479.5779.6282.2384.5882.9783.5685.0382.3081.3382.3783.3383.4982.9783.1784.88
CIFAR10-unbalance81.6881.4381.5381.8781.8881.4181.2481.4685.3584.6484.9485.8184.3084.1884.3085.1284.9084.7684.8486.20
CIFAR100-IID40.8042.0942.1144.4740.6740.5141.9142.9255.4054.5354.5656.9451.2050.0250.7452.1354.0454.0454.5157.37
CIFAR100-D141.7641.7842.4845.7841.8340.6842.2347.5054.6554.0154.4657.0051.7548.8350.1951.5053.8453.6254.2557.04
CIFAR100-D241.8142.1342.9847.2441.8441.1242.5749.9853.9152.8953.1156.3951.1347.7149.4450.6453.9353.4453.6756.39
CIFAR100-unbalance40.9043.1942.1643.2741.0541.3340.5542.9455.2753.7253.7956.4551.0150.1651.0350.3954.2154.1554.6756.80
MNIST-IID98.1598.1297.9998.2098.1198.1998.1998.0898.4798.4298.4598.4998.3898.3598.3098.3998.3998.5398.4298.48
MNIST-D198.1398.0597.9998.2098.1298.1998.1498.2298.4998.4598.4298.4398.3098.2698.2198.3598.3698.4998.4098.46
MNIST-D298.0097.9198.0398.2198.0498.0298.0198.0598.4098.4098.3598.4698.3098.2598.1298.4498.2898.4198.4698.47
MNIST-unbalance98.1598.1698.0298.1698.1398.1098.1898.1098.5398.4998.4198.1698.3498.3098.3198.3298.3798.4298.4098.56
Setting 2500 clients partial participation
CIFAR10-IID73.2673.0973.2675.7472.5872.6672.9075.0384.1982.2681.7085.1782.4981.1281.1682.7774.3479.8479.3681.78
CIFAR100-IID27.3628.4627.9327.3626.5027.2327.4724.5250.6154.5343.3854.5444.1144.4844.2540.5841.1736.9043.8053.02
Table 2. Top 1 test accuracy (%) for IID, Non-IID, and unbalanced data with 100 and 500 clients fully engaged. The comparison method is to add KD, DML, and BDD-HFL implemented with FedAvg [2], FedProx [19], FedDyn [21], FedDC [22], and FedDisco [23] five baselines. Bold numbers indicate the best results.
Table 2. Top 1 test accuracy (%) for IID, Non-IID, and unbalanced data with 100 and 500 clients fully engaged. The comparison method is to add KD, DML, and BDD-HFL implemented with FedAvg [2], FedProx [19], FedDyn [21], FedDC [22], and FedDisco [23] five baselines. Bold numbers indicate the best results.
MethodsFedAvg+KD+FML+BDD-HFLFedProx+KD+FML+BDD-HFLFedDC+KD+FML+BDD-HFLFedDyn+KD+FML+BDD-HFLFedDisco+KD+FML+BDD-HFL
Setting 1100 clients all participation
CIFAR10-IID82.1681.8481.5382.7681.8581.7681.8082.5886.1885.3985.3886.6585.2682.4884.5985.8485.4985.0685.7187.07
CIFAR10-D180.4280.8980.8682.3680.7080.4380.7582.8385.6485.0985.3086.3685.2684.1784.5085.6985.1185.1685.4985.75
CIFAR10-D279.1480.8280.9281.8240.9340.2741.5648.4484.3283.5983.6984.8384.1483.3782.9784.1783.9884.3084.1384.95
CIFAR10-unbalance86.3181.8081.6782.4981.9081.5881.7182.5686.3185.4185.7486.8585.6884.5084.9386.1585.5585.8485.8187.02
CIFAR100-IID39.6841.8141.7043.1540.3938.7939.5842.9455.5253.9954.6157.1152.0751.6851.4653.5353.5754.6754.3656.95
CIFAR100-D140.4842.0942.3844.0740.1540.2440.8445.5555.3453.7654.6657.9452.8451.6251.8452.7654.6853.4355.1357.32
CIFAR100-D240.1143.1443.2743.5640.9340.2741.5648.4454.8653.7654.6356.8951.8950.4050.6753.0153.9053.8354.1557.16
CIFAR100-unbalance40.0343.3142.2544.1239.9340.0840.7641.8255.6953.8754.5557.0652.8151.7052.1752.5554.1854.7754.2457.15
MNIST-IID98.1298.4998.5098.4598.1298.2198.1998.1498.4598.4298.4498.4598.5198.7298.6698.6398.3398.4698.4298.48
MNIST-D198.0998.4398.4898.3198.0598.1598.1198.1598.4898.4598.4598.5398.4498.7298.6798.5598.3998.4498.4698.49
MNIST-D297.9898.3898.3598.3997.9698.0698.1298.1098.5198.4498.4598.5898.4698.7098.5798.6498.4098.4998.5298.46
MNIST-unbalance98.1298.4998.3798.3598.1098.2698.3198.1898.4698.4898.4398.4898.6098.6298.7398.6998.4398.4698.4698.48
Setting 2500 clients all participation
CIFAR10-IID73.4372.8673.0175.6372.7771.5772.5375.1784.9385.3983.9386.3684.0783.3883.4384.7382.2082.0684.2585.40
CIFAR100-IID26.0327.9527.7527.8628.2226.4827.5327.0454.2538.8647.8956.2650.2251.9551.7551.8050.0538.6049.4656.38
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, W.; Yan, M.; Li, X.; Han, L. Bidirectional Decoupled Distillation for Heterogeneous Federated Learning. Entropy 2024, 26, 762. https://doi.org/10.3390/e26090762

AMA Style

Song W, Yan M, Li X, Han L. Bidirectional Decoupled Distillation for Heterogeneous Federated Learning. Entropy. 2024; 26(9):762. https://doi.org/10.3390/e26090762

Chicago/Turabian Style

Song, Wenshuai, Mengwei Yan, Xinze Li, and Longfei Han. 2024. "Bidirectional Decoupled Distillation for Heterogeneous Federated Learning" Entropy 26, no. 9: 762. https://doi.org/10.3390/e26090762

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop