Next Article in Journal
Collaborative Reliable Event Transport Based on Mobile-Assisted Sensing in Urban Digital Twin
Next Article in Special Issue
Force Tracking Control of Functional Electrical Stimulation via Hybrid Active Disturbance Rejection Control
Previous Article in Journal
Infrared Small-Target Detection Using Multiscale Local Average Gray Difference Measure
Previous Article in Special Issue
Self-Oxygen Regulator System for COVID-19 Patients Based on Body Weight, Respiration Rate, and Blood Saturation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Ring-Topology Decentralized Federated Learning with Deep Generative Models for Medical Data in eHealthcare Systems

1
School of Public Affairs, Zhejiang University, Hangzhou 310027, China
2
Zhejiang Post & Telecommunication Construction Co., Ltd., Hangzhou 310016, China
3
Inria, 78150 Le Chesnay-Rocquencourt, France
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(10), 1548; https://doi.org/10.3390/electronics11101548
Submission received: 18 April 2022 / Revised: 6 May 2022 / Accepted: 8 May 2022 / Published: 12 May 2022
(This article belongs to the Special Issue Recent Advanced Applications of Rehabilitation and Medical Robotics)

Abstract

:
By leveraging deep learning technologies, data-driven-based approaches have reached great success with the rapid increase of data generated for medical applications. However, security and privacy concerns are obstacles for data providers in many sensitive data-driven scenarios, such as rehabilitation and 24 h on-the-go healthcare services. Although many federated learning (FL) approaches have been proposed with DNNs for medical applications, these works still suffer from low usability of data due to data incompleteness, low quality, insufficient quantity, sensitivity, etc. Therefore, we propose a ring-topology-based decentralized federated learning (RDFL) scheme for deep generative models (DGM), where DGM is a promising solution for solving the aforementioned data usability issues. Our RDFL schemes provide communication efficiency and maintain training performance to boost DGMs in target tasks compared with existing FL works. A novel ring FL topology and a map-reduce-based synchronizing method are designed in the proposed RDFL to improve the decentralized FL performance and bandwidth utilization. In addition, an inter-planetary file system (IPFS) is introduced to further improve communication efficiency and FL security. Extensive experiments have been taken to demonstrate the superiority of RDFL with either independent and identically distributed (IID) datasets or non-independent and identically distributed (Non-IID) datasets.

1. Introduction

Recent years have witnessed a rapid growth of deep learning (DL) algorithms widely used to solve data-driven industrial problems in real-world medical applications [1,2]. These deep learning methods are benefited a lot by the massive amount of data collected. To improve the DL-based products, it brings great demand for different entities, e.g., huge amounts of devices belonging to different patients/hospitals, to contribute their data and train models together. In such collaborative training, the collection of massive data for centralized training causes serious privacy threats [3], which motivates federated learning (FL) [4] allowing participants to learn the model collaboratively by only synchronizing local-trained model parameters without revealing their original data.
A general federated learning system usually uses a central parameter server to coordinate the large federation of participating nodes (nodes, clients and workers are used interchangeably in this manuscript). For instance, the Conventional FL framework [5,6] uses a highly centralized architecture where a centralized node collects gradients or model parameters from data nodes to update the global model. Although some FL approaches have been proposed [7,8,9,10], the model training performance always suffers from low usability of medical data, such as data incompleteness, low quality, insufficient quantity and sensitivity. Deep generative models (DGM) like the generative adversarial network (GAN) can be used to tackle the problems mentioned above. In order to meet the data privacy constraints, distributed GAN algorithms are proposed [11]. Large communication bandwidth among nodes is required in current distributed GAN algorithms [12,13], while an intermediary is required to ensure convergence due to its architectures, which separate generators from discriminators. However, the communication bandwidth could be limited and costly in many real-world applications [14]. The center node in the current FL framework suffers from communication pressure and communication bandwidth bottleneck [15,16]. Communication-efficient distributed GAN is still an open problem, and we propose a framework that places local discriminators with local generators and synchronizes occasionally.
Additionally, the aforementioned centralized FL frameworks could bring security concerns and suffer the risk of single-point failure. Through the literature review, the decentralized FL framework [17,18,19] has been proposed. The decentralized FL framework removes the centralized node and synchronizes FL updates among the data nodes, then performs aggregation. However, it still faces challenges in communication pressure and cost, especially when blockchain is employed as an effective decentralized storage and replaces the central FL servers [18,20]. In addition, it is important to design the aggregation algorithm used in the decentralized FL framework that can achieve competitive performance under the situation of data poisoning from malicious nodes.
To tackle the aforementioned problems, a ring-topology decentralized federated learning (RDFL) framework is proposed in this paper. RDFL aims to provide communication-efficient learning across multiple data sources in a decentralized structure, which is also subject to privacy constraints. Inspired by the idea of ring-allreduce (https://andrew.gibiansky.com/blog/machine-learning/baidu-allreduce/) which is accessed on 21 February 2017, consistent hashing technique [21] is employed in the proposed RDFL to construct a ring topology of decentralized nodes, which is able to reduce the communication pressure and improve topology stability. Besides, an innovative model synchronizing method is also designed in RDFL to benefit the bandwidth utilization and decentralized FL performance. Additionally, an InterPlanetary File System (IPFS) [22] based data sharing scheme is also designed to further improve communication efficiency and reduce communication costs. The code of RDFL will be published online soon.
To sum up, the main contributions of the proposed RDFL are as follows:
(1)
A new data node topology mechanism for decentralized FL has been designed in this work. The proposed mechanism is able to reduce communication pressure and significantly improve system stability. To the best of our knowledge, this is the first attempt to conduct a data node topology design for communication-efficient decentralized FL.
(2)
A novel ring decentralized federated learning (RDFL) synchronizing method is designed to improve bandwidth utilization and training stability.
(3)
To improve the communication performance and security of the decentralized FL framework, an IPFS-based data sharing scheme is designed to reduce system communication pressure and cost.

2. Related Work

This work relates to two literature, federated learning and distributed/federated GAN.
Federated learning has emerged as a new paradigm in a distributed machine learning setup [4] and became widespread by Google’s blog post (https://ai.googleblog.com/2017/04/federated-learning-collaborative.html) that is accessed in 06 April 2017. It [4] proposes an FL process that collects locally calculated gradients and aggregates them at the central node. To help build FL tasks, some centralized FL frameworks have been proposed. Representatives of these frameworks are FATE (https://fate.fedai.org/), TensorFlow-Federated (TFF) (https://www.tensorflow.org/federated), PaddleFL (https://github.com/PaddlePaddle/PaddleFL), LEAF [6] and PySyft [5]. However, these centralized FL frameworks still have the problem of security concern, communication bottleneck and stability.
To avoid the problems caused by the centralized FL framework, the research on the decentralized FL framework has attracted much attention. In [19], it has proposed a decentralized FL algorithm based on the Gossip algorithm and model segmentation. Local models are propagated over a peer-to-peer network through sum-weight gossip. Roy et al. have proposed a peer-to-peer decentralized FL algorithm. Lalitha et al. have explored a fully decentralized FL algorithm [23]. A blockchain-based decentralized FL framework is presented in [18]. To overcome the communication problem of the decentralized FL framework, current research focuses on researching novel communication compression or model compression techniques to reduce the communication pressure. For instance, Hu et al. utilize the gossip algorithm to improve bandwidth utilization and model segmentation to reduce communication pressure [19]. Amiri et al. [24] and Konečný el al. [25] propose model quantification methods to reduce communication pressure. Tang et al. [26] and Koloskova et al. [27] introduce communication compression methods to reduce communication pressure. Besides, sharing datasets [28] and knowledge distillation [29,30] are employed in FL to improve the FL performance on Non-IID datasets. To further protect the data privacy of FL, existing research focuses on several defense methods, including differential privacy [31] and multi-party secure computing (MPC) [32]. There are also reports on applying blockchain technology to decentralized FL to improve the security [7,18,33]. Distributed GANs have been proposed recently [34,35]. Moreover, they propose a single generator at the intermediary and distributed discriminators in [34]. A gossip approach for distributed GAN that does not require an intermediary server is presented in [35]. In order to deal with non-iid data, an individual discriminator is trained separately while the centralized generator is updated to fool the weakest discriminator in [11]. All of the above works require large communications during training. Few attempts have been made to address the problem of GAN training in an FL way [36,37,38], while little attention has been paid on improving communication efficiency.

3. The Proposed RDFL

In this section, we first describe how the designed topology mechanism in RDFL that utilizes a consistent hashing algorithm to build a ring decentralized FL topology for FL nodes. Then, we describe the synchronizing method in RDFL. Finally, an IPFS-based data sharing scheme is presented to further reduce communication cost.

3.1. Ring Decentralized FL Topology

Topology Overview Consider a group of n data nodes among which there are m trusted data nodes and n m untrusted data nodes. These n data nodes are represented by the symbol D = { D P 1 , D P 2 , D P 3 , , D P n } . RDFL utilizes a consistent hashing algorithm to construct a ring topology of n data nodes. The consistent hash value H k = H a s h ( D P k i p ) [ 0 , 2 32 1 ] , D P k i p represents the ip of D P k , k [ 1 , n ] . Data nodes are distributed on the ring with an index value range [ 0 , 2 32 1 ] according to the consistent hash value. Figure 1 shows the ring topology constructed by the consistent hashing algorithm.
Malicious Node The malicious nodes can be detected with committee election methods [18]. The malicious nodes will only send local models to the nearest trusted data node found with the proposed ring topology in a clockwise direction and will not be passed anymore. In Figure 1, the green data nodes represent trusted data nodes and the gray data nodes represent Untrusty data nodes. According to the clockwise principle, Untrusty data nodes D P 2 and D P 3 send models to the trusted data provider D P 4 . Untrusty data node D P 5 sends models to the nearest trusted data node D P k . With the help of a consistent hashing algorithm, different untrusty data nodes can only send models to their corresponding trusty nodes, which reduces the communication pressure of trusty node effectively.
In order to deal with continuous untrusty nodes, a possible solution is to make the distribution of trusty nodes on the ring uniform. Hence, virtual nodes of trusted nodes can also be added to the ring topology if needed, which aims to further reduce communication pressure. Figure 2 shows a ring topology with virtual nodes. The green nodes with red dashed lines represent virtual nodes. D P 1 v 1 is the virtual node of D P 1 .

3.2. RDFL Training

Synchronizing progress Based on the ring decentralized topology constructed with the consistent hash algorithm, the trusty node follows the synchronizing progress illustrated in Figure 3. M i represents the model of data node D P i . r represents the number of rounds to execute model synchronization and m represents the number of trusted nodes. At each synchronizing round, each node sends its models in a clockwise direction, then execute Federated Averaging (FedAvg) [4] to generate a new global model and starts the next iteration.
Training progress with GAN Models The horizon training iteration is denoted by T and the index time is denoted by t. Consider nodes D P i , i 1 , 2 , , N with local dataset R i for each node, the weight of node D P i is denoted by p i . Assume each node has local discriminator and generator with corresponding parameters d i and g i , loss function L D i and L G i , local stochastic gradients θ ˜ i ( d t i , g t i ) and h ˜ i ( d t i , g t i ) and learning rate l r d ( t ) and l r g ( t ) at time t. We assume the learning rates are the same across nodes. To improve the bandwidth utilization between trusted nodes, RDFL introduces the Ring-allreduce algorithm and the clockwise principle. As show in Figure 3, the trusted node D P 1 retains the local model M 1 after distillation, D P 2 retains M 2 and D P n retains M n . Then, the trusted nodes utilize the Ring-allreduce algorithm and the clockwise rule to synchronize the local models of the trusted nodes. After synchronization, all trusted nodes have local models of other trusted nodes. The detail of RDFL training is described in Algorithm 1.
Algorithm 1 RDFL training with generative adversarial network.
Input: Set training period T, synchronizing interval K. Initialize global discriminator and generator d 0 and g 0 . Initialize local discriminator and generator parameters d 0 i = d 0 , g 0 i = g 0 for all N nodes D P i , i { 1 , 2 , , N } .
Output: The New global model
Procedure: Data Node Executes
1:
for each FL round t = 1 , 2 , 3 , , T  do
2:
 Each node calculates local stochastic gradient θ ˜ t i and h ˜ t i corresponding to local discriminator and local generator respectively while fake generated by the local generator.
3:
 Each updates its local parameter in parallel;
d t i d t 1 i + l r d ( t ) θ ˜ t i g t i g t 1 i + l r g ( t ) h ˜ t i
4:
if  t mod K = 0  then
5:
  Malicious node detection
6:
  Each trusted node receive all trusty nodes’ model parameters through the ring;
7:
  B is the subset of N stands for trusty nodes
8:
  for D P i B executes global model parameters
d t j = 1 B p j d t j g t j = 1 B p j g t j
9:
  Each node updates its local model parameters with the executed global parameters;
10:
end if
11:
end for
It needs to be pointed out that we assume all nodes on the ring participate in the communication process. If part of the nodes meet communication failures during parameter sending, an extension work could be taken by following [25]. Due to the paragraph limitation, the proof of convergence for model averaging would not be listed here, a similar proof could be referred in [36], which conducts a centralized FL framework with GAN.

3.3. IPFS-Based Data Sharing Scheme

In the most decentralized FL work [7,17,18,19,33,36], we noticed that the model parameters are transferred among data nodes directly, which occupies a lot of communication overhead and could cause serious communication cost when the blockchain is employed. For instance, gas fee is required in popular blockchain Ethereum, where the cost could be significantly high for large models. In order to reduce the risk of communication costs, an IPFS-based data sharing scheme is designed in RDFL. Data files, e.g., model parameters, in IPFS would be divided into multiple pieces stored on different nodes and IPFS will generate the IPFS hash corresponding to the file. The IPFS hash is a 46 byte string and the corresponding file can be obtained from IPFS through the IPFS hash.
As shown in Figure 4, data provider (node D P k ) sends its data, e.g., model parameters, to the data receiver (node D P h ):
  • Data provider creates an AES key.
  • Data provider stores data onto the IPFS and gets the corresponding IPFS hash.
  • Data provider encrypts the AES key in the above step using the public RSA key provided by the data receiver, which ensures that only the data receiver can conduct decryption to access the AES key.
  • Data provider send the encrypted AES key to data receiver.
  • Data provider send encrypted IPFS hash to data receiver.
  • Data receiver get the encrypt AES key and conducts decryption with its RSA private key.
  • Data receiver get the encrypted IPFS hash and conducts decryption with the received AES key in the above step.
  • Data receiver get the relevant file from IPFS with the IPFS hash.
The direct communication between data node D P k and D P h only occurs at Steps 4 and 5 in the proposed scheme, where the size of both the AES key and IPFS hash are significantly smaller than DGM or DNN model parameters. Therefore, the proposed IPFS data sharing scheme in RDFL is able to significantly benefit the system communication efficiently and reduce communication cost especially when the blockchain technique is used.

3.4. Communication and Computation Complexity

Since each node needs to train its local discriminator and generator, RDFL requires similar computations compared to FedGan [36] and increased computations (roughly doubled) for each node compared to distributed GAN [35]. The communications in the proposed RDFL are mainly limited to parameter transferring among all nodes in each round for every K steps. Assume M is the size of model parameters, including the discriminator and generator, there would be N 1 times communications in one round and the average load per communication time per node is M. Increasing K could reduce the communication frequency, which may reduce the performance of the FL train. An overview of another two decentralized FL communication methods is shown in Figure 5. A summary of communication complexity comparison is shown in Table 1. Generally, the total transferred data volume per FL round is similar for all three methods. The proposed RDFL achieves a better performance on the communication pressure of the nodes, which could benefit the system’s bandwidth utilization and increase robustness.

4. Experimental Results and Discussions

4.1. Experimental Setup

In this section, we conduct several experiments to evaluate the proposed RDFL to show its convergence, performance in generating close-to-real data, and robustness in reducing communications (by increasing synchronization interval K). The inception score (IS) is used in this paper, which is a common criterion in measuring the performance of GAN [12]. Another criterion used is the Earth Mover’s Distance (EMD), which is also known as the Wasserstein distance. In practice, EMD is approximated by comparing average softmax scores of drawn samples from real data against the generated data such that:
E M D ( ( x r , y r ) , ( x g , y g ) ) = 1 N i = 1 N ( f o ( x r i ) | y r i | ( f o ( x g i ) | y g i | )
where ( x r , y r ) are real data samples, ( x g , y g ) are generated data samples, f o is the oracle classifier mentioned above. EMD measures a relative distance between real data and fake data. Obviously, a better generator should have a lower EMD by producing realistic images closer to real images.
We build the training set of each client by randomly choosing 50% of the total training samples with replacements to simulate IID data. In order to further examine the performance of RDFL on a non-iid dataset, the latent dirichlet allocation (LDA) and the label partition method is applied to divide the dataset into N partitions [39].

4.2. RDFL Training Performance with GAN

We test RDFL on MNIST to show its performance on image datasets. MNIST consists of 10 classes of data, which we split across B = 5. The hyperparameters of the GAN model used in the experiment are listed in Table 2.
From Figure 6, the trained GAN with RDFL is able to generate close-to-real images. We check Gan with RDFL performance robustness to reduced communications and increased synchronization intervals K by setting K = 1000; 2000; 5000; 10,000; 20,000. The results are shown in the middle and right part of Figure 6, which indicate that Gan with RDFL has high performance for image data. Furthermore, its performance is robust to reducing the communications by increasing synchronization intervals K. In addition, we also conduct experiments for GAN with RDFL under the non-iid scenario. The results are shown in Figure 6. It could be seen that the GAN with RDFL could still finish training with acceptable image generation quality. We would like to encourage researchers to tackle the problem of federated learning of GANs with non-IID data in the future.

5. Conclusions

In this paper, we propose a decentralized FL framework based on DGMs called RDFL to tackle the problems in existing decentralized FL frameworks. RDFL utilizes a consistent hashing algorithm and Ring-allreduce to improve communication performance, decentralized FL performance and stability. Moreover, RDFL introduces IPFS to further improve communication performance and reduce communication cost. We hope that RDFL can facilitate the application of decentralized FL with DGMs on medical areas. Future work will be focused on improving effective aggregation methods to replace the existing FedAvg algorithm. The related code will be published at https://github.com/ZJU-DistributedAI/RDFL-GAN accessed on 7 April 2021.

Author Contributions

Conceptualization, Z.W. (Zhao Wang), C.W. and Y.H.; methodology, Z.W. (Zhao Wang); software, Y.H.; validation, Y.H., R.H. and Z.W. (Zhihao Wang); formal analysis, R.H. and Z.W. (Zhihao Wang); investigation, Y.H.; resources, Y.H.; data curation, R.H. and Z.W. (Zhihao Wang); writing—original draft preparation, Z.W. (Zhao Wang); writing—review and editing, S.Y.; visualization, Y.H., R.H. and Z.W. (Zhihao Wang); supervision, C.W.; project administration, Z.W. (Zhao Wang); funding acquisition, C.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key Research and Development Project of China (2021ZD0110400), National Natural Science Foundation of China (U19B2042) and Ningbo Natural Science Foundation (2021J167).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sai Ambati, L.; El-Gayar, O.F.; Nawar, N. Influence of the Digital Divide and Socio-Economic Factors on Prevalence of Diabetes. Issues Inf. Syst. 2020, 21, 103–113. [Google Scholar]
  2. Ambati, L.S.; El-Gayar, O.; El, O.; Nawar, N. Design Principles for Multiple Sclerosis Mobile Self-Management Applications: A Patient-Centric Perspective. In Proceedings of the AMCIS 2021 Proceedings, Virtual, 9–13 August 2021; p. 11. [Google Scholar]
  3. Ren, H.; Li, H.; Dai, Y.; Yang, K.; Lin, X. Querying in internet of things with privacy preserving: Challenges, solutions and opportunities. IEEE Netw. 2018, 32, 144–151. [Google Scholar] [CrossRef]
  4. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial Intelligence and Statistics, PMLR, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  5. Ryffel, T.; Trask, A.; Dahl, M.; Wagner, B.; Mancuso, J.; Rueckert, D.; Passerat-Palmbach, J. A generic framework for privacy preserving deep learning. arXiv 2018, arXiv:1811.04017. [Google Scholar]
  6. Caldas, S.; Duddu, S.M.K.; Wu, P.; Li, T.; Konečnỳ, J.; McMahan, H.B.; Smith, V.; Talwalkar, A. Leaf: A benchmark for federated settings. arXiv 2018, arXiv:1812.01097. [Google Scholar]
  7. Lu, Y.; Huang, X.; Dai, Y.; Maharjan, S.; Zhang, Y. Blockchain and federated learning for privacy-preserved data sharing in industrial IoT. IEEE Trans. Ind. Inform. 2019, 16, 4177–4186. [Google Scholar] [CrossRef]
  8. Hao, M.; Li, H.; Luo, X.; Xu, G.; Yang, H.; Liu, S. Efficient and privacy-enhanced federated learning for industrial artificial intelligence. IEEE Trans. Ind. Inform. 2019, 16, 6532–6542. [Google Scholar] [CrossRef]
  9. Savazzi, S.; Nicoli, M.; Rampa, V. Federated learning with cooperating devices: A consensus approach for massive IoT networks. IEEE Internet Things J. 2020, 7, 4641–4654. [Google Scholar] [CrossRef] [Green Version]
  10. Liu, Y.; James, J.; Kang, J.; Niyato, D.; Zhang, S. Privacy-preserving traffic flow prediction: A federated learning approach. IEEE Internet Things J. 2020, 7, 7751–7763. [Google Scholar] [CrossRef]
  11. Yonetani, R.; Takahashi, T.; Hashimoto, A.; Ushiku, Y. Decentralized Learning of Generative Adversarial Networks from Non-iid Data. arXiv 2019, arXiv:1905.09684. [Google Scholar]
  12. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. Gans trained by a two time-scale update rule converge to a local nash equilibrium. arXiv 2017, arXiv:1706.08500. [Google Scholar]
  13. Augenstein, S.; McMahan, H.B.; Ramage, D.; Ramaswamy, S.; Kairouz, P.; Chen, M.; Mathews, R. Generative models for effective ML on private, decentralized datasets. arXiv 2019, arXiv:1911.06679. [Google Scholar]
  14. Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated machine learning: Concept and applications. ACM Trans. Intell. Syst. Technol. (TIST) 2019, 10, 1–19. [Google Scholar] [CrossRef]
  15. Tang, Z.; Shi, S.; Chu, X. Communication-efficient decentralized learning with sparsification and adaptive peer selection. arXiv 2020, arXiv:2002.09692. [Google Scholar]
  16. Philippenko, C.; Dieuleveut, A. Artemis: Tight convergence guarantees for bidirectional compression in federated learning. arXiv 2020, arXiv:2006.14591. [Google Scholar]
  17. He, C.; Tan, C.; Tang, H.; Qiu, S.; Liu, J. Central server free federated learning over single-sided trust social networks. arXiv 2019, arXiv:1910.04956. [Google Scholar]
  18. Li, Y.; Chen, C.; Liu, N.; Huang, H.; Zheng, Z.; Yan, Q. A blockchain-based decentralized federated learning framework with committee consensus. IEEE Netw. 2020, 35, 234–241. [Google Scholar] [CrossRef]
  19. Hu, C.; Jiang, J.; Wang, Z. Decentralized federated learning: A segmented gossip approach. arXiv 2019, arXiv:1908.07782. [Google Scholar]
  20. Zhao, Y.; Zhao, J.; Jiang, L.; Tan, R.; Niyato, D. Mobile edge computing, blockchain and reputation-based crowdsourcing iot federated learning: A secure, decentralized and privacy-preserving system. arXiv 2019, arXiv:1906.10893. [Google Scholar]
  21. Lamping, J.; Veach, E. A fast, minimal memory, consistent hash algorithm. arXiv 2014, arXiv:1406.2294. [Google Scholar]
  22. Benet, J. Ipfs-content addressed, versioned, p2p file system. arXiv 2014, arXiv:1407.3561. [Google Scholar]
  23. Lalitha, A.; Shekhar, S.; Javidi, T.; Koushanfar, F. Fully decentralized federated learning. In Proceedings of the Third workshop on Bayesian Deep Learning (NeurIPS), Montreal, QC, Canada, 7 December 2018. [Google Scholar]
  24. Amiri, M.M.; Gunduz, D.; Kulkarni, S.R.; Poor, H.V. Federated learning with quantized global model updates. arXiv 2020, arXiv:2006.10672. [Google Scholar]
  25. Konečnỳ, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. arXiv 2016, arXiv:1610.05492. [Google Scholar]
  26. Tang, H.; Gan, S.; Zhang, C.; Zhang, T.; Liu, J. Communication compression for decentralized training. Adv. Neural Inf. Process. Syst. 2018, 31, 7652–7662. [Google Scholar]
  27. Koloskova, A.; Lin, T.; Stich, S.U.; Jaggi, M. Decentralized deep learning with arbitrary communication compression. arXiv 2019, arXiv:1907.09356. [Google Scholar]
  28. Zhao, Y.; Li, M.; Lai, L.; Suda, N.; Civin, D.; Chandra, V. Federated learning with non-iid data. arXiv 2018, arXiv:1806.00582. [Google Scholar] [CrossRef]
  29. Jeong, E.; Oh, S.; Kim, H.; Park, J.; Bennis, M.; Kim, S.L. Communication-efficient on-device machine learning: Federated distillation and augmentation under non-iid private data. arXiv 2018, arXiv:1811.11479. [Google Scholar]
  30. Itahara, S.; Nishio, T.; Koda, Y.; Morikura, M.; Yamamoto, K. Distillation-Based Semi-Supervised Federated Learning for Communication-Efficient Collaborative Training with Non-IID Private Data. arXiv 2020, arXiv:2008.06180. [Google Scholar] [CrossRef]
  31. Geyer, R.C.; Klein, T.; Nabi, M. Differentially private federated learning: A client level perspective. arXiv 2017, arXiv:1712.07557. [Google Scholar]
  32. Melis, L.; Song, C.; De Cristofaro, E.; Shmatikov, V. Exploiting unintended feature leakage in collaborative learning. In Proceedings of the 2019 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 20–22 May 2019; pp. 691–706. [Google Scholar]
  33. Kim, H.; Park, J.; Bennis, M.; Kim, S.L. Blockchained on-device federated learning. IEEE Commun. Lett. 2019, 24, 1279–1283. [Google Scholar] [CrossRef] [Green Version]
  34. Li, F.; Ma, L.; Cai, J. Multi-discriminator generative adversarial network for high resolution gray-scale satellite image colorization. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 3489–3492. [Google Scholar]
  35. Hardy, C.; Le Merrer, E.; Sericola, B. Gossiping GANs: Position paper. In Proceedings of the Second Workshop on Distributed Infrastructures for Deep Learning, Rennes, France, 10 December 2018; pp. 25–28. [Google Scholar]
  36. Rasouli, M.; Sun, T.; Rajagopal, R. Fedgan: Federated generative adversarial networks for distributed data. arXiv 2020, arXiv:2006.07228. [Google Scholar]
  37. Fan, C.; Liu, P. Federated generative adversarial learning. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Beijing, China, 29 October–1 November 2020; pp. 3–15. [Google Scholar]
  38. Rajotte, J.F.; Mukherjee, S.; Robinson, C.; Ortiz, A.; West, C.; Ferres, J.L.; Ng, R.T. Reducing bias and increasing utility by federated generative modeling of medical images using a centralized adversary. arXiv 2021, arXiv:2101.07235. [Google Scholar]
  39. He, C.; Li, S.; So, J.; Zhang, M.; Wang, H.; Wang, X.; Vepakomma, P.; Singh, A.; Qiu, H.; Shen, L.; et al. Fedml: A research library and benchmark for federated machine learning. arXiv 2020, arXiv:2007.13518. [Google Scholar]
Figure 1. Ring topology.
Figure 1. Ring topology.
Electronics 11 01548 g001
Figure 2. Ring topology with virtual nodes.
Figure 2. Ring topology with virtual nodes.
Electronics 11 01548 g002
Figure 3. Ring decentralized federated learning.
Figure 3. Ring decentralized federated learning.
Electronics 11 01548 g003
Figure 4. Workflow of IPFS data sharing scheme.
Figure 4. Workflow of IPFS data sharing scheme.
Electronics 11 01548 g004
Figure 5. Overview of decentralized FL communication methods: p2p (left) and FL Gossip (right).
Figure 5. Overview of decentralized FL communication methods: p2p (left) and FL Gossip (right).
Electronics 11 01548 g005
Figure 6. Illustration of FL training quality with GAN on MNIST IID (left) and non-IID (right), where number of nodes B = 5 . (left) Generated images on K = 2000 , (middle) IS vs. Iterations with K [1000, 2000, 5000, 10,000, 20,000], (right) EMD vs. Iterations with K [1000, 2000, 5000, 10,000, 20,000].
Figure 6. Illustration of FL training quality with GAN on MNIST IID (left) and non-IID (right), where number of nodes B = 5 . (left) Generated images on K = 2000 , (middle) IS vs. Iterations with K [1000, 2000, 5000, 10,000, 20,000], (right) EMD vs. Iterations with K [1000, 2000, 5000, 10,000, 20,000].
Electronics 11 01548 g006aElectronics 11 01548 g006b
Table 1. Communication complexity analysis.
Table 1. Communication complexity analysis.
Decentralized FL FrameworkCommunication Times/RoundNode Pressure (MB/c)Total Transferred Data Volume per Round (MB)
P2P1 N × M N 2 M
FL Gossip [19] r o u n d ( N 1 2 ) 2 M 2 N M × r o u n d ( N 1 2 )
RDFL N 1 M N ( N 1 ) M
Table 2. GAN hyperparameters. The learning rates for the generator and discriminator are both equal to the same value, across all cases. BN stands for batch normalization, and Trans Conv stands for transpose convolution, Conv means convolution.
Table 2. GAN hyperparameters. The learning rates for the generator and discriminator are both equal to the same value, across all cases. BN stands for batch normalization, and Trans Conv stands for transpose convolution, Conv means convolution.
OperationKernelStridesFeature MapsBNNon-Linearity
G ( z ) 100 × 1 × 1 input
Trans Conv4 × 41 × 1256YReLU
Trans Conv4 × 42 × 2128YReLU
Trans Conv4 × 42 × 264YReLU
Trans Conv4 × 42 × 23NTanh
D(x)32 × 32 × 3 input
Conv4 × 42 × 232YLeakyReLU
Conv4 × 42 × 264YLeakyReLU
Conv4 × 42 × 2128YLeakyReLU
Conv4 × 41 × 11N-
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, Z.; Hu, Y.; Yan, S.; Wang, Z.; Hou, R.; Wu, C. Efficient Ring-Topology Decentralized Federated Learning with Deep Generative Models for Medical Data in eHealthcare Systems. Electronics 2022, 11, 1548. https://doi.org/10.3390/electronics11101548

AMA Style

Wang Z, Hu Y, Yan S, Wang Z, Hou R, Wu C. Efficient Ring-Topology Decentralized Federated Learning with Deep Generative Models for Medical Data in eHealthcare Systems. Electronics. 2022; 11(10):1548. https://doi.org/10.3390/electronics11101548

Chicago/Turabian Style

Wang, Zhao, Yifan Hu, Shiyang Yan, Zhihao Wang, Ruijie Hou, and Chao Wu. 2022. "Efficient Ring-Topology Decentralized Federated Learning with Deep Generative Models for Medical Data in eHealthcare Systems" Electronics 11, no. 10: 1548. https://doi.org/10.3390/electronics11101548

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop