Secure Smart Communication Efficiency in Federated Learning: Achievements and Challenges
Abstract
:1. Introduction
2. Problem Statement
2.1. Background
2.2. The Components of FL Systems
2.2.1. The Processes of an FL System
2.2.2. The Different Types of FL Systems
- Horizontal Federated Learning: This type of FL system is when the data from various devices have a similar set of features in terms of the domain but with different instances. This is the original sense of FL learning, where data from each domain are homogeneous, as depicted in Figure 3. Then, domains contribute together to train the global ML model [24]. This can be explained using the original example that is presented by Google [6], wherein the global model is an aggregate of multiple locally trained participating devices [37].
- Vertical Federated Learning: The data distributed across in a vertical FL setting are common data between unrelated domains. This could perhaps be called a feature-based learning scheme as the datasets involved in the training process perhaps share the same sample ID space but may differ in feature space. An example could be where a bank and an e-commerce business in the same city have a mutual user base, as shown in Figure 4. The bank has user-sensitive data, such as credit card ratings or revenue. At the same time, the e-commerce business has a purchase and browsing history. Here, two different domains can use their data to maybe create a prediction model based on the user and product information [29].
- Federated Transfer Learning: This type of system is different from the aforementioned systems, where neither the samples nor the features have many similarities [38]. An example could be where two data sources, such as a bank in the United States and an e-commerce business in Canada, are restricted by geography but still have a small intersection with each other, being different institutions, similar to a vertical FL. However, this is just the method of partitioning the data by the ML model being similar to the traditional ML method of transfer learning, where the ML model used is a pre-trained model on a similar dataset. This method can provide better results in some cases compared to a newly built ML model [24]; this is further shown in Figure 5.
2.3. Publication Analysis
3. Research Questions and Communication Efficient Methods
3.1. RQ1-What Are Some of the Challenges Presented in FL with Regards to Communication?
- Number of participating devices: Having a high number of participating devices in an FL environment has its advantage, wherein the ML model could be trained on more data and there could be a possible increase in performance and accuracy. However, the large number of devices participating in multiple FL training rounds at the same time could create a communication bottleneck. In some cases, a high number of clients could also increase the overall computational cost [24,52].
- Network bandwidth: In contrast to the traditional ML approach, the FL approach reduces the cost substantially; however, the communication bandwidth still needs to be preserved [24]. The participating devices may not always have the bandwidth needed. They could be participating under unreliable network conditions. Factors such as having a difference between the upload speed and download speed could result in delays, such as model uploads by participants [30] to the central server, which could lead to a potential bottleneck, leading to disrupting the FL environment.
- Limited edge node computation: The computation is now dependent on edge devices rather than powerful GPUs and CPUs. The edge devices could have limitations towards computation, power resources, storage, and limited link bandwidth [53]. The authors in [53] compared the training time between a central server and an edge device. They elaborated that an image classification model with over 60 million parameters can be trained in just a few minutes over a GPU, reaching speeds of 56 Gbps. However, even with a powerful smartphone connected over 5G, it could take much longer, reaching an average speed of 50 Mbps.
- Statistical heterogeneity: Another possible source for a communication bottleneck or where communication costs can rise could be statistical heterogeneity, where the data are non-independent and identically (non-i.i.d.) distributed [54]. In an FL environment, the data are only locally present on each participating device. They are gathered and collected by the participants on their independent devices based on their usage pattern and local environment. An individual participant’s dataset in an FL environment could not be representative of the population distribution of the other participants and their datasets [6]. The size of data gathered and distributed amongst devices can typically vary heavily [55]. Therefore, this type of fluctuation in the size of the dataset could affect communication by causing a delay in model updates and other attributes. A device with a larger dataset could take longer to update, whereas a smaller one could be carried out with updates. However, the global model might not be aggregated until all individual client models are trained and uploaded, causing a bottleneck.
3.2. RQ2-How Can Communication Be More Efficient in an FL Environment?
3.2.1. Local Updating
3.2.2. Client Selection
3.2.3. Reducing Model Updates
3.2.4. Decentralized Training and Peer-to-Peer Learning
3.2.5. Compression Schemes
- a.
- Reducing the size of the object / ML model from the clients to the server, i.e., that used to update the overall global model.
- b.
- Reducing the size of the global model that is shared with the clients on the network, i.e., the model on which the clients start local training using the available data.
- c.
- Any changes that are made to the overall training algorithm that make training the global training model more computationally efficient.
4. Discussion
5. Conclusions and Future Expectations
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Weichert, D.; Link, P.; Stoll, A.; Rüping, S.; Ihlenfeldt, S.; Wrobel, S. A review of machine learning for the optimization of production processes. Int. J. Adv. Manuf. Technol. 2019, 104, 1889–1902. [Google Scholar] [CrossRef]
- Pazzani, M. Comprehensible knowledge discovery: Gaining insight from data. In Proceedings of the First Federal Data Mining Conference and Exposition, Citeseer, Newport Beach, CA, USA, 14–17 August 1997; pp. 73–82. [Google Scholar]
- Meyer, G.; Adomavicius, G.; Johnson, P.E.; Elidrisi, M.; Rush, W.A.; Sperl-Hillen, J.M.; O’Connor, P.J. A machine learning approach to improving dynamic decision making. Inf. Syst. Res. 2014, 25, 239–263. [Google Scholar] [CrossRef]
- L’heureux, A.; Grolinger, K.; Elyamany, H.F.; Capretz, M.A. Machine learning with big data: Challenges and approaches. IEEE Access 2017, 5, 7776–7797. [Google Scholar] [CrossRef]
- Albrecht, J.P. How the GDPR will change the world. Eur. Data Prot. L. Rev. 2016, 2, 287. [Google Scholar] [CrossRef]
- McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial Intelligence and Statistics, PMLR, Fort Lauderdale, FL, USA, 9–11 May 2017; pp. 1273–1282. [Google Scholar]
- Truex, S.; Baracaldo, N.; Anwar, A.; Steinke, T.; Ludwig, H.; Zhang, R.; Zhou, Y. A hybrid approach to privacy-preserving federated learning. In Proceedings of the 12th ACM Workshop on Artificial Intelligence and Security, Los Angeles, CA, USA, 7–11 November 2019; pp. 1–11. [Google Scholar]
- Aledhari, M.; Razzak, R.; Parizi, R.M.; Saeed, F. Federated learning: A survey on enabling technologies, protocols, and applications. IEEE Access 2020, 8, 140699–140725. [Google Scholar] [CrossRef] [PubMed]
- Yazdinejad, A.; Srivastava, G.; Parizi, R.M.; Dehghantanha, A.; Choo, K.K.R.; Aledhari, M. Decentralized Authentication of Distributed Patients in Hospital Networks Using Blockchain. IEEE J. Biomed. Health Inform. 2020, 24, 2146–2156. [Google Scholar] [CrossRef]
- House, W. Consumer data privacy in a networked world: A framework for protecting privacy and promoting innovation in the global digital economy. White House Washington DC 2012, 1, 120. [Google Scholar]
- Rieke, N.; Hancox, J.; Li, W.; Milletari, F.; Roth, H.R.; Albarqouni, S.; Bakas, S.; Galtier, M.N.; Landman, B.A.; Maier-Hein, K.; et al. The future of digital health with federated learning. NPJ Digit. Med. 2020, 3, 1–7. [Google Scholar] [CrossRef] [PubMed]
- Liu, Y.; James, J.; Kang, J.; Niyato, D.; Zhang, S. Privacy-preserving Traffic Flow Prediction: A Federated Learning Approach. IEEE Internet Things J. 2020, 7, 7751–7763. [Google Scholar] [CrossRef]
- Yazdinejad, A.; Parizi, R.M.; Dehghantanha, A.; Karimipour, H. Federated learning for drone authentication. Ad. Hoc. Netw. 2021, 120, 102574. [Google Scholar] [CrossRef]
- Pokhrel, S.R.; Choi, J. Federated learning with blockchain for autonomous vehicles: Analysis and design challenges. IEEE Trans. Commun. 2020, 68, 4734–4746. [Google Scholar] [CrossRef]
- Zhao, Y.; Zhao, J.; Jiang, L.; Tan, R.; Niyato, D. Mobile edge computing, blockchain and reputation-based crowdsourcing iot federated learning: A secure, decentralized and privacy-preserving system. arXiv 2019, arXiv:1906.10893. [Google Scholar]
- Mothukuri, V.; Khare, P.; Parizi, R.M.; Pouriyeh, S.; Dehghantanha, A.; Srivastava, G. Federated Learning-based Anomaly Detection for IoT Security Attacks. IEEE Internet Things J. 2021, 9, 2545–2554. [Google Scholar] [CrossRef]
- Attota, D.C.; Mothukuri, V.; Parizi, R.M.; Pouriyeh, S. An Ensemble Multi-View Federated Learning Intrusion Detection for IoT. IEEE Access 2021, 9, 117734–117745. [Google Scholar] [CrossRef]
- Saharkhizan, M.; Azmoodeh, A.; Dehghantanha, A.; Choo, K.K.R.; Parizi, R.M. An Ensemble of Deep Recurrent Neural Networks for Detecting IoT Cyber Attacks Using Network Traffic. IEEE Internet Things J. 2020, 7, 8852–8859. [Google Scholar] [CrossRef]
- Hard, A.; Rao, K.; Mathews, R.; Ramaswamy, S.; Beaufays, F.; Augenstein, S.; Eichner, H.; Kiddon, C.; Ramage, D. Federated learning for mobile keyboard prediction. arXiv 2018, arXiv:1811.03604. [Google Scholar]
- Chen, M.; Mathews, R.; Ouyang, T.; Beaufays, F. Federated learning of out-of-vocabulary words. arXiv 2019, arXiv:1903.10635. [Google Scholar]
- Yang, T.; Andrew, G.; Eichner, H.; Sun, H.; Li, W.; Kong, N.; Ramage, D.; Beaufays, F. Applied federated learning: Improving google keyboard query suggestions. arXiv 2018, arXiv:1812.02903. [Google Scholar]
- Ramaswamy, S.; Mathews, R.; Rao, K.; Beaufays, F. Federated learning for emoji prediction in a mobile keyboard. arXiv 2019, arXiv:1906.04329. [Google Scholar]
- Tian, L. Federated Learning: Challenges, Methods and Future Directions. 2019. Available online: https://blog.ml.cmu.edu/2019/11/12/federated-learning-challenges-methods-and-future-directions/ (accessed on 13 November 2020).
- Mothukuri, V.; Parizi, R.M.; Pouriyeh, S.; Huang, Y.; Dehghantanha, A.; Srivastava, G. A survey on security and privacy of federated learning. Future Gener. Comput. Syst. 2021, 115, 619–640. [Google Scholar] [CrossRef]
- Canetti, R.; Feige, U.; Goldreich, O.; Naor, M. Adaptively secure multi-party computation. In Proceedings of the Twenty-Eighth Annual ACM Symposium on Theory of Computing, Philadelphia, PA, USA, 22–24 May 1996; pp. 639–648. [Google Scholar]
- Dwork, C. Differential privacy: A survey of results. In Proceedings of the International Conference on Theory and Applications of Models of Computation, Xi’an, China, 25–29 April 2008; Springer: Berlin/Heidelberg, Germany; pp. 1–19. [Google Scholar]
- Tramèr, F.; Kurakin, A.; Papernot, N.; Goodfellow, I.; Boneh, D.; McDaniel, P. Ensemble adversarial training: Attacks and defenses. arXiv 2017, arXiv:1705.07204. [Google Scholar]
- Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings, R.; et al. Advances and open problems in federated learning. arXiv 2019, arXiv:1912.04977. [Google Scholar]
- Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated machine learning: Concept and applications. Acm Trans. Intell. Syst. Technol. (TIST) 2019, 10, 1–19. [Google Scholar] [CrossRef]
- Lim, W.Y.B.; Luong, N.C.; Hoang, D.T.; Jiao, Y.; Liang, Y.C.; Yang, Q.; Niyato, D.; Miao, C. Federated learning in mobile edge networks: A comprehensive survey. IEEE Commun. Surv. Tutorials 2020, 22, 2031–2063. [Google Scholar] [CrossRef]
- Połap, D.; Srivastava, G.; Lin, J.C.W.; Woźniak, M. Federated Learning Model with Augmentation and Samples Exchange Mechanism. In Proceedings of the International Conference on Artificial Intelligence and Soft Computing, Zakopane, Poland, 20–24 June 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 214–223. [Google Scholar]
- Połap, D.; Srivastava, G.; Yu, K. Agent architecture of an intelligent medical system based on federated learning and blockchain technology. J. Inf. Secur. Appl. 2021, 58, 102748. [Google Scholar] [CrossRef]
- Ahmed, U.; Srivastava, G.; Lin, J.C.W. Reliable customer analysis using federated learning and exploring deep-attention edge intelligence. Future Gener. Comput. Syst. 2022, 127, 70–79. [Google Scholar] [CrossRef]
- Xu, R.; Baracaldo, N.; Zhou, Y.; Anwar, A.; Ludwig, H. Hybridalpha: An efficient approach for privacy-preserving federated learning. In Proceedings of the 12th ACM Workshop on Artificial Intelligence and Security, London, UK, 15 November 2019; pp. 13–23. [Google Scholar]
- Kang, J.; Xiong, Z.; Niyato, D.; Zou, Y.; Zhang, Y.; Guizani, M. Reliable federated learning for mobile networks. IEEE Wirel. Commun. 2020, 27, 72–80. [Google Scholar] [CrossRef]
- Zhu, H.; Zhang, H.; Jin, Y. From federated learning to federated neural architecture search: A survey. Complex Intell. Syst. 2021, 7, 639–657. [Google Scholar] [CrossRef]
- Tian, Z.; Zhang, R.; Hou, X.; Liu, J.; Ren, K. FederBoost: Private Federated Learning for GBDT. arXiv 2020, arXiv:2011.02796. [Google Scholar]
- Chen, Y.; Qin, X.; Wang, J.; Yu, C.; Gao, W. Fedhealth: A federated transfer learning framework for wearable healthcare. IEEE Intell. Syst. 2020, 35, 83–93. [Google Scholar] [CrossRef]
- Xu, J.; Glicksberg, B.S.; Su, C.; Walker, P.; Bian, J.; Wang, F. Federated learning for healthcare informatics. J. Healthc. Inform. Res. 2021, 5, 1–19. [Google Scholar] [CrossRef] [PubMed]
- Yin, L.; Feng, J.; Xun, H.; Sun, Z.; Cheng, X. A Privacy-Preserving Federated Learning for Multiparty Data Sharing in Social IoTs. IEEE Trans. Netw. Sci. Eng. 2021, 8, 2706–2718. [Google Scholar] [CrossRef]
- Li, L.; Fan, Y.; Tse, M.; Lin, K.Y. A review of applications in federated learning. Comput. Ind. Eng. 2020, 149, 106854. [Google Scholar] [CrossRef]
- Brisimi, T.S.; Chen, R.; Mela, T.; Olshevsky, A.; Paschalidis, I.C.; Shi, W. Federated learning of predictive models from federated electronic health records. Int. J. Med. Inform. 2018, 112, 59–67. [Google Scholar] [CrossRef]
- Nasajpour, M.; Karakaya, M.; Pouriyeh, S.; Parizi, R.M. Federated Transfer Learning For Diabetic Retinopathy Detection Using CNN Architectures. In Proceedings of the SoutheastCon 2022, IEEE, Mobile, AL, USA, 26 March–3 April 2022; pp. 655–660. [Google Scholar]
- Li, Q.; Wen, Z.; Wu, Z.; Hu, S.; Wang, N.; Li, Y.; Liu, X.; He, B. A survey on federated learning systems: Vision, hype and reality for data privacy and protection. IEEE Trans. Knowl. Data Eng. 2021, early access. [Google Scholar] [CrossRef]
- Pfitzner, B.; Steckhan, N.; Arnrich, B. Federated Learning in a Medical Context: A Systematic Literature Review. Acm Trans. Internet Technol. (TOIT) 2021, 21, 1–31. [Google Scholar] [CrossRef]
- Lyu, L.; Yu, H.; Yang, Q. Threats to federated learning: A survey. arXiv 2020, arXiv:2003.02133. [Google Scholar]
- Rahman, K.J.; Ahmed, F.; Akhter, N.; Hasan, M.; Amin, R.; Aziz, K.E.; Islam, A.M.; Mukta, M.S.H.; Islam, A.N. Challenges, applications and design aspects of federated learning: A survey. IEEE Access 2021, 9, 124682–124700. [Google Scholar] [CrossRef]
- Li, T.; Sahu, A.K.; Talwalkar, A.; Smith, V. Federated learning: Challenges, methods, and future directions. IEEE Signal Process. Mag. 2020, 37, 50–60. [Google Scholar] [CrossRef]
- Konečnỳ, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. arXiv 2016, arXiv:1610.05492. [Google Scholar]
- Zheng, S.; Shen, C.; Chen, X. Design and Analysis of Uplink and Downlink Communications for Federated Learning. IEEE J. Sel. Areas Commun. 2020, 39, 2150–2167. [Google Scholar] [CrossRef]
- Luping, W.; Wei, W.; Bo, L. CMFL: Mitigating communication overhead for federated learning. In Proceedings of the 2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS), Dallas, TX, USA, 7–10 July 2019; pp. 954–964. [Google Scholar]
- Chen, Y.; Sun, X.; Jin, Y. Communication-efficient federated deep learning with layerwise asynchronous model update and temporally weighted aggregation. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 4229–4238. [Google Scholar] [CrossRef] [PubMed]
- Shi, Y.; Yang, K.; Jiang, T.; Zhang, J.; Letaief, K.B. Communication-efficient edge AI: Algorithms and systems. IEEE Commun. Surv. Tutorials 2020, 22, 2167–2191. [Google Scholar] [CrossRef]
- Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated optimization in heterogeneous networks. arXiv 2018, arXiv:1812.06127. [Google Scholar]
- Sattler, F.; Wiedemann, S.; Müller, K.R.; Samek, W. Robust and communication-efficient federated learning from non-iid data. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 3400–3413. [Google Scholar] [CrossRef] [PubMed]
- Zhang, S.; Choromanska, A.E.; LeCun, Y. Deep learning with elastic averaging SGD. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 685–693. [Google Scholar]
- Zhao, Y.; Li, M.; Lai, L.; Suda, N.; Civin, D.; Chandra, V. Federated learning with non-iid data. arXiv 2018, arXiv:1806.00582. [Google Scholar] [CrossRef]
- Smith, V.; Forte, S.; Ma, C.; Takáč, M.; Jordan, M.I.; Jaggi, M. CoCoA: A general framework for communication-efficient distributed optimization. J. Mach. Learn. Res. 2017, 18, 8590–8638. [Google Scholar]
- Briggs, C.; Fan, Z.; Andras, P. Federated learning with hierarchical clustering of local updates to improve training on non-IID data. arXiv 2020, arXiv:2004.11791. [Google Scholar]
- Singh, A.; Yadav, A.; Rana, A. K-means with Three different Distance Metrics. Int. J. Comput. Appl. 2013, 67, 13–17. [Google Scholar] [CrossRef]
- Reisizadeh, A.; Mokhtari, A.; Hassani, H.; Jadbabaie, A.; Pedarsani, R. Fedpaq: A communication-efficient federated learning method with periodic averaging and quantization. In Proceedings of the International Conference on Artificial Intelligence and Statistics, PMLR, Online, 26–28 August 2020; pp. 2021–2031. [Google Scholar]
- Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.; Stich, S.; Suresh, A.T. Scaffold: Stochastic controlled averaging for federated learning. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 5132–5143. [Google Scholar]
- Shamir, O.; Srebro, N.; Zhang, T. Communication-efficient distributed optimization using an approximate newton-type method. In Proceedings of the International Conference on Machine Learning, PMLR, Bejing, China, 22–24 June 2014; pp. 1000–1008. [Google Scholar]
- Reddi, S.J.; Konečnỳ, J.; Richtárik, P.; Póczós, B.; Smola, A. AIDE: Fast and communication efficient distributed optimization. arXiv 2016, arXiv:1608.06879. [Google Scholar]
- Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smithy, V. Feddane: A federated newton-type method. In Proceedings of the 2019 IEEE 53rd Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 3–6 November 2019; pp. 1227–1231. [Google Scholar]
- AbdulRahman, S.; Tout, H.; Mourad, A.; Talhi, C. FedMCCS: Multicriteria client selection model for optimal IoT federated learning. IEEE Internet Things J. 2020, 8, 4723–4735. [Google Scholar] [CrossRef]
- Xu, J.; Wang, H. Client selection and bandwidth allocation in wireless federated learning networks: A long-term perspective. IEEE Trans. Wirel. Commun. 2020, 20, 1188–1200. [Google Scholar] [CrossRef]
- Nishio, T.; Yonetani, R. Client selection for federated learning with heterogeneous resources in mobile edge. In Proceedings of the ICC 2019-2019 IEEE International Conference on Communications (ICC), Shanghai, China, 20–24 May 2019; pp. 1–7. [Google Scholar]
- Anh, T.T.; Luong, N.C.; Niyato, D.; Kim, D.I.; Wang, L.C. Efficient training management for mobile crowd-machine learning: A deep reinforcement learning approach. IEEE Wirel. Commun. Lett. 2019, 8, 1345–1348. [Google Scholar] [CrossRef]
- Cho, Y.J.; Wang, J.; Joshi, G. Client Selection in Federated Learning: Convergence Analysis and Power-of-Choice Selection Strategies. arXiv 2020, arXiv:2010.01243. [Google Scholar]
- Zhang, W.Z.; Elgendy, I.A.; Hammad, M.; Iliyasu, A.M.; Du, X.; Guizani, M.; Abd El-Latif, A.A. Secure and Optimized Load Balancing for Multitier IoT and Edge-Cloud Computing Systems. IEEE Internet Things J. 2020, 8, 8119–8132. [Google Scholar] [CrossRef]
- Kamp, M.; Adilova, L.; Sicking, J.; Hüger, F.; Schlicht, P.; Wirtz, T.; Wrobel, S. Efficient decentralized deep learning by dynamic model averaging. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Dublin, Ireland, 10–14 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 393–409. [Google Scholar]
- Bui, T.D.; Nguyen, C.V.; Swaroop, S.; Turner, R.E. Partitioned variational inference: A unified framework encompassing federated and continual learning. arXiv 2018, arXiv:1811.11206. [Google Scholar]
- Smith, V.; Chiang, C.K.; Sanjabi, M.; Talwalkar, A. Federated multi-task learning. arXiv 2017, arXiv:1705.10467. [Google Scholar]
- Guha, N.; Talwalkar, A.; Smith, V. One-shot federated learning. arXiv 2019, arXiv:1902.11175. [Google Scholar]
- Nguyen, H.T.; Sehwag, V.; Hosseinalipour, S.; Brinton, C.G.; Chiang, M.; Poor, H.V. Fast-convergent federated learning. IEEE J. Sel. Areas Commun. 2020, 39, 201–218. [Google Scholar] [CrossRef]
- Tang, H.; Gan, S.; Zhang, C.; Zhang, T.; Liu, J. Communication compression for decentralized training. arXiv 2018, arXiv:1803.06443. [Google Scholar]
- Reisizadeh, A.; Taheri, H.; Mokhtari, A.; Hassani, H.; Pedarsani, R. Robust and communication-efficient collaborative learning. arXiv 2019, arXiv:1907.10595. [Google Scholar]
- He, C.; Tan, C.; Tang, H.; Qiu, S.; Liu, J. Central server free federated learning over single-sided trust social networks. arXiv 2019, arXiv:1910.04956. [Google Scholar]
- Lalitha, A.; Kilinc, O.C.; Javidi, T.; Koushanfar, F. Peer-to-peer federated learning on graphs. arXiv 2019, arXiv:1901.11173. [Google Scholar]
- Roy, A.G.; Siddiqui, S.; Pölsterl, S.; Navab, N.; Wachinger, C. Braintorrent: A peer-to-peer environment for decentralized federated learning. arXiv 2019, arXiv:1905.06731. [Google Scholar]
- Rothchild, D.; Panda, A.; Ullah, E.; Ivkin, N.; Stoica, I.; Braverman, V.; Gonzalez, J.; Arora, R. Fetchsgd: Communication-efficient federated learning with sketching. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual Event, 13–18 July 2020; pp. 8253–8265. [Google Scholar]
- Spring, R.; Kyrillidis, A.; Mohan, V.; Shrivastava, A. Compressing gradient optimizers via count-sketches. arXiv 2019, arXiv:1902.00179. [Google Scholar]
- Li, S.; Qi, Q.; Wang, J.; Sun, H.; Li, Y.; Yu, F.R. GGS: General Gradient Sparsification for Federated Learning in Edge Computing. In Proceedings of the ICC 2020-2020 IEEE International Conference on Communications (ICC), Virtual, 7–11 June 2020; pp. 1–7. [Google Scholar]
- Hu, R.; Gong, Y.; Guo, Y. Sparsified Privacy-Masking for Communication-Efficient and Privacy-Preserving Federated Learning. arXiv 2020, arXiv:2008.01558. [Google Scholar]
- Sattler, F.; Wiedemann, S.; Müller, K.R.; Samek, W. Sparse binary compression: Towards distributed deep learning with minimal communication. In Proceedings of the 2019 IEEE International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar]
- Nicholls, J. Quantization in Deep Learning. 2018. Available online: https://medium.com/@joel_34050/quantization-in-deep-learning-478417eab72b/ (accessed on 28 April 2021).
- Jiang, P.; Agrawal, G. A linear speedup analysis of distributed deep learning with sparse and quantized communication. Adv. Neural Inf. Process. Syst. 2018, 31, 2525–2536. [Google Scholar]
- Amiri, M.M.; Gunduz, D.; Kulkarni, S.R.; Poor, H.V. Federated learning with quantized global model updates. arXiv 2020, arXiv:2006.10672. [Google Scholar]
- Koloskova, A.; Lin, T.; Stich, S.U.; Jaggi, M. Decentralized deep learning with arbitrary communication compression. arXiv 2019, arXiv:1907.09356. [Google Scholar]
- Dai, X.; Yan, X.; Zhou, K.; Yang, H.; Ng, K.K.; Cheng, J.; Fan, Y. Hyper-sphere quantization: Communication-efficient sgd for federated learning. arXiv 2019, arXiv:1911.04655. [Google Scholar]
- Shlezinger, N.; Chen, M.; Eldar, Y.C.; Poor, H.V.; Cui, S. Federated learning with quantization constraints. In Proceedings of the ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 8851–8855. [Google Scholar]
- Shlezinger, N.; Chen, M.; Eldar, Y.C.; Poor, H.V.; Cui, S. UVeQFed: Universal vector quantization for federated learning. IEEE Trans. Signal Process. 2020, 69, 500–514. [Google Scholar] [CrossRef]
- Liu, L.; Zhang, J.; Song, S.; Letaief, K.B. Hierarchical Quantized Federated Learning: Convergence Analysis and System Design. arXiv 2021, arXiv:2103.14272. [Google Scholar]
Reference | Year | Objective | Security | Privacy | Communication | Challenges | Future |
---|---|---|---|---|---|---|---|
[29] | 2019 | Providing a survey of existing works on FL, discussing frameworks, concepts, and applications. | ✓ | ✓ | ✗ | ✓ | ✓ |
[46] | 2020 | Introduction of concept of FL, predominantly covering threat model attacks. | ✓ | ✓ | ✗ | ✓ | ✓ |
[8] | An FL survey focusing on hardware, software and technologies, and real-life applications | ✓ | ✓ | ✓ | ✓ | ✓ | |
[30] | Introduction of FL, presenting some existing challenges and their solutions | ✓ | ✓ | ✓ | ✓ | ✓ | |
[24] | A survey on security and privacy of federated learning | ✓ | ✓ | ✗ | ✓ | ✓ | |
[48] | A detailed survey introducing FL and the challenges. | ✓ | ✓ | ✓ | ✓ | ✓ | |
[41] | An overview of FL characteristics and applications within three specific domains | ✓ | ✓ | ✓ | ✓ | ✓ | |
[39] | 2021 | A survey of FL in healthcare, covering common topics of introduction of technology, challenges, etc. | ✓ | ✓ | ✓ | ✓ | ✓ |
[24] | A comprehensive survey posing research questions with regard to FL and privacy and security | ✓ | ✓ | ✓ | ✓ | ✓ | |
[44] | A thorough categorization of FL based on six main aspects to enable effective deployment of FL models | ✓ | ✓ | ✓ | ✓ | ✓ | |
[45] | A systematic literature review of FL research studies with a concentration on the medical domain | ✓ | ✓ | ✓ | ✓ | ✓ | |
[47] | A comprehensive review of three main aspects of FL: design, applications, and challenges | ✓ | ✓ | ✓ | ✓ | ✓ |
Ref. | Section | Model and Technology | Remarks |
---|---|---|---|
[59] | Local Updating | Hierarchical clustering technique | An FL+HC technique separating client clusters similarity of local updates |
[61] | FedPAQ | Using periodic averaging to aggregate and achieve global model updates | |
[62] | SCAFFOLD algorithm | An algorithm that provides better convergence rates over non-iid data | |
[55] | Compression Schemes-Sparsification | STC method | Providing compression for both upstream and downstream communications |
[82] | FetchSGD | Compresses the gradient based on client’s local data | |
[84] | General gradient sparsification (GSS) | Batch normalization layer with local gradients mitigating the impact of delayed gradients and not increasing the communication overhead | |
[85] | CPFed | A sparsified masking model providing compression and differential privacy | |
[86] | Sparse binary compression (SBS) | Introducing temporal sparsity, where gradients are not communicated after every local iteration | |
[61] | Compression Schemes-Quantization | FedPAQ | Using quantization techniques based upon model accuracy |
[89] | Lossy FL algorithm (LFL) | Quantizing models before broadcasting | |
[91] | Hyper-sphere quantization (HSQ) framework | Ability to reduce the cost of communication per iteration | |
[92] | UVeQFed | Algorithm convergence of model minimizes the loss function | |
[94] | Heir-Local-QSGD | Leveraging client–edge–cloud network hierarchy and quantized models updates | |
[81] | Decentralized Training or Peer-to-peer Learning | BrainTorrent | A peer-to-peer learning framework where models converge faster and reach good accuracy |
[78] | QuanTimed-DSGD | decentralized gradient-based optimization imposing iteration deadlines for devices | |
[66] | Client Selection | FedMCCS | A multi-criteria client selection that considers IoT device specification and network condition |
[67] | Resource allocation model | Optimizing learning performance in how clients are selected and how bandwidth is allocated | |
[68] | FedCS | The framework allows the server to aggregate as many clients as possible within a certain time-frame | |
[70] | Power-of-choice | A communication and computation-efficient client selection framework | |
[72] | Reduced Model Updates | A decentralized deep learning model | Ability to handle different phases of the model training well |
[73] | A partitioned variational inference (PVI) | A Bayesian neural network over FL that is synchronous and asynchronous for model updates across machines | |
[75] | One-shot federated learning | A single round of communication performed between central server and connected devices | |
[76] | FOLB | Intelligent sampling of devices in each round of model training to optimize the convergence speed |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Pouriyeh, S.; Shahid, O.; Parizi, R.M.; Sheng, Q.Z.; Srivastava, G.; Zhao, L.; Nasajpour, M. Secure Smart Communication Efficiency in Federated Learning: Achievements and Challenges. Appl. Sci. 2022, 12, 8980. https://doi.org/10.3390/app12188980
Pouriyeh S, Shahid O, Parizi RM, Sheng QZ, Srivastava G, Zhao L, Nasajpour M. Secure Smart Communication Efficiency in Federated Learning: Achievements and Challenges. Applied Sciences. 2022; 12(18):8980. https://doi.org/10.3390/app12188980
Chicago/Turabian StylePouriyeh, Seyedamin, Osama Shahid, Reza M. Parizi, Quan Z. Sheng, Gautam Srivastava, Liang Zhao, and Mohammad Nasajpour. 2022. "Secure Smart Communication Efficiency in Federated Learning: Achievements and Challenges" Applied Sciences 12, no. 18: 8980. https://doi.org/10.3390/app12188980
APA StylePouriyeh, S., Shahid, O., Parizi, R. M., Sheng, Q. Z., Srivastava, G., Zhao, L., & Nasajpour, M. (2022). Secure Smart Communication Efficiency in Federated Learning: Achievements and Challenges. Applied Sciences, 12(18), 8980. https://doi.org/10.3390/app12188980