A Trusted Federated Learning Method Based on Consortium Blockchain
Abstract
:1. Introduction
- Smart Contract Design: We introduce a novel smart contract framework that automates the verification of participants, manages model updates, and ensures the auditability of the entire learning process. Smart contracts are utilized to validate the computational power of participants, prevent malicious updates, and provide transparency in model aggregation.
- Central Management Chain and Private Chains: To address scalability and privacy concerns, we propose a decentralized architecture that includes a central management chain and multiple private chains. The central management chain coordinates model aggregation and manages cross-chain communication, while private chains enable participants to train local models without exposing sensitive data. This structure ensures that the system is scalable, secure, and privacy-preserving.
- Cross-Chain Communication Protocol: A critical challenge in the blockchain–FL integration is the overhead introduced by cross-chain communication. We design an efficient cross-chain communication protocol that enables low-latency model aggregation and synchronization between the central management chain and private chains. This protocol ensures that updates are securely shared across chains while minimizing communication delays.
- We propose a hybrid blockchain architecture that combines a central management chain and multiple private chains, addressing both scalability and privacy concerns in federated learning.
- We design and implement smart contracts that automate participant validation, model update management, and provide transparent audit logs, improving the security and trustworthiness of the learning process.
- We develop a cross-chain communication protocol that efficiently manages data exchange and model aggregation between chains, significantly reducing communication overhead.
2. Related Work
2.1. Federated Learning
2.2. Blockchain in Federated Learning
2.3. Challenges in Existing Systems
- Real-Time Security: Blockchain’s consensus algorithms, while ensuring security, can introduce latency, making it difficult to achieve real-time verification of model updates. This is particularly problematic in time-sensitive applications such as real-time health monitoring and autonomous driving.
- Cross-Chain Communication: In blockchain-based federated learning systems that use multiple private chains, ensuring efficient and secure communication between these chains and the central management chain is a challenge. Existing methods struggle with optimizing cross-chain transactions and managing resources effectively.
- Resource Management: Managing computational and communication resources across heterogeneous devices is complex. Blockchain, with its decentralized architecture, requires additional computational resources for maintaining the ledger and executing consensus protocols, which can further exacerbate the resource limitations of certain federated learning participants.
3. Preliminary
3.1. Blockchain Basics
- Efficiency and Scalability: Consortium blockchains achieve significantly higher transaction throughput (200–1000 TPS) and lower latency (seconds) compared to public blockchains, making them suitable for federated learning, where frequent model updates are required.
- Privacy and Security: Consortium blockchains restrict access to authorized participants, ensuring a high level of privacy critical for domains like healthcare and finance. Public blockchains, in contrast, expose all transaction data to the network, which compromises privacy.
- Cost-Effectiveness: Consortium blockchains operate on lightweight consensus mechanisms (e.g., Proof of Authority), avoiding the energy-intensive processes of public blockchains, such as Proof of Work, thereby reducing operational costs.
- Trust Model: The semi-trusted nature of consortium blockchains aligns well with federated learning, where participants (e.g., hospitals, financial institutions) are known entities. A fully decentralized trust model, as in public blockchains, is unnecessary in such environments.
- Decentralized trust: No single participant controls the ledger.
- Immutability: Past model updates are irreversible.
- Smart contract enforcement: Automation of model verification and aggregation.
3.2. Federated Learning Architecture
- represent the set of participants;
- denote the local model parameters of participant at training round t;
- be the aggregation function used to update the global model.
3.3. Challenges in Secure Federated Learning
- Privacy Risks: Although FL prevents raw data from being shared, the exchange of model parameters can still leak information about the local data. For instance, membership inference attacks allow adversaries to infer whether specific data samples were part of a participant’s dataset. Mathematically, given a set of model updates , an adversary can compute gradients and exploit them to reconstruct private information from .
- Malicious Data and Model Poisoning: A participant may upload poisoned model parameters to corrupt the global model. The goal of an adversary is to maximize the global loss by sending manipulated parameters:
- Communication Bottlenecks: In FL, model parameters must be exchanged between the participants and the central server during each training round. The total communication cost can be expressed as
- Asynchronous Updates and System Heterogeneity: In practice, participants may have different computational capabilities and network conditions, leading to asynchronous updates. Let represent the local computation time of participant , and be the time required for a global aggregation round. The straggler effect arises when , slowing down the overall learning process and leading to suboptimal global model updates.
3.4. Formal Problem Definition
- Objective: Minimize the global loss function under secure, privacy-preserving conditions using blockchain for model aggregation and participant validation.
- Constraints:
- –
- Ensure is updated only with verified, non-malicious updates.
- –
- Minimize communication costs .
- –
- Maintain privacy by preventing information leakage through model updates.
4. Proposed Scheme
4.1. System Architecture
- Central Management Chain: This is the core blockchain that manages global model aggregation and cross-chain communication. It acts as the orchestrator of the entire system, receiving updates from private chains and executing smart contracts to ensure model consistency and security. The CMC records global model parameters and logs all transactions for auditable and transparent operations.
- Private Chains: Each private chain is formed by a group of participants with similar computational power. Participants train local models on their respective datasets, and model parameters are periodically shared within their private chain. Once the local models are updated and validated, the private chain sends aggregated model updates to the CMC.
4.2. Cross-Chain Communication
- Routing Protocols: Cross-chain routing is responsible for handling requests and data exchange between the CMC and private chains. When a private chain completes its local model aggregation, the routing protocol forwards the aggregated parameters to the CMC. Similarly, the CMC, after updating the global model, disseminates the new model parameters to all private chains through the routing protocol.
- Smart Contracts: Smart contracts automate the validation and management of model updates. Specifically, the Model Update Contract (MUC) ensures that the model updates sent from private chains meet predefined criteria before being incorporated into the global model, as Algorithm 3 shows below. Each cross-chain communication instance is logged using smart contracts to ensure transparency and accountability.
4.3. Model Update Process
- Local Model Training: Each participant in a private chain trains a local model on its dataset . After completing local training, each participant shares its model parameters with other members in the same private chain.
- Private Chain Aggregation: Within each private chain , the participants’ model parameters are aggregated using an averaging function , forming the private chain’s local global model :
- Cross-Chain Communication to CMC: The aggregated model from each private chain is sent to the CMC via the cross-chain routing protocol. The CMC aggregates these models using the weighted aggregation function :The global model is updated by weighting the contributions from each private chain based on their computational power.
- Redistribution of Global Model: Once the global model is updated, it is distributed back to all private chains. Each participant receives the updated global model, which it uses as the starting point for the next round of local training.
4.4. Computational Power-Based Chain Grouping
4.5. Smart Contracts
4.5.1. Supervisory Contract
Algorithm 1 Supervisory Contract |
|
4.5.2. Computational Power Registration Contract
Algorithm 2 Computational Power Registration Contract |
|
4.5.3. Model Update Contract
Algorithm 3 Model Update Contract |
|
4.5.4. Private Chain Model Fusion Contract
Algorithm 4 Private Chain Model Fusion Contract |
|
4.6. Security Features
- Auditable Logging: Every interaction, including model updates and cross-chain communications, is logged on the blockchain. These logs are immutable and can be audited to ensure that all actions adhere to predefined protocols.
- Consensus Algorithms: The system uses a Proof of Authority (PoA) consensus algorithm in the central management chain, ensuring that only trusted nodes validate model updates and cross-chain communications. The PoA algorithm minimizes the risk of malicious actors influencing the system while ensuring efficiency.
- Smart Contract-Based Regulation: Smart contracts enforce the rules governing the federated learning process, including participant validation, model update criteria, and reward mechanisms. These contracts ensure that only valid model updates are incorporated into the global model and that participants follow the agreed-upon protocols.
- Tamper-Resistant Model Updates: By utilizing blockchain’s immutability, the system ensures that, once a model update is recorded, it cannot be altered. This prevents tampering with model updates, further ensuring the integrity of the global model.
5. Security Analysis
5.1. Trust Model
- Data Integrity: Every transaction, including model updates and participant interactions, is recorded immutably on the blockchain. Let represent the transaction recording the model update from participant at round t. Each transaction is linked to its predecessor via a cryptographic hash , ensuring that the history of updates cannot be altered:
- Transparency: All participants can verify the sequence of transactions and model updates stored on the blockchain. The transparency of the system is achieved by allowing all participants to audit the blockchain’s contents, which include the model updates and decisions made by the smart contracts.
- Auditability: Smart contracts enforce predefined rules governing model updates. Let be the Model Update Contract, which verifies each model update before it is incorporated into the global model:All validations are logged on the blockchain, providing a transparent and auditable history of model updates. This ensures that only valid updates are included in the final model.
5.2. Privacy Protection
- Local Data Retention: Participants in the system do not share raw data; instead, they retain their data locally and only share model parameters. Let represent the local dataset of participant . Instead of sharing , each participant trains a local model and shares only the parameters . This ensures that sensitive information in is never exposed to other participants or the central authority.
- Decentralized Data Exchange: The use of blockchain enables decentralized, secure data exchanges. Model parameters are exchanged through the blockchain using cryptographic techniques that ensure the confidentiality of the updates. Each update is encrypted before being transmitted over the blockchain, and only authorized participants with the correct decryption keys can access the model updates. Let represent the encryption of with key k, ensuring thatThis prevents unauthorized access to model updates and ensures that, even in the event of an eavesdropping attack, no useful information is leaked.
- Differential Privacy: To further enhance privacy, the system can employ differential privacy techniques to protect participants’ data from being inferred through model updates. Let represent a differentially private version of the model parameters. The differential privacy mechanism adds noise to the model parameters before they are shared:
5.3. Resilience to Attacks
5.3.1. Defense Against Model Poisoning Attacks
- Cross-chain validation: Model updates are aggregated within each private chain and verified by the Model Update Contract . This contract checks the consistency of the updates, ensuring that poisoned updates do not deviate significantly from the valid updates sent by other participants:
- Weighted Aggregation: Even if a poisoned update is included, the blockchain’s consensus and aggregation mechanisms reduce its impact on the global model by weighting contributions based on the participant’s trust score and computational power. Therefore, the effect of any individual malicious participant is minimized.
5.3.2. Defense Against Data Tampering
5.3.3. Defense Against Collusion Among Malicious Participants
5.3.4. Defense Against Adaptive Adversarial Attacks
5.3.5. Defense Against Model Inversion Attack
5.3.6. Defense Against Membership Inference Attack
5.3.7. Experimental Results and Analysis
6. Performance Comparison and Evaluation
6.1. Experimental Setup
- Model Accuracy: The primary evaluation metric is the test set accuracy, calculated as the ratio of correctly classified images to the total number of test samples.
- Convergence Speed: To measure communication efficiency, we use the Time of Accuracy (ToA@x) metric, where x represents a target accuracy. The ToA@x metric records the number of communication rounds needed to reach a specified accuracy level.
- Security Overhead: We also measure the computational overhead introduced by blockchain integration, particularly in cross-chain communication and validation processes.
6.2. Experimental Results
6.2.1. Effect of Blockchain Integration on Communication Efficiency and Accuracy
- High accuracy and fast convergence: For applications requiring precise results and rapid convergence (e.g., healthcare diagnostics or financial forecasting), a smaller (e.g., ) is recommended. This ensures strict computational power alignment, reducing variance within private chains and enhancing model consistency.
- Bandwidth-constrained or heterogeneous environments: In scenarios with limited communication bandwidth or highly diverse participant capabilities, a larger (e.g., ) may be preferable to reduce communication overhead, even if convergence speed slightly decreases.
- Balanced trade-offs: For general-purpose federated learning tasks, a moderate (e.g., ) provides a good balance between communication efficiency and model accuracy.
6.2.2. Influence of IID and Non-IID Data on Performance
- IID Scenario: All participants have an evenly distributed dataset, ideal for federated learning as all clients contribute equally to the global model.
- Slightly IID Scenario (): Data are moderately skewed across participants, with some variability in data distribution. This scenario reflects a more practical setting wherein data are not uniformly distributed.
- Non-IID Scenario (): Data distribution is highly uneven, with each participant having a limited subset of data classes. This represents the most challenging scenario for model convergence and accuracy.
6.2.3. Impact of Consortium Blockchain-Based Grouping with Different FL Algorithms
6.3. Extended Experimental Validation in Real Scenarios
6.3.1. ISIC 2017 Dataset for Lesion Diagnosis Classification
6.3.2. NuCLS Dataset for Nucleus Classification
6.4. Blockchain-Induced Latency Analysis
- Consensus Delay: The Proof of Authority (PoA) consensus algorithm introduces latency proportional to the number of validator nodes, as each transaction must be validated and signed by these nodes.
- Transaction Processing Delay: Each submitted model update is treated as a transaction on the blockchain. The time required to validate and package these transactions into a block adds to the system latency.
- Block Generation Time: The interval between successive blocks determines how quickly transactions are confirmed. Shorter intervals reduce latency but may lead to increased computational overhead and potential for chain forks.
6.5. Discussion
- Enhanced Communication Efficiency: The consortium blockchain-based federated learning system optimizes participant grouping, leading to fewer communication rounds needed for model convergence, particularly under non-IID conditions.
- Improved Model Accuracy: The scheme provides robust accuracy improvements in non-IID scenarios by enforcing controlled and validated participant grouping, ensuring that diverse and representative data contribute to the global model.
- Security and Computational Overhead: While blockchain integration adds computational overhead due to consensus mechanisms and cross-chain communication, the enhanced security and auditability offset these costs. The tamper-resistant and auditable nature of blockchain provides added value in sensitive applications.
7. Conclusions
7.1. Theoretical and Practical Implications
- Theoretical Implications: From a theoretical perspective, this study contributes to the field by formalizing a trust model for federated learning based on blockchain technology. The integration of blockchain ensures that model updates are auditable, verifiable, and tamper-resistant, establishing a decentralized trust framework. Additionally, the system’s cross-chain communication and smart contract-based validation processes provide a solid theoretical foundation for handling secure data exchanges in distributed learning environments.
- Practical Implications: Practically, the system enhances the efficiency and security of federated learning processes. By grouping participants based on their computational power, we optimize model update frequency and reduce communication delays. The use of blockchain provides a secure infrastructure for federated learning, making the system applicable in various industries where data privacy and trust are critical, such as healthcare, finance, and smart cities. Moreover, the system’s ability to retain data locally ensures compliance with stringent privacy regulations, such as GDPR, while still enabling collaborative model training.
7.2. Limitations
- Cross-Chain Communication Latency: Although the use of cross-chain communication protocols ensures secure data transmission, there is an inherent latency associated with blockchain transactions, particularly in large-scale networks with numerous participants. This latency may affect the timeliness of model updates in real-time applications.
- Computational Overhead: The additional computational requirements for maintaining the blockchain, verifying transactions, and executing smart contracts introduce overhead that may not be suitable for participants with very limited resources. Although computational power-based grouping mitigates this to some extent, further optimization is required to ensure scalability across a broader range of devices.
- Security Beyond the Current Model: While the proposed system defends against common threats such as model poisoning and data tampering, more sophisticated attacks, such as adaptive adversarial attacks, require the integration of advanced security measures like homomorphic encryption and zero-knowledge proofs. These additions could further enhance the system’s robustness but may also increase computational complexity.
7.3. Future Work
- Optimizing Cross-Chain Communication: Future work can focus on reducing the latency of cross-chain communications. Developing more efficient routing protocols and transaction validation mechanisms could significantly improve the system’s scalability and real-time performance.
- Advanced Security Mechanisms: Incorporating advanced cryptographic techniques, such as homomorphic encryption or secure multi-party computation, can enhance the privacy and security of model updates. These methods would allow for secure computation on encrypted data, ensuring that sensitive information remains confidential even during model aggregation.
- Broadening Application Scope: The proposed blockchain-based FL system has the potential to be extended to other machine learning paradigms, such as unsupervised and reinforcement learning. Additionally, its application in domains like the Internet of Things (IoT), where data privacy and secure communication are paramount, could further demonstrate the practical utility of the system.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; PMLR: New York, NY, USA, 2017; pp. 1273–1282. [Google Scholar]
- Xu, J.; Glicksberg, B.S.; Su, C.; Walker, P.; Bian, J.; Wang, F. Federated learning for healthcare informatics. J. Healthc. Inform. Res. 2021, 5, 1–19. [Google Scholar] [CrossRef]
- Debauche, O.; Nkamla Penka, J.B.; Mahmoudi, S.; Lessage, X.; Hani, M.; Manneback, P.; Lufuluabu, U.K.; Bert, N.; Messaoudi, D.; Guttadauria, A. RAMi: A new real-time Internet of Medical Things architecture for elderly patient monitoring. Information 2022, 13, 423. [Google Scholar] [CrossRef]
- Chatterjee, P.; Das, D.; Rawat, D.B. Federated learning empowered recommendation model for financial consumer services. IEEE Trans. Consum. Electron. 2023, 70, 2508–2516. [Google Scholar] [CrossRef]
- Xu, C.; Mao, Y. An improved traffic congestion monitoring system based on federated learning. Information 2020, 11, 365. [Google Scholar] [CrossRef]
- Akallouch, M.; Akallouch, O.; Fardousse, K.; Bouhoute, A.; Berrada, I. Prediction and Privacy Scheme for Traffic Flow Estimation on the Highway Road Network. Information 2022, 13, 381. [Google Scholar] [CrossRef]
- Javeed, D.; Saeed, M.S.; Kumar, P.; Jolfaei, A.; Islam, S.; Islam, A.N. Federated learning-based personalized recommendation systems: An overview on security and privacy challenges. IEEE Trans. Consum. Electron. 2023, 70, 2618–2627. [Google Scholar] [CrossRef]
- Fang, M.; Cao, X.; Jia, J.; Gong, N. Local model poisoning attacks to Byzantine-Robust federated learning. In Proceedings of the 29th USENIX security symposium (USENIX Security 20), Denver, CO, USA, 14–17 August 2020; pp. 1605–1622. [Google Scholar]
- Lyu, X.; Han, Y.; Wang, W.; Liu, J.; Wang, B.; Liu, J.; Zhang, X. Poisoning with cerberus: Stealthy and colluded backdoor attack against federated learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 9020–9028. [Google Scholar]
- Asad, M.; Shaukat, S.; Hu, D.; Wang, Z.; Javanmardi, E.; Nakazato, J.; Tsukada, M. Limitations and future aspects of communication costs in federated learning: A survey. Sensors 2023, 23, 7358. [Google Scholar] [CrossRef]
- Almanifi, O.R.A.; Chow, C.O.; Tham, M.L.; Chuah, J.H.; Kanesan, J. Communication and computation efficiency in federated learning: A survey. Internet Things 2023, 22, 100742. [Google Scholar] [CrossRef]
- Albaseer, A.; Abdallah, M.; Al-Fuqaha, A.; Erbad, A. Data-driven participant selection and bandwidth allocation for heterogeneous federated edge learning. IEEE Trans. Syst. Man, Cybern. Syst. 2023, 53, 5848–5860. [Google Scholar] [CrossRef]
- Djebrouni, Y.; Benarba, N.; Touat, O.; De Rosa, P.; Bouchenak, S.; Bonifati, A.; Felber, P.; Marangozova, V.; Schiavoni, V. Bias mitigation in federated learning for edge computing. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024, 7, 1–35. [Google Scholar] [CrossRef]
- Nakamoto, S. Bitcoin: A peer-to-peer electronic cash system. Decentralized Business Review. 2008. Available online: https://bitcoin.org/bitcoin.pdf (accessed on 11 November 2024).
- Zhu, J.; Cao, J.; Saxena, D.; Jiang, S.; Ferradi, H. Blockchain-empowered federated learning: Challenges, solutions, and future directions. ACM Comput. Surv. 2023, 55, 1–31. [Google Scholar] [CrossRef]
- Ma, C.; Li, J.; Shi, L.; Ding, M.; Wang, T.; Han, Z.; Poor, H.V. When federated learning meets blockchain: A new distributed learning paradigm. IEEE Comput. Intell. Mag. 2022, 17, 26–33. [Google Scholar] [CrossRef]
- Wei, W.; Zhou, Y.; Li, D.; Hong, X. Double-Layer Blockchain-Based Decentralized Integrity Verification for Multi-chain Cross-Chain Data. In Proceedings of the International Conference on Neural Information Processing, Changsha, China, 20–23 November 2023; Springer: Singapore, 2023; pp. 264–279. [Google Scholar]
- Dib, O.; Brousmiche, K.L.; Durand, A.; Thea, E.; Hamida, E.B. Consortium blockchains: Overview, applications and challenges. Int. J. Adv. Telecommun 2018, 11, 51–64. [Google Scholar]
- Pfeiffer, K.; Rapp, M.; Khalili, R.; Henkel, J. Federated learning for computationally constrained heterogeneous devices: A survey. ACM Comput. Surv. 2023, 55, 1–27. [Google Scholar] [CrossRef]
- Ye, M.; Fang, X.; Du, B.; Yuen, P.C.; Tao, D. Heterogeneous federated learning: State-of-the-art and research challenges. ACM Comput. Surv. 2023, 56, 1–44. [Google Scholar] [CrossRef]
- Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated optimization in heterogeneous networks. Proc. Mach. Learn. Syst. 2020, 2, 429–450. [Google Scholar]
- Xu, Y.; Liao, Y.; Xu, H.; Ma, Z.; Wang, L.; Liu, J. Adaptive control of local updating and model compression for efficient federated learning. IEEE Trans. Mob. Comput. 2022, 22, 5675–5689. [Google Scholar] [CrossRef]
- Hu, Y.; Liu, T.; Yang, C.; Huang, Y.; Suo, S. A flexible model compression and resource allocation scheme for federated learning. IEEE Trans. Mach. Learn. Commun. Netw. 2023, 1, 168–184. [Google Scholar] [CrossRef]
- Wang, Z.; Zhang, Z.; Tian, Y.; Yang, Q.; Shan, H.; Wang, W.; Quek, T.Q. Asynchronous federated learning over wireless communication networks. IEEE Trans. Wirel. Commun. 2022, 21, 6961–6978. [Google Scholar] [CrossRef]
- Truex, S.; Liu, L.; Chow, K.H.; Gursoy, M.E.; Wei, W. LDP-Fed: Federated learning with local differential privacy. In Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking, Heraklion, Greece, 27 April 2020; pp. 61–66. [Google Scholar]
- Wu, X.; Zhang, Y.; Shi, M.; Li, P.; Li, R.; Xiong, N.N. An adaptive federated learning scheme with differential privacy preserving. Future Gener. Comput. Syst. 2022, 127, 362–372. [Google Scholar] [CrossRef]
- Mugunthan, V.; Polychroniadou, A.; Byrd, D.; Balch, T.H. Smpai: Secure multi-party computation for federated learning. In Proceedings of the NeurIPS 2019 Workshop on Robust AI in Financial Services, Vancouver, BA, Canada, 8–14 December 2019; MIT Press: Cambridge, MA, USA, 2019; Volume 21. [Google Scholar]
- Sotthiwat, E.; Zhen, L.; Li, Z.; Zhang, C. Partially encrypted multi-party computation for federated learning. In Proceedings of the 2021 IEEE/ACM 21st International Symposium on Cluster, Cloud and Internet Computing (CCGrid), Melbourne, Australia, 10–13 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 828–835. [Google Scholar]
- Aziz, R.; Banerjee, S.; Bouzefrane, S.; Le Vinh, T. Exploring homomorphic encryption and differential privacy techniques towards secure federated learning paradigm. Future Internet 2023, 15, 310. [Google Scholar] [CrossRef]
- Al-Huthaifi, R.; Li, T.; Huang, W.; Gu, J.; Li, C. Federated learning in smart cities: Privacy and security survey. Inf. Sci. 2023, 632, 833–857. [Google Scholar] [CrossRef]
- Nguyen, D.C.; Ding, M.; Pham, Q.V.; Pathirana, P.N.; Le, L.B.; Seneviratne, A.; Li, J.; Niyato, D.; Poor, H.V. Federated learning meets blockchain in edge computing: Opportunities and challenges. IEEE Internet Things J. 2021, 8, 12806–12825. [Google Scholar] [CrossRef]
- Taherdoost, H. Blockchain and machine learning: A critical review on security. Information 2023, 14, 295. [Google Scholar] [CrossRef]
- Shayan, M.; Fung, C.; Yoon, C.J.; Beschastnikh, I. Biscotti: A blockchain system for private and secure federated learning. IEEE Trans. Parallel Distrib. Syst. 2020, 32, 1513–1525. [Google Scholar] [CrossRef]
- Li, Y.; Chen, C.; Liu, N.; Huang, H.; Zheng, Z.; Yan, Q. A blockchain-based decentralized federated learning framework with committee consensus. IEEE Netw. 2020, 35, 234–241. [Google Scholar] [CrossRef]
- Qu, Y.; Gao, L.; Luan, T.H.; Xiang, Y.; Yu, S.; Li, B.; Zheng, G. Decentralized privacy using blockchain-enabled federated learning in fog computing. IEEE Internet Things J. 2020, 7, 5171–5183. [Google Scholar] [CrossRef]
- Kalapaaking, A.P.; Khalil, I.; Atiquzzaman, M. Smart Policy Control for Securing Federated Learning Management System. IEEE Trans. Netw. Serv. Manag. 2023, 20, 1600–1611. [Google Scholar] [CrossRef]
- Moudoud, H.; Cherkaoui, S.; Khoukhi, L. Towards a secure and reliable federated learning using blockchain. In Proceedings of the 2021 IEEE Global Communications Conference (GLOBECOM), Madrid, Spain, 7–11 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 01–06. [Google Scholar]
- Madill, E.; Nguyen, B.; Leung, C.K.; Rouhani, S. ScaleSFL: A sharding solution for blockchain-based federated learning. In Proceedings of the Fourth ACM International Symposium on Blockchain and Secure Critical Infrastructure, Nagasaki, Japan, 30 May–3 June 2022; pp. 95–106. [Google Scholar]
- Hsu, T.M.H.; Qi, H.; Brown, M. Federated visual classification with real-world data distribution. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part X 16. Springer: Cham, Switzerland, 2020; pp. 76–92. [Google Scholar]
Feature | Public Blockchain | Consortium Blockchain | Hybrid Blockchain |
---|---|---|---|
Transaction Throughput (TPS) | Low (10–20) | High (200–1000) | Medium (50–200) |
Latency (ms) | High (minutes to hours) | Low (seconds) | Medium (seconds to minutes) |
Resource Usage | High | Low | Medium |
Consensus Mechanism | PoW/PoS | PoA | Combination of PoW/PoA |
Energy Consumption | High | Low | Medium |
Privacy | Minimal | High | Moderate |
Scalability | Limited | High | Moderate |
Cost | High | Low | Moderate |
Attack Type | Dataset | ASR Without Defense (%) | ASR with Defense (%) | Model Accuracy (%) |
---|---|---|---|---|
Model Inversion | FashionMNIST | 85 | 12 | 88.5 |
CIFAR-10 | 80 | 15 | 87.5 | |
Membership Inference | FashionMNIST | 72 | 18 | 89.2 |
CIFAR-10 | 70 | 20 | 87.8 | |
Model Poisoning | FashionMNIST | 55 | 15 | 90.1 |
CIFAR-10 | 50 | 18 | 88.3 |
Hardware and Software | Setup |
---|---|
CPU | Intel(R) Core(TM) i9-9900X CPU @ 3.50 GHz |
Memory | 128 G |
GPU | NVIDIA GeForce RTX 2080 Ti × 8 |
CUDA Version | 12.0 |
Programming Language | python3.9 |
Operating System | Ubuntu 18.04.6 LTS |
Federated Learning Framework | Pytorch 1.10.2 |
Algorithm | Computational Power Alignment () | Average Accuracy | Highest Accuracy | [email protected] [email protected] | |||||
---|---|---|---|---|---|---|---|---|---|
F-MNIST | CIFAR-10 | F-MNIST | CIFAR-10 | F-MNIST | CIFAR-10 | ||||
FedAvg | - | 0.8318 | 0.6498 | 0.8567 | 0.6809 | 393 | - | ||
CBFedAvg | 0.8759 | 0.7254 | 0.8916 | 0.7375 | 90 | 210 | |||
FedAvg | - | 0.8631 | 0.7046 | 0.8713 | 0.7121 | 175 | 669 | ||
CBFedAvg | 0.8892 | 0.7468 | 0.8982 | 0.7609 | 72 | 173 | |||
FedAvg | - | 0.8778 | 0.7115 | 0.8803 | 0.7153 | 127 | 462 | ||
CBFedAvg | 0.8913 | 0.7401 | 0.8962 | 0.7506 | 95 | 186 |
Dataset | FashionMNIST | CIFAR-10 | |||||
---|---|---|---|---|---|---|---|
IID | IID | ||||||
FedAvg | Average Accuracy | 0.9118 | 0.8631 | 0.7522 | 0.8120 | 0.7046 | 0.6680 |
Highest Accuracy | 0.9127 | 0.8713 | 0.7811 | 0.8134 | 0.7121 | 0.6772 | |
[email protected] [email protected] | 24 | 175 | - | 66 | 669 | - | |
CBFedAvg | Average Accuracy | 0.9308 | 0.9032 | 0.8366 | 0.8328 | 0.7459 | 0.7367 |
Highest Accuracy | 0.9315 | 0.9040 | 0.8417 | 0.8336 | 0.7508 | 0.7492 | |
[email protected] [email protected] | 20 | 58 | - | 49 | 148 | 206 |
Algorithm | Average Accuracy | Highest Accuracy | [email protected] [email protected] | |||
---|---|---|---|---|---|---|
F-MNIST | CIFAR-10 | F-MNIST | CIFAR-10 | F-MNIST | CIFAR-10 | |
FedAvg | 0.8631 | 0.7046 | 0.8713 | 0.7121 | 175 | 669 |
FedProx | 0.8747 | 0.7100 | 0.8802 | 0.7192 | 143 | 401 |
FedIR | 0.8786 | 0.7202 | 0.8827 | 0.7266 | 106 | 293 |
CBFedAvg | 0.9032 | 0.7459 | 0.9042 | 0.7508 | 58 | 148 |
CBFedProx | 0.9061 | 0.7627 | 0.9153 | 0.7708 | 53 | 139 |
CBFedIR | 0.9118 | 0.7823 | 0.9186 | 0.7854 | 48 | 118 |
Model | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) |
---|---|---|---|---|
Baseline FL | 88.2 | 85.4 | 84.6 | 85.0 |
CBFL | 91.7 | 89.5 | 88.1 | 88.8 |
Model | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) |
---|---|---|---|---|
Baseline FL | 85.3 | 82.1 | 81.4 | 81.7 |
CBFL | 88.6 | 86.4 | 85.8 | 86.1 |
Configuration | Average Response Time (ms) | Max Response Time (ms) | Impact on Accuracy (%) |
---|---|---|---|
PoA (5 validators) | 150 | 300 | −0.2 |
PoA (10 validators) | 230 | 450 | −0.5 |
Dynamic Block Size (PoS) | 180 | 400 | −0.3 |
Batch Aggregation | 100 | 200 | −0.1 |
Strategy | Average Latency Reduction (%) | Accuracy Impact (%) |
---|---|---|
Optimized Consensus (PoA) | 35 | −0.1 |
Batch Aggregation | 30 | −0.05 |
Hierarchical Blockchain | 40 | −0.2 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yin, X.; Wu, X.; Zhang, X. A Trusted Federated Learning Method Based on Consortium Blockchain. Information 2025, 16, 14. https://doi.org/10.3390/info16010014
Yin X, Wu X, Zhang X. A Trusted Federated Learning Method Based on Consortium Blockchain. Information. 2025; 16(1):14. https://doi.org/10.3390/info16010014
Chicago/Turabian StyleYin, Xiaojun, Xijun Wu, and Xinming Zhang. 2025. "A Trusted Federated Learning Method Based on Consortium Blockchain" Information 16, no. 1: 14. https://doi.org/10.3390/info16010014
APA StyleYin, X., Wu, X., & Zhang, X. (2025). A Trusted Federated Learning Method Based on Consortium Blockchain. Information, 16(1), 14. https://doi.org/10.3390/info16010014