Next Article in Journal
An Online Evaluation Method for Random Number Entropy Sources Based on Time-Frequency Feature Fusion
Next Article in Special Issue
ExSPIN: Explicit Feedback-Based Self-Play Fine-Tuning for Text-to-SQL Parsing
Previous Article in Journal
Chinese Named Entity Recognition for Automobile Fault Texts Based on External Context Retrieving and Adversarial Training
Previous Article in Special Issue
Causal Discovery and Reasoning for Continuous Variables with an Improved Bayesian Network Constructed by Locality Sensitive Hashing and Kernel Density Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Class-Hidden Client-Side Watermarking in Federated Learning

1
School of Information Engineering, Yangzhou University, Yangzhou 225009, China
2
Jiangsu Province Engineering Research Center of Knowledge Management and Intelligent Service, Yangzhou 225127, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(2), 134; https://doi.org/10.3390/e27020134
Submission received: 22 November 2024 / Revised: 19 January 2025 / Accepted: 25 January 2025 / Published: 27 January 2025
(This article belongs to the Special Issue Applications of Information Theory to Machine Learning)

Abstract

:
Federated learning consists of a central aggregator and multiple clients, forming a distributed structure that effectively protects data privacy. However, since all participants can access the global model, the risk of model leakage increases, especially when unreliable participants are involved. To safeguard model copyright while enhancing the robustness and secrecy of the watermark, this paper proposes a client-side watermarking scheme. Specifically, the proposed method introduces an additional watermark class, expanding the output layer of the client model into an N + 1 -class classifier. The client’s local model is then trained using both the watermark dataset and the local dataset. Notably, before uploading to the server, the parameters of the watermark class are removed from the output layer and stored locally. Additionally, the client uploads amplified parameters to address the potential weakening of the watermark during the aggregation. After aggregation, the global model is distributed to the clients for local training. Through multiple rounds of iteration, the saved watermark parameters are continuously updated until the global model converges. On the MNIST, CIFAR-100, and CIFAR-10 datasets, the watermark detection rates on VGG-16 and ResNet-18 reached 100%. Furthermore, extensive experiments demonstrate that this method has minimal impact on model performance and exhibits strong robustness against pruning and fine-tuning attacks.

1. Introduction

Federated learning [1,2] is a distributed machine learning method that effectively addresses data privacy concerns and alleviates problems caused by data siloing [3,4,5]. However, the performance of such models comes at the cost of an expensive training process, requiring substantial training data, specialized hardware, and infrastructure. Federated learning is commonly applied in sensitive domains, such as healthcare and finance, where the security of the model directly impacts individuals’ health and financial well-being. Nevertheless, allowing all clients access to the global model in federated learning increases the risk of model leakage [6,7,8]. Therefore, protecting the copyrights of federated learning models is of paramount importance; otherwise, it may lead to business malicious competition, knowledge theft, or other malicious purposes [9,10,11,12].
To protect the intellectual property of deep learning models [13,14], scholars have introduced the concept of model watermarking [15,16,17]. Model watermarking involves embedding unique identifiers or features into the parameters, structure, or outputs of the model to mark ownership information. If the model is illegally copied or stolen, the watermark can help identify the copyright. Additionally, watermarking techniques can be combined with other security measures such as model encryption, access control, etc., to further enhance the security and copyright protection of the model. Currently, watermarking techniques are divided into white-box watermarking [18,19,20] and black-box watermarking [21,22,23]. White-box watermarking involves accessing the model’s internal parameters and embedding the watermark by modifying parameters or altering the model structure. The verification process also requires access to internal information [24,25]. White-box watermarks typically offer higher robustness and security because the watermark information is directly embedded into the model’s internal structure. However, since access to the model’s internal information is required, white-box watermarking may introduce certain computational and storage overheads, and may potentially leak some sensitive information about the model. On the other hand, black-box watermarking does not require access to the model’s internal information during embedding and verification processes. It embeds watermark information into the model’s input and output behaviors, relying solely on specific model outputs for watermark validation. Therefore, black-box watermarking is generally more lightweight. However, black-box watermarks may be more vulnerable to attacks, as the watermark information relies solely on the model’s output behavior, which may be subject to external interference or attacks.
Although model watermarking techniques have been widely applied in various fields, their application in federated learning models still faces numerous challenges. Federated learning, as a distributed learning framework, involves participants training models locally, with the central server unable to directly access the data. This architecture introduces new issues, such as watermark attenuation after aggregation, aggregation failures due to model structure changes (structure-based white-box watermarking), and impacts on the accuracy of the original task (black-box watermarking). These challenges make watermarking techniques designed for centralized learning unsuitable for federated learning. To address these challenges, researchers have proposed watermarking algorithms specifically designed for federated learning models. These algorithms are generally classified into two types: server-side watermarking [26,27] and client-side watermarking [28,29]. Server-side watermarking is intended to protect the global model, which is collaboratively trained by participants and aggregated on the central server. It embeds specific markers or features to identify the ownership of the model, preventing unauthorized replication. Client-side watermarking is used for ownership verification or tracking, preventing unauthorized copying or theft of the model. This watermarking is embedded during the local training process on participants’ devices, with the server not needing to directly access the training data.
These model watermarking methods have been adjusted and optimized in both white-box and black-box watermarking techniques to adapt to federated learning scenarios. However, white-box federated model watermarking algorithms require modifying the model weights and training based on specific loss functions to embed watermark information. While this approach enhances the robustness of the watermark, it significantly increases the algorithm’s complexity and training overhead. In contrast, black-box federated model watermarking algorithms embed triggers into the original task data, allowing the model to learn the mapping between the triggers and specific classes in the original task. Due to the significant differences between the trigger features and the original task features, this inevitably impacts the accuracy of the original task, particularly negatively affecting the accuracy of the classes containing the watermark samples.
To address the aforementioned issues, we propose an innovative watermarking algorithm that ensures a high watermark verification success rate while minimizing the impact on the original task, making it almost negligible, and avoiding an increase in model complexity. This approach adds an additional “watermark” class to the training data instead of embedding a watermark behavior directly into a specific class of the original task, thereby preventing interference with the classification task. This method makes the watermark more covert and difficult to detect while not affecting the model’s original task. Specifically, we assume that the initiator’s client is trustworthy. First, we create a private watermark dataset and modify the model structure from an N-class classification model to an (N + 1)-class classification model. During the client’s training process, the watermark is embedded into the model. Before submitting the model for aggregation, the client restores the original model structure and retains the output layer parameters related to the watermark class, continuously updating these parameters. Meanwhile, untrustworthy clients focus solely on the original task. During verification, the stored watermark parameters are reintegrated into the model’s output layer to confirm the ownership of the federated learning model. Additionally, we introduce a scaling factor ξ to address issues of aggregation failure and watermark attenuation. The main contributions of our work include the following:
This study proposes a watermarking method that ensures watermark labels are independent of the primary task labels, thereby avoiding interference between watermark information and task outputs. This independence ensures that watermark embedding does not affect the performance of the model’s primary task, guaranteeing high fidelity.
By modifying the model’s structure, the watermark can be effectively embedded into different types of models without relying on any specific class, thereby enhancing the adaptability of the watermark.
Various experiments demonstrate that the proposed method has strong robustness against pruning and fine-tuning attacks.
The remainder of this paper is structured as follows: Section 2 provides an overview of existing deep learning and federated learning watermarking technologies. Section 3 presents a general description of the threat model and the requirements for watermarks. The proposed method and algorithm are discussed in Section 4. Section 5 evaluates and analyzes the experimental results. Finally, Section 6 concludes this work.

2. Related Works

2.1. Centralized and Federated Learning Watermarking Schemes

White-box Watermarking Schemes: White-box watermarking refers to the method of directly embedding a watermark into a model by modifying its weights or structure. In centralized learning, white-box watermarking schemes are mainly categorized into weight-based [30,31,32,33] and structure-based watermarking [19,34,35]. In weight-based watermarking schemes, Uchida et al. [31] were the first to propose a method of embedding a watermark into a DNN model by imposing an additional regularization term on the weight parameters. Later, Liu et al. [32] introduced a method to enhance watermark robustness by selecting fewer but important parameters for embedding the watermark while introducing regularization terms. Additionally, Yan et al. [33] proposed a method of injecting virtual neurons into the model to resist neural structure obfuscation. In these weight-based white-box watermarking schemes, access to the internal information of the DNN is required during the validation phase to detect the watermark.
In structure-based watermarking schemes, methods such as in [19] use passports for deep neural network ownership verification, where using incorrect passports significantly degrades the performance of the original task. Zhang et al. and Fan et al. [34,35] improved this approach, enhancing the watermark’s robustness against deletion attacks and ambiguity. Although structural watermarking schemes have certain advantages, such as being independent of model parameters and offering enhanced resistance to deletion attacks, they also come with drawbacks and challenges that need to be carefully considered and balanced in practical applications.
Black-box Watermarking Schemes: Black-box watermarking typically uses specific trigger samples to embed unique watermark information to verify model ownership. Adi et al. [36] embedded a watermark into the model using a trigger set made up of predefined abstract samples known only to the model owner. Li et al. [37] generated trigger construction algorithms using frequency-domain image watermarking and verified ownership of remote suspicious models by issuing trigger queries to them. Similarly, Pyone et al. [18] employed learnable transformational images to embed watermark patterns into the model, enabling remote verification of model ownership. Since black-box watermarks generally rely only on the model’s output features or markers, attackers may confuse or destroy the watermark by forging similar features, thus complicating watermark verification.

2.2. Federated Learning Watermarking Schemes

Server-side Watermarking Schemes: In federated learning, server-side watermarking typically involves embedding the watermark into the global model after parameter aggregation, which is then distributed to the clients [26,27,38,39]. Merkle-Sign [40] uses Merkle trees [41] to verify ownership, suitable for collaborative client-server environments. The server embeds client and server keys in each round, uploads them to the Merkle tree, and eventually updates all client keys. FedTracker [42] employs a dual-layer protection scheme, including a global watermark mechanism and a local fingerprint mechanism, respectively providing ownership verification and traceability. Furthermore, Chen et al. [27] introduced a detector to capture the relationship between the critical sample feature distributions of the target model’s output and critical sample labels, effectively verifying ownership based on the detector’s accuracy.
Client-side Watermarking Schemes: Client-side watermarking in federated learning is a technique used to verify model ownership [28,29,39,43], involving embedding specific identifiers or information into the client’s model during training to ensure the model’s ownership. The main difference from server-side watermarking lies in their embedding locations. Liu et al. [43] predesigned a set of noise patterns as watermarks embedded into client models, influencing the aggregation process of the global model. FedIPR [28] is designed such that each client can embed its private watermark, based on private features and backdoors, to independently assert ownership and verify the watermark. Li et al. [28] allow each client to embed its private watermark, which can be either white-box or black-box, to independently assert ownership and detect free riders. Liang et al. [29] utilize replaceable client-side watermarking techniques to update the federated learning model every few rounds, enabling each model to use a unique watermark.

3. Threat Model and Watermark Requirements

3.1. Threat Model

Our current work faces two primary threat models. Firstly, it aims to prevent unauthorized replication and misuse of the model by other clients. In such scenarios, malicious clients may illicitly acquire well-trained models and employ them for commercial competition, privacy infringement, or other unethical purposes. This could lead to intellectual property violations, harming the interests of the model owner, and potentially compromising the reliability and security of the entire federated learning system. Secondly, there is a risk of privacy breaches due to dishonest servers. These servers may attempt to steal model parameters during the training process to gain unfair advantages, or inadequate security measures during transmission or storage may lead to data leaks.

3.2. Watermarking Scenario and Requirements

In this work, we consider a scenario where n clients collaborate with an owner (O) to jointly train a deep learning model f to address complex pattern recognition tasks. It is important to note that the owner (O) is not only a participant in the training process but also the initiator of the model. As described in Section 3.1, potential attackers (A) may exist either among the participating clients or as external malicious users. To effectively resist potential attacks, our algorithm needs to possess the following characteristics:
Fidelity: In the field of digital watermarking, fidelity refers to whether the quality, accuracy, and integrity of the original data are effectively maintained after embedding the watermark.
Detectability: Watermark detectability denotes the ability to extract and successfully identify watermark information from a model that has embedded the watermark. A watermark with high detectability can be accurately discovered and extracted during the verification phase.
Robustness: Robustness is the most important characteristic of watermarking algorithms, and the watermark embedded in the model should have strong resistance to various attacks while maintaining the performance of the model.
Secrecy: A good watermark should be imperceptible, meaning that users or consumers should not be able to detect the watermark’s presence when using the model for predictions.
False negative ratio: The false negative ratio indicates the probability that watermark samples are incorrectly classified as clean categories during watermark verification. A lower false negative ratio signifies higher accuracy in watermark detection.

4. Proposed Method

This section presents the watermark dataset generation, watermark embedding, and detection procedures of the proposed watermarking scheme. During the process of generating the watermark dataset, we create a user-specific watermark, which is used to construct a user-specific, private watermark dataset. During the watermark embedding phase, the model is trained using both the watermark dataset and the local original dataset. A new watermark class is introduced in the output layer for training, achieving watermark embedding. Subsequently, restore the model structure, and the parameters of the watermark class are saved to ensure their secrecy and invisibility, thus enabling copyright traceability and identification. In the watermark detection phase, an independent third-party certification authority is responsible for generating the watermark dataset and verifying the ownership of the suspicious model using the results obtained from the watermark dataset.

4.1. Watermark Dataset Generation

The watermark dataset is constructed by randomly selecting a small number of samples from the training data. Our method involves randomly picking samples from each class. The creation of the watermark dataset includes selecting text or identifiers associated with identity information and embedding them into designated locations within the samples. In this work, the watermark is embedded in the lower right corner of the samples (Figure 1). The process can be mathematically expressed as follows:
D w W i j = D i j + P 0 < i < C , j [ 0 , Z ]
Here, D w represents the watermark dataset, W i j refers to each sample in the watermark dataset, D i j denotes the original sample, and P signifies a particular pattern. In this equation, i is the class index, j indicates the sample index randomly selected from each class (with a selection of 5 samples), and Z denotes the maximum number of samples per class.

4.2. Watermark Embedding

The watermark dataset has its own labels, which are independent of the original dataset. Therefore, our expected goal during watermark embedding is that in the watermarked models, watermark samples are classified into the corresponding watermark classes, as described in Equation (2). In the clean models, since there are no classes corresponding to the watermark, watermark samples are classified based on content other than the watermark pattern, as described in Equation (3).
τ c = δ w ( X c ) τ w = δ w ( X w )
τ c = δ c ( X c ) τ c = δ c ( X w )
here, X and δ represent the input sample and the model, respectively. τ represents the predicted class of the sample by the model, with the subscripts c and w denoting clean (no watermark) and watermarked, respectively.
During the watermark embedding process, participants in the training are divided into two groups: the model initiator, referred to as client O, and the remaining clients (Figure 2). During the model training phase, the server and clients work with a shared model configured for N classifications. Under this framework, the other clients train solely on their local datasets without any additional operations. In contrast, client O trains on both the local dataset and the watermark dataset.
After training, the locally updated models from all clients are uploaded to the server for aggregation, and the aggregated model is distributed back to each client for the next training round. This process is detailed in Algorithm 1, which illustrates the steps of watermark embedding performed by the owner client and the subsequent server-side aggregation: Entropy 27 00134 i001
Step 1: As defined in Equation (4), during the local training rounds of client O, after receiving the model parameters G i , client O modifies the model’s output layer from N classes to N + 1 classes. Subsequently, the original task dataset is combined with the generated watermark dataset to form D a l l .
D a l l = D o + D w
where D o represents the local dataset of client O, D w denotes the watermark dataset, and D a l l is the dataset composed of both. During the local training process, the selected loss function updates the parameters based on the dataset D a l l for local training of the model parameters. Equation (5) defines the loss function, which is consistent with the typical federated learning loss function.
W o ( t + 1 ) = arg min W 1 N o ( x o j , y o j ) D a l l L ( W , x o j , y o j )
where W o ( t + 1 ) is the model parameters updated by participant o after t rounds of training. N o means the size of the local training dataset owned by participant o. j is the index of the dataset. L ( W , x o j , y o j ) denotes the loss function used to quantify the discrepancy between the model’s output for the input sample x o j and its corresponding true label y o j .
Step 2: After completing local training, the watermark parameters in the output layer are stored, and the model’s output is adjusted from N + 1 classes back to N classes (as shown in Figure 2). This step aims to conceal the watermark class information and ensure that the local model of client O is consistent with the structure of the aggregated model. Since the model has no commercial value in the initial training stages, the watermark embedding begins when the model reaches commercial value (epoch = e) and continues until the end of training (epoch = E), as shown in Algorithm 1.
G ( t + 1 ) = G t + 1 m × i = 1 m ( W i ( t + 1 ) G t )
G ( t + 1 ) = G t + η n × i = 1 m ( W i ( t + 1 ) G t )
Additionally, during training, client O may not be selected for model aggregation in every iteration, meaning that the watermark information from its local training may not be sent to the server for aggregation every time. To address this issue, we introduce a scaling factor ξ [44] to amplify the client’s parameters. This is a widely used mechanism aimed at adjusting each client’s contribution to model updates, thereby balancing the differences in data volume and importance among clients.
i = 1 m 1 ( W i ( t + 1 ) G t ) 0
W i ( t + 1 ) n η × ( G ( t + 1 ) G t ) + G t
Global model aggregation can be expressed by Equation (6). We modify it to Equation (7), where η represents the global learning rate. When η = n m , these two equations are equivalent. Due to the nature of federated learning, the aggregation process gradually reduces the bias until balanced, as shown in Equation (8). Therefore, Equation (7) can be simplified to Equation (9), where ξ = n η , indicating that the weight W i ( t + 1 ) is amplified by a factor of ξ .
ξ = C l i e n t a l l C l i e n t s e c t
As noted in [44], when n η = C l i e n t a l l C l i e n t s e t , smaller weights can be well preserved. Thus, in our training, we use Equation (10) and adopt the method in Equation (11) to amplify the watermark-containing weights by a factor of ξ [44], addressing the issue of the watermark not being aggregated in every iteration.
W o t = ξ × W o t
Here, C l i e n t a l l is the total number of clients participating in the training, and C l i e n t s e c t is the number of clients selected for aggregation in a particular iteration. W o t represents the local model parameters of client O uploaded to the server, which are amplified by ξ before uploading. Subsequently, client O uploads the model to the server for further model aggregation and updating.
Step 3: In federated learning, model updates are performed by the server through the averaging of the local model parameters received from the clients. This process involves combining the parameters from each client’s local model to form a global model, which reflects the collective contributions of all participants. The model aggregation procedure is as follows:
G ( t + 1 ) = i = 1 N D i D W i ( t + 1 )
Here, G ( t + 1 ) represents the parameters of the global model at the ( t + 1 ) -th round, while W i ( t + 1 ) refers to the parameters of the local model of the i-th participant at the same round. D and D i represent the entire dataset and the dataset of the i-th client, respectively. The aggregation of local model parameters from all participants is represented by the summation symbol ∑ in the formula.
By iterating through steps 1 to 3, the model undergoes iterative training and watermark embedding until it is aggregated and the embedded watermark can be reliably extracted, obtaining the final watermarked model. This process not only ensures the effective preservation and utilization of watermark information but also optimizes the N + 1 classification model. During local training, we train watermark datasets so that the model performs watermark embedding while maintaining the performance of the original task.

4.3. Watermark Verification

If a suspected organization uses the model M ^ without authorization from the model owner O, the owner has the right to involve an independent third-party certification agency to confirm the model’s copyright and produce relevant documentation. The watermark verification procedure is illustrated in Figure 3 and detailed in Algorithm 2. The third-party agency can carry out the watermark verification by following these steps:
Step 1: The model owner provides parameters P a r a W and watermark patterns, and third-party organizations can randomly select samples to generate watermark verification sets.
Step 2: The trained watermark class parameters P a r a W provided by the model owner are added into the output layer of the target model M ^ .
Step 3: Use the suspected model M ^ to test the watermark dataset generated in Step 1. First, input the watermark samples into the target model M ^ and record the model’s predictions for these samples. Based on the predicted results and the true labels of the watermark samples, calculate the prediction accuracy of the target model on the watermark samples. Specifically, the watermark class is considered private information of the model owner. Consistent watermark patterns are used throughout the training, testing, and validation processes. By determining whether the accuracy of watermark verification exceeds the predefined threshold γ , if the threshold is surpassed, it can be concluded that the target model M ^ has been illegally stolen. The value of γ is primarily adjusted based on the size of the watermark set and the categories of the samples [20,36], and since the watermark set dynamically changes according to the size of the dataset and sample categories, γ is adjusted dynamically. Entropy 27 00134 i002

5. Experiments and Analysis

5.1. Experiment Settings

This study utilized the PyTorch 1.13.1 framework [45]. The hardware configuration included an NVIDIA GeForce RTX 4060 Ti GPU with 8 GB of memory, an Intel Core i5-13600KF CPU with 14 physical cores and 20 logical cores, and 31.82 GB of memory. The experiment involved 100 participants in a federated learning setup. The dataset was partitioned into 100 simulated client datasets based on independent and identically distributed (IID) samples, while under the non-independent and identically distributed (Non-IID) setting, the dataset was generated according to a Dirichlet distribution. The training process consisted of 250 communication rounds. In each round, the learning rate was set to 0.01, the momentum to 0.5, and the minibatch size to 32, with selected clients training locally for 10 epochs. In the comparative experiments, to ensure fairness, we applied the same parameter settings across all setups, and all experimental data were validated using the test set.

5.1.1. Dataset and Model Settings

In this study, we used the MNIST, CIFAR-10, and CIFAR-100 datasets for experiments.
MNIST [46]: This dataset contains 60,000 28 × 28 grayscale images of handwritten digits and 10,000 test images, with about 6000 training images and 1000 test images per digit. Due to its simplicity, MNIST is commonly used for quick testing and comparison of image classification models.
CIFAR-10 [47]: CIFAR-10 comprises 60,000 color images across 10 categories, with 6000 images per category, each sized 32 × 32 pixels. The dataset includes diverse samples with varying angles, sizes, lighting conditions, and backgrounds, making it a standard benchmark for image classification tasks.
CIFAR-100 [48]: CIFAR-100 contains 60,000 images divided into 100 categories, with 600 images per category. Each image is 32 × 32 pixels, covering a broader range of objects and more complex content than CIFAR-10, suitable for testing deep learning models on more challenging tasks.
To evaluate our method, we selected two well-known network architectures: VGG-16 [49] and ResNet-18 [50]. ResNet-18 consists of an initial convolutional layer, four ResBlocks, and a fully connected layer, while VGG-16 includes 13 convolutional layers and 3 fully connected layers. Our method, applied to the fully connected layer, is compatible with most models.

5.1.2. Watermark Dataset Setting

To embed the watermark successfully without wasting computational resources, we conducted a series of experiments using 5, 10, and 15 samples per class to create the watermark dataset. We examined how the size of the watermark dataset affects the primary task and the accuracy of watermark detection. Assuming the original dataset contains L classes, the sizes of the watermark dataset are 5 L, 10 L, and 15 L. According to our experimental results, it can be known that the decline in the original task accuracy caused by the watermark dataset size varying from 5 L to 15 L is insignificant, with a maximum decrease of only 0.06% (Figure 4a). This indicates that the impact of the watermark dataset size on the main task can be considered insignificant. Additionally, the figures demonstrate that a watermark dataset of 5 L samples already achieves a 100% success rate in watermark embedding (Figure 4b). In summary, choosing a dataset of 5 L samples for watermarking can efficiently embed the watermark with minimal interference in the original task. Considering the idea of minimizing computational resources, we randomly select 5 samples from each class for subsequent experiments.

5.2. Performance Analysis Under IID

5.2.1. Fidelity

A comparative analysis between the accuracy of the pristine model and the watermarked model was conducted on the test set. In the experiment, a total of 250 iterations were completed throughout the training phase. Given the absence of commercial significance in the initial stages of training, watermark embedding was deemed unnecessary. Consequently, in our experiments, watermark embedding commenced from the 100th iteration onwards.
We compared the proposed class-hidden scheme with the FedIPR [28], WAFFLE [38], FedTracker [42], and FedCRMW [51] schemes while evaluating the impact of watermark embedding on the original task accuracy (see Figure 5). To ensure that the experimental results are statistically significant and not due to random fluctuations, we conducted 100 rounds of experiments with the same parameters for each group. A comparison of accuracy before and after watermark embedding showed that the proposed scheme resulted in accuracy reductions of 0.15%, 0.13%, 0.09%, 0.07%, 0.12%, and 0.29%. In comparison, the FedIPR [28] scheme led to accuracy drops of 1.48%, 1.6%, 1.79%, 1.18%, 1.51%, and 1.43%; the WAFFLE [38] scheme showed accuracy decreases of 1.35%, 1.68%, 1.72%, 1.25%, 1.47%, and 1.55%; the FedTracker [42] scheme caused accuracy decreases of 1.93%, 2.91%, 1.8%, 1.73%, 1.28%, and 1.25%; and finally, the FedCRMW [51] scheme exhibited accuracy reductions of 1.31%, 1.62%, 1.65%, 1.22%, 1.44%, and 1.63%. Based on the results above, we can confirm that the experimental findings are statistically significant and not due to random variations and that the proposed method has the least impact on the original task. Therefore, we conclude that the proposed class-hidden embedding scheme successfully embeds watermark information while maintaining high accuracy on the original task.

5.2.2. Detectability

A watermark dataset is first constructed to verify the detectability of the model watermark. Subsequently, during the training of the local model on client O, the parameters associated with watermark information in the output layer were continuously updated and saved. Finally, the detectability of the model is evaluated by calculating the watermark detection rate of the global model.
As shown in Figure 6, it demonstrates the accuracy of the original tasks and watermark detection when embedding the watermark for the ResNet-18 and VGG-16 models on the MNIST, CIFAR-10, and CIFAR-100 datasets. For the VGG-16 model, the watermark detection accuracy is above 80% after 40 iterations and remains at 100% after 60 iterations (Figure 6a,b). We observe that the watermark detection accuracy on the ResNet-18 model surpasses 95% for the MNIST, CIFAR-10, and CIFAR-100 datasets after 10 iterations and remains at 100% after 25 iterations (Figure 6d,e). Our proposed method demonstrates that after embedding the watermark, the accuracy of the primary task remains unaffected, indicating that our approach maintains a high watermark detection capability.

5.2.3. Robustness

Watermark robustness refers to the capacity of the watermark to stay detectable and recognizable under various attacks and processing conditions. In this paper, we evaluated the robustness of the watermark by testing its resilience against fine-tuning and pruning attacks.
Fine-tuning Attack: Fine-tuning refers to the additional training of a pretrained model using an additional dataset to adapt it to a new task. This process is achieved by making small adjustments and optimizing the parameters of the existing model. In the experiment, we performed fine-tuning with a dataset that excluded watermark samples, and no parameters of any layer were frozen. Figure 7 shows the changes in the watermark detection rates of our proposed method compared with those of the other three methods after 50 fine-tuning iterations. In the CIFAR-10/VGG-16 experiment, we set the scaling factor ξ to the predefined value of 50%. As can be clearly seen from Figure 7, with this setting, watermark information is easily lost during the fine-tuning process, resulting in our method having a slightly lower watermark detection rate compared with FedIPR [28], WAFFLE [38], and FedCRMW [51]. However, in the other five experiments, we used the predefined ξ , further validating the conclusions in [44]. The final results show that our method consistently outperforms FedIPR [28], WAFFLE [38], and FedCRMW [51], maintaining a 100% detection rate. Overall, our method demonstrates good performance during the fine-tuning process.
Pruning Attack: Pruning attacks involve intentionally removing or disrupting embedded watermark information in a model by applying pruning operations. In this study, we utilized a weight-based pruning method to modify the model. Table 1 presents a comparison of the robustness of the VGG-16 model under pruning attacks. It can be observed that as the pruning ratio increases, the accuracy of all watermark schemes on the original task decreases. However, compared with the methods of FedIPR [28], WAFFLE [38], FedCRMW [51], and FedTracker [42], our proposed watermarking scheme demonstrates a smaller reduction in accuracy on the original task. Moreover, even with a pruning ratio as high as 90%, our method maintains a watermark detection rate of 100%. Overall, under the same pruning ratio, our scheme exhibits stronger resilience to pruning attacks.
Forging Attack: A forging attack is a security threat where attackers fabricate data or images to bypass security defenses. In this context, attackers may have pinpointed the location where the watermark is embedded in our model. Consequently, they could attempt to forge additional parameter weights to replace the original ones, aiming to bypass the watermark and illicitly use the model. In our experiment, we tested two models on CIFAR-10, using 90 clients for each model, with the number of clients gradually increasing. The final results, shown in Figure 8, reveal that after the weights were replaced, the model’s accuracy plummeted to about 1%. The accuracy of FedIPR [28], WAFFLE [38], FedTracker [42], and FedCRMW [51] also dropped to 1%. This accuracy indicates that the model becomes non-functional after the weight replacement, making it commercially useless, and further demonstrates the robustness of our method in resisting forging attacks.
Overwriting Attack: The overwriting attack involves feeding the model a large amount of forged data, aiming to overwrite or replace the originally embedded watermark or identifying features, thereby preventing the model from correctly recognizing or verifying the original watermark. For the method proposed in this paper, malicious clients may input numerous other watermark features in an attempt to overwrite the model’s features, leading to a decrease in the model’s watermark accuracy. Table 2 presents the accuracy of our model’s private watermark after being overwritten by 1 to 10 ambiguous watermarks. The results show that although the model’s watermark accuracy decreases after overwriting, the impact is almost negligible, with the accuracy remaining above 99%. However, the proposed method also has certain limitations, notably the inability to identify the uniqueness of watermarks. While the accuracy of the private watermark does not significantly decline, the model is also capable of verifying the ambiguous watermarks used for overwriting.

5.2.4. Secrecy

The proposed scheme combines the watermark dataset with client O’s local dataset for training. Before aggregation, the watermark class parameters are removed from the model’s output layer and stored locally, ensuring that the watermark dataset associated with the watermark class remains on the client side and is not shared with other participants. Additionally, the labels of the watermark dataset belong to independent categories outside the predefined labels. If client O does not provide the watermark parameters in the output layer, predictions on the watermark dataset will use its original semantic information labels. Only when the watermark dataset and the trained output layer watermark parameters are correctly provided, can the model output the correct labels. For other clients or observers who do not have the watermark class parameters, the model’s predicted labels for the watermark dataset will correspond to its original semantic information, making it impossible to infer the existence of the watermark.
As shown in Figure 9, the activation maps and weights of the watermarked model are identical to those of the clean model, demonstrating that the proposed watermarking scheme exhibits strong secrecy, making it difficult for attackers to perceive the presence of a watermark in the model.

5.2.5. False Negative Ratio

The false negative rate refers to the probability that watermark samples are predicted as non-watermark categories during watermark verification. To evaluate the false negative rate, we separately assessed watermark verification on the CIFAR-10 dataset using both a clean ResNet-18 model and a ResNet-18 model embedded with watermarks. The results show that the clean model accurately classified watermark samples into their original semantic categories (Figure 10a). For models containing watermarks, the false negative rate can be categorized into two scenarios. The first scenario involves unaltered models with embedded watermarks, where the watermark dataset detection rate is 100% (Figure 10b), and the false negative rate is 0. The second scenario involves models with watermarks that have undergone some pruning or fine-tuning; while the detection accuracy of watermark samples may decrease, the watermark can still be detected (Figure 10c). In summary, whether in a clean model, a model embedded with watermarks, or a model subjected to attacks after watermark embedding, the false negative rate of the watermark proposed in this paper is very low, and its accuracy is sufficient to support copyright verification.
Additionally, we conducted watermark verification experiments on different types of samples, including original watermark samples, clean samples, and ambiguous watermark samples, using VGG-16 and ResNet-18 models. To enhance the credibility of the experimental results, we conducted 90 rounds of experiments, averaging every 10 rounds. As shown in Table 3, the verification success rate for original watermark samples reached 100%, the success rate for clean samples was 0%, and the success rate for ambiguous watermark samples was around 65%. Based on these results, it can be effectively demonstrated that the proposed method has a high watermark verification success rate.

5.2.6. Efficiency Evaluation

We compared the efficiency of our method with recent white-box watermarking algorithms on the VGG-16 model. As shown in Table 4, the global model’s training time increased by only 0.13 s, 0.07 s, and 0.08 s, indicating that our watermarking approach added virtually no extra training burden during a single global training round.
In contrast, white-box methods such as FedTracker [42], FedIPR [28], and FedCRMW [51] employed more complex watermark embedding strategies, which resulted in additional time for embedding and training watermarks during the global model training process. Specifically, FedTracker [42] increased training time by 57.31 s, 56.71 s, and 37.07 s across the three datasets, FedIPR [28] added 104.04 s, 115.8 s, and 108.85 s, and FedCRMW [51] increased it by 67.74 s, 60.12 s, and 52.70 s, respectively.
In terms of verification efficiency, our method is comparable to other algorithms, requiring only 0.064 milliseconds per verification, while FedIPR [28] requires approximately 0.09 milliseconds, FedCRMW [51] 0.077 milliseconds, and FedTracker [42] 0.081 milliseconds.
The experimental results clearly demonstrate that our method achieves higher verification efficiency with minimal training overhead, highlighting its notable advantages.

5.3. Performance Analysis Under Non-IID

In federated learning, the non-independent and identically distributed (Non-IID) nature of client data is a common issue. Specifically, the datasets held by each client exhibit significant heterogeneity, with notable differences in labels and sample sizes within their training sets. Non-IID data negatively impact the effectiveness of federated learning. Given that Non-IID scenarios are more prevalent in real-world applications, experiments under such conditions are more meaningful.
We conducted experiments under Non-IID conditions using CIFAR-10, MNIST, and CIFAR-100 datasets, with VGG-16 as the chosen model. As shown in Figure 11, our method not only maintained the original task accuracy but also slightly improved it, with the watermark accuracy remaining around 98%. This demonstrates that our method remains effective even under Non-IID conditions.

5.3.1. Robustness Against Fine-Tuning Attack

Similar to fine-tuning in the IID setting, we fine-tuned the client’s model using the client’s own dataset. The key difference is that, in the Non-IID setting, the data from different categories in the client’s dataset are imbalanced. The experimental results are shown in Figure 12.
As shown in Figure 12a, after 50 rounds of fine-tuning, the accuracy of the original task did not change significantly, with an overall variation of about 0.3%. From Figure 12b, it can be observed that, except for the CIFAR-10 dataset, where the watermark accuracy dropped to 85.4%, likely due to severe class imbalance in the client’s training data causing the model to fall into a local optimum, the watermark accuracy for the other two datasets did not show any noticeable change and remained above 98%. This conclusion is consistent with the results in the IID setting, indicating that after fine-tuning, our model is still able to perform well enough to support the verification task.

5.3.2. Robustness Against Pruning Attack

We also conducted pruning experiments on the model under the Non-IID setting, with the pruning rate gradually increasing from 10% to 90%. As shown in Figure 13a, the accuracy of the original task does not significantly decrease until the pruning rate exceeds 60%. Figure 13b clearly shows that even when the pruning rate reaches 90%, our watermark accuracy remains above 98%. This strongly demonstrates that, even under Non-IID conditions, our model retains high robustness against pruning attacks.

5.3.3. Robustness Against Overwriting Attack

To evaluate whether the vulnerability of the watermark increases under the Non-IID condition, we conducted overwriting attack experiments on the model using the same attack setup as in the IID setting. The results are shown in Figure 14a. Whether applying one or ten ambiguous watermarks, our model still maintained an accuracy of over 98%. Although the model performs excellently in terms of watermark accuracy, unfortunately, it cannot achieve uniqueness. Additionally, we conducted an ambiguous watermark detection test, and as shown in the right-side of Figure 14b, the accuracy of each ambiguous watermark reached up to 60%, highlighting a limitation of our approach.

6. Conclusions

This paper introduces a class-hidden watermarking scheme to mitigate the impact of black-box watermarking on the accuracy of the original task in federated learning. The method modifies the local model structure during client training by adding a watermark class to embed the watermark. To maintain stealth, the parameters of the watermark class are hidden when uploading the model parameters. Experimental results show that, whether in IID or Non-IID scenarios, after embedding the watermark, the accuracy of the original classification task remains high, with the watermark’s impact on the original classification task being less than 0.3%, which is almost negligible. This scheme not only demonstrates good robustness, effectively resisting fine-tuning, pruning, forging, and overwriting attacks but also preserves the secrecy of the watermark. Therefore, in the context of federated learning, this scheme shows significant potential for practical applications, providing security guarantees for model safety in federated learning environments. Future work could focus on identifying malicious users, which would not only help with copyright verification but also enable the accurate tracking of leakage sources, identifying the specific client responsible for the leak. This may involve exploring user identification techniques and developing robust mechanisms to trace data leakage back to its origin, thus enhancing security measures in federated learning environments.

Author Contributions

Conceptualization, W.C. and J.C.; methodology, W.C. and W.Z.; software, W.Z.; data curation, W.Z. and C.Z.; writing—original draft preparation, C.Z. and W.Z.; writing—review and editing, W.C. and J.C.; funding acquisition: W.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (No. 42201444).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The associated datasets “MNIST”, “CIFAR-10” and “CIFAR-100” used for demonstration are publicly accessible from “http://yann.lecun.com/exdb/mnist/, accessed on 18 March 2024 ”. The links for “CIFAR-10” and “CIFAR-100” are “https://www.cs.toronto.edu/~kriz/cifar.html, accessed on 18 March 2024”, respectively.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-Efficient Learning of Deep Networks from Decentralized Data. PMLR 2017, 54, 1273–1282. [Google Scholar]
  2. Wu, D.; Wang, N.; Zhang, J.; Zhang, Y.; Xiang, Y.; Gao, L. A Blockchain-based Multi-layer Decentralized Framework for Robust Federated Learning. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar] [CrossRef]
  3. Salem, A.; Zhang, Y.; Humbert, M.; Berrang, P.; Fritz, M.; Backes, M. ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models. In Proceedings of the NDSS, San Diego, CA, USA, 24–27 February 2019; pp. 24–27. [Google Scholar]
  4. Liu, L.; Wang, Y.; Liu, G.; Peng, K.; Wang, C. Membership Inference Attacks Against Machine Learning Models via Prediction Sensitivity. IEEE Trans. 2023, 20, 2341–2347. [Google Scholar] [CrossRef]
  5. Zhu, C.; Zhang, J.; Sun, X.; Chen, B.; Meng, W. ADFL: Defending backdoor attacks in federated learning via adversarial distillation. Comput. Secur. 2023, 132, 103366. [Google Scholar] [CrossRef]
  6. Oh, S.J.; Schiele, B.; Fritz, M. Towards Reverse-Engineering Black-Box Neural Networks. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning; Springer: Cham, Switzerland, 2019; pp. 121–144. [Google Scholar] [CrossRef]
  7. Oh, S.J.; Augustin, M.; Schiele, B.; Fritz, M. Towards Reverse-Engineering Black-Box Neural Networks. arXiv 2018, arXiv:1711.01768. [Google Scholar] [CrossRef]
  8. Orekondy, T.; Schiele, B.; Fritz, M. Knockoff Nets: Stealing Functionality of Black-Box Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4954–4963. [Google Scholar] [CrossRef]
  9. Wang, B.; Gong, N.Z. Stealing Hyperparameters in Machine Learning. In Proceedings of the 2018 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 20–24 May 2018; pp. 36–52. [Google Scholar]
  10. Yan, M.; Fletcher, C.W.; Torrellas, J. Cache Telepathy: Leveraging Shared Resource Attacks to Learn DNN Architectures. In Proceedings of the USENIX, Online, 12–14 August 2020; pp. 2003–2020. [Google Scholar] [CrossRef]
  11. Jagielski, M.; Carlini, N.; Berthelot, D.; Kurakin, A.; Papernot, N. High Accuracy and High Fidelity Extraction of Neural Networks. In Proceedings of the USENIX, Online, 12–14 August 2020; pp. 1345–1362. [Google Scholar]
  12. Zhu, Y.; Cheng, Y.; Zhou, H.; Lu, Y. Hermes Attack: Steal DNN Models with Lossless Inference Accuracy. In Proceedings of the USENIX, Virtual Event, 14–16 July 2021; pp. 1973–1988. [Google Scholar]
  13. Wang, J.; Wu, H.; Zhang, X.; Yao, Y. Watermarking in Deep Neural Networks via Error Back-propagation. Soc. Imaging Sci. Technol. 2020, 32, 1–9. [Google Scholar] [CrossRef]
  14. Namba, R.; Sakuma, J. Robust Watermarking of Neural Network with Exponential Weighting. In Proceedings of the ACM, Auckland, New Zealand, 9–12 July 2019; pp. 228–240. [Google Scholar] [CrossRef]
  15. Ong, D.S.; Chan, C.S.; Ng, K.W.; Fan, L.; Yang, Q. Protecting Intellectual Property of Generative Adversarial Networks From Ambiguity Attacks. In Proceedings of the IEEE, CVPR, Nashville, TN, USA, 19–25 June 2021; pp. 3630–3639. [Google Scholar] [CrossRef]
  16. Chen, X.; Chen, T.; Zhang, Z.; Wang, Z. You are caught stealing my winning lottery ticket! Making a lottery ticket claim its ownership. In Proceedings of the NeurIPS, Online, 6–14 December 2021; pp. 1780–1791. [Google Scholar] [CrossRef]
  17. Lim, J.H.; Chan, C.S.; Ng, K.W.; Fan, L.; Yang, Q. Protect, show, attend and tell: Empowering image captioning models with ownership protection. Pattern Recognit. 2022, 122, 108285. [Google Scholar] [CrossRef]
  18. April Pyone, M.M.; Kiya, H. Piracy-Resistant DNN Watermarking by Block-Wise Image Transformation with Secret Key. In Proceedings of the ACM, Virtual Event, 22–25 June 2021; pp. 159–164. [Google Scholar] [CrossRef]
  19. Fan, L.; Ng, K.W.; Chan, C.S.; Yang, Q. DeepIPR: Deep Neural Network Ownership Verification With Passports. IEEE Trans 2022, 44, 6122–6139. [Google Scholar] [CrossRef] [PubMed]
  20. Szyller, S.; Atli, B.G.; Marchal, S.; Asokan, N. DAWN: Dynamic Adversarial Watermarking of Neural Networks. In Proceedings of the ACM, Virtual Event, 22–25 June 2021; pp. 4417–4425. [Google Scholar] [CrossRef]
  21. Fkirin, A.; Attiya, G.; El-Sayed, A.; Shouman, M.A. Copyright protection of deep neural network models using digital watermarking: A comparative study. Multim. Tools Appl. 2022, 81, 15961–15975. [Google Scholar] [CrossRef] [PubMed]
  22. Han, B.; Jhaveri, R.H.; Wang, H.; Qiao, D.; Du, J. Application of Robust Zero-Watermarking Scheme Based on Federated Learning for Securing the Healthcare Data. IEEE J. 2023, 27, 804–813. [Google Scholar] [CrossRef] [PubMed]
  23. Xue, M.; Sun, S.; Zhang, Y.; Wang, J.; Liu, W. Active intellectual property protection for deep neural networks through stealthy backdoor and users’ identities authentication. Appl. Intell. 2022, 52, 16497–16511. [Google Scholar] [CrossRef]
  24. Jia, H.; Choquette-Choo, C.A.; Chandrasekaran, V.; Papernot, N. Entangled Watermarks as a Defense against Model Extraction. In Proceedings of the USENIX, Virtual Event, 14–16 July 2021; pp. 1937–1954. [Google Scholar] [CrossRef]
  25. Rouhani, B.D.; Chen, H.; Koushanfar, F. DeepSigns: An End-to-End Watermarking Framework for Ownership Protection of Deep Neural Networks. In Proceedings of the ACM ASPLOS, Providence, RI, USA, 13–17 April 2019; pp. 485–497. [Google Scholar] [CrossRef]
  26. Shao, S.; Yang, W.; Gu, H.; Lou, J.; Qin, Z.; Fan, L.; Yang, Q.; Ren, K. FedTracker: Furnishing Ownership Verification and Traceability for Federated Learning Model. arXiv 2022, arXiv:2211.07160. [Google Scholar] [CrossRef]
  27. Chen, J.; Li, M.; Cheng, Y.; Zheng, H. FedRight: An effective model copyright protection for federated learning. Comput. Secur. 2023, 135, 103504. [Google Scholar] [CrossRef]
  28. Li, B.; Fan, L.; Gu, H.; Li, J.; Yang, Q. FedIPR: Ownership Verification for Federated Deep Neural Network Models. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 4521–4536. [Google Scholar] [CrossRef] [PubMed]
  29. Liang, J.; Wang, R. FedCIP: Federated Client Intellectual Property Protection with Traitor Tracking. arXiv 2023, arXiv:2306.01356. [Google Scholar] [CrossRef]
  30. Yin, Z.; Yin, H.; Zhang, X. Neural Network Fragile watermarking With No Model Performance Degradation. In Proceedings of the IEEE ICIP, Bordeaux, France, 16–19 October 2022; pp. 3958–3962. [Google Scholar] [CrossRef]
  31. Uchida, Y.; Nagai, Y.; Sakazawa, S.; Satoh, S. Embedding Watermarks into Deep Neural Networks. In Proceedings of the ACM ICMR, Bucharest, Romania, 6–9 June 2017; pp. 269–277. [Google Scholar] [CrossRef]
  32. Liu, H.; Weng, Z.; Zhu, Y. Watermarking Deep Neural Networks with Greedy Residuals. ICML 2021, 139, 6978–6988. [Google Scholar]
  33. Yan, Y.; Pan, X.; Zhang, M.; Yang, M. Rethinking White-Box Watermarks on Deep Learning Models under Neural Structural Obfuscation. In Proceedings of the USENIX, Anaheim, CA, USA, 9–11 August 2023; Available online: https://www.usenix.org/conference/usenixsecurity23/presentation/yan (accessed on 18 March 2024).
  34. Zhang, J.; Chen, D.; Liao, J.; Zhang, W.; Hua, G.; Yu, N. Passport-aware Normalization for Deep Model Protection. In Proceedings of the NeurIPS, Virtual, 6–12 December 2020. [Google Scholar] [CrossRef]
  35. Fan, L.; Ng, K.W.; Chan, C.S. Rethinking Deep Neural Network Ownership Verification: Embedding Passports to Defeat Ambiguity Attacks. In Proceedings of the NeurIPS, Vancouver, BC, Canada, 8–14 December 2019; pp. 4716–4725. [Google Scholar]
  36. Adi, Y.; Baum, C.; Cissé, M.; Pinkas, B.; Keshet, J. Turning Your Weakness Into a Strength: Watermarking Deep Neural Networks by Backdooring. In Proceedings of the USENIX, Baltimore, MD, USA, 15–17 August 2018; pp. 1615–1631. [Google Scholar]
  37. Li, M.; Zhong, Q.; Zhang, L.Y.; Du, Y.; Zhang, J.; Xiang, Y. Protecting the Intellectual Property of Deep Neural Networks with Watermarking: The Frequency Domain Approach. In Proceedings of the 2020 IEEE 19th International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom), Guangzhou, China, 29 December–1 January 2020; pp. 402–409. [Google Scholar] [CrossRef]
  38. Tekgul, B.G.A.; Xia, Y.; Marchal, S.; Asokan, N. WAFFLE: Watermarking in Federated Learning. In Proceedings of the IEEE SRDS, Chicago, IL, USA, 20–23 September 2021; pp. 310–320. [Google Scholar] [CrossRef]
  39. Yang, W.; Shao, S.; Yang, Y.; Liu, X.; Liu, X.; Xia, Z.; Schaefer, G.; Fang, H. Watermarking in Secure Federated Learning: A Verification Framework Based on Client-Side Backdooring. ACM Trans. Intell. Syst. Technol. 2024, 15, 5:1–5:25. [Google Scholar] [CrossRef]
  40. Li, F.Q.; Wang, S.L.; Liew, A.W.C. Towards practical watermark for deep neural networks in federated learning. arXiv 2021, arXiv:2105.03167. [Google Scholar] [CrossRef]
  41. Becker, G. Merkle signature schemes, merkle trees and their cryptanalysis. Ruhr-Univ. Bochum Tech. Rep. 2008, 12, 19. [Google Scholar]
  42. Shao, S.; Yang, W.; Gu, H.; Qin, Z.; Fan, L.; Yang, Q. FedTracker: Furnishing Ownership Verification and Traceability for Federated Learning Model. IEEE Trans. Dependable Secur. Comput. 2024, 22, 114–131. [Google Scholar] [CrossRef]
  43. Liu, X.; Shao, S.; Yang, Y.; Wu, K.; Yang, W.; Fang, H. Secure Federated Learning Model Verification: A Client-side Backdoor Triggered Watermarking Scheme. In Proceedings of the IEEE SMC, Melbourne, Australia, 17–20 October 2021; pp. 2414–2419. [Google Scholar] [CrossRef]
  44. Bagdasaryan, E.; Veit, A.; Hua, Y.; Estrin, D.; Shmatikov, V. How To Backdoor Federated Learning. PMLR 2020, 108, 2938–2948. [Google Scholar]
  45. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic Differentiation in PyTorch. 2017. Available online: https://api.semanticscholar.org/CorpusID:40027675 (accessed on 18 March 2024).
  46. LeCun, Y.; Cortes, C. Mnist Handwritten Digit Database; ATT Labs: New York, NY, USA, 2010. [Google Scholar]
  47. Alex Krizhevsky, G.H. Learning Multiple Layers of Features from Tiny Images; Technical Report; University of Toronto: Toronto, ON, Canada, 2009. [Google Scholar]
  48. Krizhevsky, A.; Nair, V.; Hinton, G. CIFAR-100 (Canadian Institute for Advanced Research). 2009. Available online: https://www.cs.toronto.edu/~kriz/cifar.html (accessed on 18 March 2024).
  49. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the ICLR, San Diego, CA, USA, 7–9 May 2015; Available online: http://arxiv.org/abs/1409.1556 (accessed on 18 March 2024).
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE CVPR, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef]
  51. Nie, H.; Lu, S. FedCRMW: Federated model ownership verification with compression-resistant model watermarking. Expert Syst. Appl. 2024, 249, 123776. [Google Scholar] [CrossRef]
Figure 1. Samples in the watermark dataset.
Figure 1. Samples in the watermark dataset.
Entropy 27 00134 g001
Figure 2. Watermark embedding process.
Figure 2. Watermark embedding process.
Entropy 27 00134 g002
Figure 3. Watermark verification process.
Figure 3. Watermark verification process.
Entropy 27 00134 g003
Figure 4. The impact of watermark dataset size: (a) The impact on the accuracy of the original task. (b) The impact on watermark detection rate.
Figure 4. The impact of watermark dataset size: (a) The impact on the accuracy of the original task. (b) The impact on watermark detection rate.
Entropy 27 00134 g004
Figure 5. Comparison of original task accuracy.
Figure 5. Comparison of original task accuracy.
Entropy 27 00134 g005
Figure 6. Accuracy of original task and watermark across different training epochs.
Figure 6. Accuracy of original task and watermark across different training epochs.
Entropy 27 00134 g006
Figure 7. Comparison of watermark accuracy across fine-tuning epochs.
Figure 7. Comparison of watermark accuracy across fine-tuning epochs.
Entropy 27 00134 g007
Figure 8. Comparison of original task accuracy against forging attack.
Figure 8. Comparison of original task accuracy against forging attack.
Entropy 27 00134 g008
Figure 9. The activation maps and weight distributions of the watermarked model and the clean model on VGG-16.
Figure 9. The activation maps and weight distributions of the watermarked model and the clean model on VGG-16.
Entropy 27 00134 g009
Figure 10. Distribution of prediction results for watermark datasets on clean and watermarked models: (a) Clean model. (b) Unmodified watermarked model. (c) Modified watermarked model.
Figure 10. Distribution of prediction results for watermark datasets on clean and watermarked models: (a) Clean model. (b) Unmodified watermarked model. (c) Modified watermarked model.
Entropy 27 00134 g010
Figure 11. Accuracy of original task and watermark under Non-IID conditions.
Figure 11. Accuracy of original task and watermark under Non-IID conditions.
Entropy 27 00134 g011
Figure 12. Accuracy of original task and watermark against fine-tuning under Non-IID conditions.
Figure 12. Accuracy of original task and watermark against fine-tuning under Non-IID conditions.
Entropy 27 00134 g012
Figure 13. Accuracy of original task and watermark against pruning under Non-IID conditions.
Figure 13. Accuracy of original task and watermark against pruning under Non-IID conditions.
Entropy 27 00134 g013
Figure 14. Watermark accuracy against overwriting attack under Non-IID conditions.
Figure 14. Watermark accuracy against overwriting attack under Non-IID conditions.
Entropy 27 00134 g014
Table 1. Accuracy of original task and watermark under different pruning rates on VGG-16.
Table 1. Accuracy of original task and watermark under different pruning rates on VGG-16.
DatasetsEvaluation
Metrics
MethodsPruning Rate
10% 20% 30% 40% 50% 60% 70% 80% 90%
cifar-10Original
Task
Acc (%)
FedIPR84.1584.0883.8283.7782.1167.1846.7521.7713.24
WAFFLE80.3379.6579.7876.6171.3530.1325.8715.989.74
FedCRMW85.3285.2985.2585.1784.3515.0210.0510.0510.00
FedTracker85.1284.9784.8884.8676.3325.0515.8111.836.17
Proposed84.9584.9084.8284.8084.6484.5964.5529.1915.91
Watermark
Acc (%)
FedIPR10010010010010054.1232.1515.4210.37
WAFFLE10010098.5598.5580.3180.3376.1566.4132.98
FedCRMW98.4198.4198.4198.4198.4115.876.356.356.35
FedTracker10010099.1190.1682.1310.5217.7310.927.28
Proposed100100100100100100100100100
mnistOriginal
Task
Acc (%)
FedIPR99.5699.4699.4499.3199.1298.9265.7232.1315.17
WAFFLE99.1599.1299.0498.9997.2176.5232.1315,9110.11
FedCRMW98.8198.8198.898.3591.4822.1511.3511.3511.35
FedTracker99.1999.1198.1595.3890.7162.1446.1728.9313.84
Proposed99.3199.3299.3499.3499.3587.3269.2750.0529.94
Watermark
Acc (%)
FedIPR10010010098.6198.2388.9065.3146.2134.18
WAFFLE10010010010091.3480.2350.4236.114.87
FedCRMW10010010088.8949.2123.817.947.947.94
FedTracker10010010095.1751.8435.9621.7517.966.92
Proposed100100100100100100100100100
cifar-100Original
Task
Acc (%)
FedIPR64.2164.1764.0145.1845.7532.1220.2115.147.19
WAFFLE60.1960.0659.8859.6631.2116.908.114.182.32
FedCRMW60.2660.1160.0641.2941.1120.9518.337.314.92
FedTracker64.4164.1963.1651.8343.7326.1813.849.175.73
Proposed64.5564.5563.1763.0740.1520,8416.3114.298.59
Watermark
Acc (%)
FedIPR10010010096.1996.1460.4254.8632.1815.81
WAFFLE10010010010049.1631.5819.284.130.00
FedCRMW10010010079.1548.9237.9520.7315.756.95
FedTracker10010010087.9260.1751.8539.7727.1510.83
Proposed100100100100100100100100100
Table 2. Watermark accuracy under different numbers of ambiguous watermarks.
Table 2. Watermark accuracy under different numbers of ambiguous watermarks.
MethodsModelsDatasetsNumber of Ambiguous Watermarks
1 2 3 4 5 6 7 8 9 10
ProposedVGG-16CIFAR-1099.3499.2499.2599.3199.3199.2599.3499.399.3499.35
MNIST99.4999.4499.4699.4599.5599.5499.5699.5599.4799.48
CIFAR-10099.0899.0699.1499.0499.1699.1399.1199.0599.1099.05
ResNet-18CIFAR-1099.5399.4499.4699.5199.5499.5699.4599.599.599.48
MNIST99.6399.6399.699.5799.5599.6299.5999.6599.5999.64
CIFAR-10099.2899.2899.2799.2999.399.399.2799.2499.2999.34
WAFFLEVGG-16CIFAR-1065.1165.1265.0965.0865.1365.1465.165.1465.1465.11
MNIST70.0870.0870.1370.1570.0870.170.1370.0970.1670.06
CIFAR-10059.2959.2559.2659.2459.2659.359.2859.2559.2559.32
ResNet-18CIFAR-1069.3469.2869.2769.3569.3269.3569.3369.2869.3569.36
MNIST73.7873.7973.8173.7573.7873.7973.7973.8173.8373.85
CIFAR-10062.7862.7862.7762.8162.7762.8162.8362.8362.8562.82
FedIPRVGG-16CIFAR-1087.6587.6687.6587.7187.6687.7187.7487.787.6787.73
MNIST89.7589.7589.7589.7589.6689.6589.7489.6789.6989.71
CIFAR-10075.7275.6475.6575.7175.7475.6475.7275.7575.7175.71
ResNet-18CIFAR-1089.9489.9489.8489.9489.9289.8689.9289.8689.8489.93
MNIST91.9591.8791.9691.8691.8591.991.8591.8491.9191.96
CIFAR-10080.9580.9180.8580.8980.980.9280.8880.8980.8680.95
FedCRMWVGG-16CIFAR-1075.3575.4375.3775.3875.3675.3775.4575.4275.3475.42
MNIST79.5679.5779.6579.5679.5979.5879.6179.6179.6279.61
CIFAR-10071.6171.6471.6271.5971.6571.6271.6371.6371.5771.58
ResNet-18CIFAR-1087.5687.6587.6587.6387.5487.5887.5687.6587.6387.6
MNIST85.1685.0585.185.185.1385.1285.1185.1485.185.12
CIFAR-10081.1681.0581.1481.1281.0781.0681.0781.0881.0581.14
FedTrackerVGG-16CIFAR-1091.4691.4691.5591.5591.4991.4691.5691.591.5191.53
MNIST90.9490.9290.9190.9190.9190.9290.9390.8690.8790.87
CIFAR-10093.8593.8493.9193.8593.8693.8793.993.8793.8993.92
ResNet-18CIFAR-1096.1296.0696.0896.1596.0796.0696.1696.0896.0896.15
MNIST95.6895.7695.6795.7195.6795.6995.7695.6895.6595.76
CIFAR-10094.3694.3494.3394.394.3494.2694.2794.3694.3694.29
Table 3. Comparison of watermark accuracy with different types of samples.
Table 3. Comparison of watermark accuracy with different types of samples.
ModelAcc(%)DatasetsNumber of Experiments
10 20 30 40 50 60 70 80 90
VGG-16WatermarkCIFAR-10100100100100100100100100100
MNIST100100100100100100100100100
CIFAR-100100100100100100100100100100
CleanCIFAR-100.000.000.000.000.000.000.000.000.00
MNIST0.000.000.000.000.000.000.000.000.00
CIFAR-1000.000.000.000.000.000.000.000.000.00
AmbiguousCIFAR-1063.9264.6160.5163.7761.2662.8260.9760.7363.99
MNIST61.2163.4858.8258.1662.2259.0262.6459.9959.08
CIFAR-10066.0963.8465.2364.2463.4366.8165.0165.0465.72
ResNet-18WatermarkCIFAR-10100100100100100100100100100
MNIST100100100100100100100100100
CIFAR-100100100100100100100100100100
CleanCIFAR-100.000.000.000.000.000.000.000.000.00
MNIST0.000.000.000.000.000.000.000.000.00
CIFAR-1000.000.000.000.000.000.000.000.000.00
AmbiguousCIFAR-1062.3762.5163.4663.2267.9164.1162.0965.8464.23
MNIST61.0563.3666.3361.1966.2166.4663.1061.0761.03
CIFAR-10063.2961.9358.1763.2261.4959.5161.3559.8959.06
Table 4. Efficiency experiment results for different watermarking methods.
Table 4. Efficiency experiment results for different watermarking methods.
MethodsDatasetsGlobal Iteration (s)Watermark Embedding (s)Watermark Verification (ms)
CleanCIFAR-10257.01--
MNIST233.04--
CIFAR-100290.09--
ProposedCIFAR-10257.140.130.064
MNIST233.110.070.064
CIFAR-100290.170.080.062
FedIPRCIFAR-10361.05104.040.083
MNIST348.84115.800.074
CIFAR-100398.94108.850.091
FedCRMWCIFAR-10324.7567.740.073
MNIST293.1660.120.071
CIFAR-1006343.6152.700.077
FedTrackerCIFAR-10314.3257.310.078
MNIST289.7556.710.072
CIFAR-100327.1637.070.081
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, W.; Zhang, C.; Zhang, W.; Cai, J. Class-Hidden Client-Side Watermarking in Federated Learning. Entropy 2025, 27, 134. https://doi.org/10.3390/e27020134

AMA Style

Chen W, Zhang C, Zhang W, Cai J. Class-Hidden Client-Side Watermarking in Federated Learning. Entropy. 2025; 27(2):134. https://doi.org/10.3390/e27020134

Chicago/Turabian Style

Chen, Weitong, Chi Zhang, Wei Zhang, and Jie Cai. 2025. "Class-Hidden Client-Side Watermarking in Federated Learning" Entropy 27, no. 2: 134. https://doi.org/10.3390/e27020134

APA Style

Chen, W., Zhang, C., Zhang, W., & Cai, J. (2025). Class-Hidden Client-Side Watermarking in Federated Learning. Entropy, 27(2), 134. https://doi.org/10.3390/e27020134

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop