Next Article in Journal
Research on Microbial Community Structure in Different Blocks of Alkaline–Surfactant–Polymer Flooding to Confirm Optimal Stage of Indigenous Microbial Flooding
Previous Article in Journal
Identifying Key Success Factors for Industry 4.0 Implementation: An Empirical Analysis Using SEM and fsQCA
Previous Article in Special Issue
A Graph Neural Network Approach with Improved Levenberg–Marquardt for Electrical Impedance Tomography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluating the Impact of Mobility on Differentially Private Federated Learning †

Department of Information and Telecommunication Engineering, Incheon National University, Incheon 22012, Republic of Korea
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in IEEE/IFIP Network Operations and Management Symposium, Budapest, Hungary, 25–29 April 2022.
Appl. Sci. 2024, 14(12), 5245; https://doi.org/10.3390/app14125245
Submission received: 22 May 2024 / Revised: 10 June 2024 / Accepted: 13 June 2024 / Published: 17 June 2024

Abstract

:
This paper investigates differential privacy in federated learning. This topic has been actively examined in conventional network environments, but few studies have investigated it in the Internet of Vehicles, especially considering various mobility patterns. In particular, this work aims to measure and enumerate the trade-off between accuracy of performance and the level of data protection and evaluate how mobility patterns affect it. To this end, this paper proposes a method considering three factors: learning models, vehicle mobility, and a privacy algorithm. By taking into account mobility patterns, local differential privacy is enhanced with an adaptive clipping method and applied to a mobility-based federated learning model. Experiments run the model on vehicular networks with two different mobility scenarios representing a non-accident traffic situation and traffic events, respectively. Results show that our privacy-enhanced federated learning models degrade accuracy performance by 2.96–3.26% on average, which is compared to the performance drop (42.97% on average) in conventional federated learning models.

1. Introduction

The future smart transportation system is expected to see collaboration among intelligent autonomous driving systems, named the Internet of Vehicles, which will ultimately improve road safety [1]. While recent research has applied machine learning techniques to realize this vision, security issues still remain in the process of collecting data from each vehicle and transferring them to a central server. Federated learning has been proposed to address these issues by allowing each vehicle to perform model training locally and only sending updated model parameters to the server [2]. As an emerging technology, however, it still has various technical challenges that have not yet been solved [3].
For instance, inference attacks in joint learning on autonomous vehicles, where an attacker exploits a training model to infer information that should be unknown, may cause privacy problems [4,5,6]. Recent differential privacy may alleviate this problem; it is possible to protect personal information while maintaining data distribution characteristics by adding noise to the data [7]. However, when differential privacy is applied to model training, there is a trade-off between privacy preservation and model accuracy, showing lower accuracy than non-privacy models. The primary goal of this research is to measure and enumerate this trade-off in the Internet of Vehicles. While recent studies are actively exploring the performance issue of differentially private federated learning models in conventional network environments, there are few studies that address this topic in the Internet of Vehicles, especially taking into account mobility patterns.
To clarify the accuracy of performance and protection level in the Internet of Vehicles, this paper proposes a method considering three factors: learning models, vehicle mobility, and a privacy algorithm. First, the method considers two learning models: a conventional learning model that is widely used in federated learning settings and a mobility-based federated learning model that uses vehicular knowledge generated according to vehicles’ movement patterns. Next, it imports two mobility models: reference point group mobility and random waypoint. These models represent normal situations and road events like traffic accidents, respectively. Regarding the third factor, this paper proposes to adopt a local differential privacy with adaptive clipping method after considering the nature of the Internet of Vehicles. After applying our privacy algorithm to the learning models, experiments are run with various noise values, and the accuracy performance of the models is measured under different mobility scenarios. In non-accident traffic situations, the experimental results show that privacy-enhanced federated learning models degrade performance by 3.26% on average, which is acceptable; elevated levels of data protection could compensate for the loss. In the case of traffic events, two learning models show very different outputs. With the mobility-based model, the privacy enhancement does not harm performance much, with a 2.96% drop. With the conventional model, however, the degradation rate reaches 42.97% (from 73.17% to 41.73%).
The main contributions of this paper are summarized as follows:
  • This work designs a method considering three factors—learning models, vehicle mobility, and a privacy algorithm—in order to evaluate the impact of mobility on differentially private federated learning.
  • This work modifies a differential privacy algorithm that takes the features of IoV into account. As a result, it works seamlessly on a mobility-based federated learning model.
  • The experimental results discover that a conventional learning model that is widely used in federated learning does not work well in a mobile, vehicular environment, especially when enhanced with a privacy method, and that an adaptive clipping method is more useful than prior methods because it guarantees model performance and the target degree of privacy preservation.
The rest of the paper is organized as follows: Section 2 reviews two underlying technologies: federated learning and differential privacy. Section 3 briefly introduces the intrinsic properties of the Internet of Vehicles and shows how the underlying technologies have been applied to them. Section 4 proposes a new method for evaluating the impact of mobility, including a new differential privacy algorithm running on mobility-based federated learning for the Internet of Vehicles. Section 5 sets up our experiments, which is followed by analyses of the experimental results in Section 6. Section 7 analyzes the communication cost and error convergence of the proposed learning model. Finally, Section 8 concludes this work.

2. Technology Reviews

2.1. Federated Learning

In a conventional deep learning setup, a central server collects data from geographically distributed devices and consumes all the collected data for training, testing, and updating models. This can cause communication efficiency problems and various security problems [8]. An emerging federated learning (FL) method is expected to resolve parts of the problems by not requiring raw data to be transmitted to the central server. Instead, it allows distributed clients (participants) to train a model at their side without sending data to the server. As the model is trained locally, participants only send the updated parameters to the server. Figure 1 shows the action processes of federated learning [9].
  • A server coordinator randomly selects clients to participate in federated learning. The selected participants download a current global model from the coordinator. Figure 1 assumes all four clients are selected as participants.
  • Each participant collects data locally and uses them to train the received models.
  • After training, each participant transmits the parameters of the locally updated global model to the coordinator. Then, the coordinator integrates the values received from all of the participants and learns (updates) the global model again.
  • The process is repeated from step 1 unless a predefined condition on the model is satisfied. Otherwise, the global model is finalized and distributed to all the participants.
The most notable feature of federated learning is the distribution of data and workload among training participants. Thus, local data are not easily exposed to large amounts of data breaches or sharing with third parties. This allows various parties to share learning models while protecting sensitive information. Distributing the workload of learning algorithms across participants can substantially improve performance. Due to these benefits, federated learning has been applied to a variety of scenarios.

2.1.1. Federated Learning on Mobility Applications

This work is interested in federated learning models in mobility applications. Federated learning fits latency-sensitive mobility applications well in the sense that it improves network efficiency by communicating only model parameters instead of large amounts of raw data and that a local model can quickly adapt to changes in the environment of a device. Most of the previous studies have tried to solve research problems related to the mobility of humans and moving objects such as vehicles. Human activity recognition aims to detect the movements of a human, such as sitting, running, jumping, lying down, or going upstairs, using cameras and sensors [10,11,12,13,14,15]. Location prediction focuses on finding the next locations of users [16,17,18]. Point-of-interest recommendation suggests a location to users who might be interested [19]. The authors in [20,21] use federated learning for the indoor localization of humans, i.e., finding their positions accurately. Similar to this, Brik et al. [22] propose a method that tracks an object in an indoor environment. Autonomous control for autonomous vehicles aims to track vehicles so as to reduce traffic accidents [23]. In references [24,25], the authors try to solve collision avoidance problems by sensing the movement of many vehicles. Traffic sign recognition is another interesting topic that classifies traffic signs by using cameras and on-vehicle sensors [26,27]. Energy demand prediction aims to predict when an electrical vehicle needs to be charged by using travel information about the vehicle, such as velocity, distance, and the number of drives taken [28,29].

2.1.2. Inference Attack on Federated Learning

Although federated learning provides a privacy solution by keeping local data on users’ sides, learning parameters can still be subject to various attacks. A malicious entity can see the parameters sent by users, and an honest but curious aggregator can identify a user’s contribution when computing the change in the global model. Among the various threat models in federated learning, this work is interested in inference attacks. An attacker infers users’ sensitive information as the parameters can leak unintended features of training data [4,5,30,31]. The simplest way is to infer the model based on continuous queries and responses. Figure 2 captures a sample scenario of an inference attack on federated learning. Recent research has reported various forms of inference attacks [32,33]. In order to defend from the attack, many previous works have adopted well-known privacy-preserving techniques: homomorphic encryption, secure multiparty computation, and differential privacy. Among them, this work explores the last one.

2.2. Differential Privacy

Differential privacy (DP) ensures a similar distribution of responses to queries, regardless of whether an individual’s record exists in a database [7]. Because an attacker cannot distinguish responses, DP makes it impossible to infer more information about a particular individual. In this way, it protects personal data while ensuring the usefulness of data utilization. DP is defined below [34].
Definition 1.
(ε, δ): differential privacy
If databases D and D with a record difference for any algorithm M are established as follows, then (ε, δ)—differential privacy—is satisfied.
ε > 0 , δ 0 , P r [ M ( D ) = t ] e ε P r [ M ( D ) = t ] + δ .
If δ is 0, then δ-difference privacy is satisfied. The smaller the δ value, the smaller the distribution difference between the two databases, thus ensuring greater privacy. In general, the value of δ is less than 1 / | D | .
In the literature, differential privacy has been applied to federated learning scenarios. The authors in [35,36] investigate a federated learning setup where a coordinator is assumed to be trusted and is responsible for adding noise to aggregate local parameters in order to ensure the privacy of all of the data from users. In another setting, the assumption of a trusted coordinator is eliminated, and data owners are responsible for perturbing private information locally before uploading it to the untrusted coordinator [37,38,39]. With an untrusted coordinator, the research works in [40,41] examine an additive noise mechanism where local parameters are perturbed with less noise but then encrypted to ensure privacy before forwarding to the coordinator.

3. Internet of Vehicles

This section begins with a brief introduction to the terminologies related to the Internet of Vehicles. Connected vehicles refer to vehicles that can communicate with their external environments via wireless communication technologies [42]. With the connectivity property, they are also represented as vehicle-to-everything (V2X); here, “everything” often includes neighboring vehicles, road infrastructure, pedestrians, satellites, and so on. Vehicular networking, as evolved from vehicular ad hoc networking, represents networking perspectives among connected objects on top of the connectivity. Connected vehicles and vehicular networking can be considered as building blocks of the Internet of Vehicles (IoV), which represents a mobile and distributed computing environment. It is characterized by collecting, sharing, and processing information in a collaborative manner between connected vehicles [1]. These technologies together are expected to enable the evolution to the next-generation road infrastructure: cooperative intelligent transportation systems (C-ITSs). While the terms are somewhat different, this work uses them interchangeably as it is more concerned with how to apply federated learning and differential privacy to them.
Two of the most distinguishing features of IoV are the mobility and time–space validity of data. First, vehicles move at a very high speed along roads. In addition, the speed can change in a wide range, for example, when a vehicle traveling at 60 km/h stops in front of a crosswalk. The high speed and varying mobility make the connectivity unstable, which eventually adversely affects the operations of federated learning. Next, data generated in IoV tend to have locality in terms of space and time. Vehicle-generated content has its own spatial scope of utility and temporal scope of validity. For instance, road safety messages are only interesting to vehicles in the vicinity and may only be valid for 30 min, even in the same location. The time–space validity of the data implies the scalability of data processing applications like federated learning, since old data are obsolete.

3.1. Federated Learning on Internet of Vehicles

While the previous section lists studies on federated learning in vehicle scenarios, most of them simply utilize the scenarios as applications. Few address systematic perspectives of IoV. There remain many issues that need to be addressed before applying federated learning to real-world vehicles [9,43]. We note that many of these issues are attributed to the mobility of vehicles, and the following are representative examples: (1) Since vehicles tend to only move on roads, training data between participating vehicles may be biased if model training is not considered. This can cause overfitting problems. (2) Due to the high mobility of vehicles, the characteristics of training data are extremely dynamic, and the amount of data also changes quickly. (3) When a coordinator selects participating vehicles randomly, it is possible that vehicles with very important information (e.g., those witnessing an accident) may not be selected as training nodes. (4) Because each vehicle has different motion characteristics and data distributions, the input data during training have non-independent and identically distributed (non-IID) characteristics. In order to deal with the problems, Deveaux et al. [43,44] investigate an interesting area: vehicular knowledge networking. They first examine the meaning and scope of knowledge on vehicular networks and propose a networking framework needed for describing, generating, storing, and sharing knowledge among vehicles. Using the framework, they show federated learning orchestration for a vehicular environment; a coordinator selects vehicles carefully by making use of vehicular knowledge to improve learning performance.

3.2. Differential Privacy on Internet of Vehicles

In the literature on IoV, differential privacy has often been applied for the protection of trajectory data and location privacy. Zhang et al. [45] develop a geographically indistinguishable method that leverages differential privacy to obscure a vehicle’s location within a specific range. Qian et al. [46] propose a location privacy-preserving method in vehicle-based crowdsensing networks. Each vehicle obfuscates its own location data using differential privacy and then submits them to an application platform. Ma et al. [47] adopt differential privacy when releasing a vehicle’s trajectory data. It uses a dynamic sampling method to process the data and a privacy budget allocation method for location protection. Cai et al. [48] propose a method of releasing a trajectory using a differential privacy dataset in IoV. The method constructs a noisy prefix tree based on the spatiotemporal characteristics of the trajectory data and utilizes a Markov chain to minimize the cost of adding noise. Li et al. [49] present a quadtree-based spatial decomposition algorithm to protect the location privacy of electric vehicles in a vehicle-to-grid network. The algorithm leverages a quadtree and Bernoulli random sampling to enhance relative error with respect to the privacy budget. Sun et al. [50] propose a method of synthesizing private and realistic trajectory data by using a geography-aware grid that integrates the public geography structures of a target area.
In another side of the literature, differential privacy has been used in IoV for general purposes. Zhao et al. [38] investigate the integration of federated learning with differential privacy to enable crowdsourcing applications for both privacy and network efficiency. Batool et al. [51] propose a two-layer framework for private data sharing in vehicular networks. Each vehicle adds noise to local raw data and then forwards them to nearby a roadside unit that is responsible for training a local model with the perturbed data. The unit then uploads the trained parameters to a coordinating server. Olowononi et al. [52] apply both federated learning and differential privacy to connected vehicles to make vehicular cyber–physical systems resilient to adversarial attacks. To enhance resiliency, they use a layer-wise relevance propagation method that can regulate the selection of perturbation values. Khaliq et al. [53] investigate a privacy-preserved parking recommender system. They propose a mutual authentication mechanism using elliptic curve cryptography and make use of differential privacy to perturb data when sharing users’ historical parking data with third-party recommender systems. Sani et al. [54] summarize privacy-preserving techniques including differential privacy applied to electric vehicles.
There is a very short list of research that applied both federated learning and differential privacy to the Internet of Vehicles. To the best of the authors’ knowledge, few prior studies have applied them in an environment considering mobility patterns and evaluated the impact of mobility on their performance, which are the final goals of this work.

4. Proposed Method

4.1. Mobility-Based Federated Learning

With respect to federated learning, this work considers two models: a conventional model and a mobility-based model. In particular, the vehicular knowledge networking introduced in the previous section efficiently identifies and resolves federating issues in the processes of client selection and communication, such as bias in training data, fairness, and non-IID arising from two features of IoV (vehicle mobility and the time–space validity of data) [3]. In this sense, this work takes it as our mobility-based federated learning model. Figure 3 illustrates the model’s architecture and operations. Each vehicle periodically collects its own semantic information (i.e., knowledge) that includes its location, training environment, and kinematic information (step 1 in the figure). In the process of participant selection, vehicles share their own knowledge with a coordinator, who then identifies which vehicles have the necessary data for model training and invites appropriate ones to join.

4.2. Mobility-Enhanced Differential Privacy

This section presents the design of a novel differential privacy algorithm that takes the features of IoV into account, and as a result, it works seamlessly on the mobility-based federated learning model. Technically, the algorithm uses local differential privacy to handle mobility and an adaptive clipping method to deal with the locality of IoV data that are related to fairness and non-IID. Local differential privacy is a method of applying differential privacy when one does not trust a server. Differential privacy is executed at the user level, and obfuscated data are then submitted to the server. In terms of federated learning, noise is added to the updated parameter values of each vehicle, and the values are transmitted to a coordinator. In this research, user-level (local) differential privacy is used to ensure stronger privacy. To this end, a proposed algorithm updates a combined averaging algorithm running on the coordinator. A general method of ensuring local differential privacy in federated learning is to limit the contribution of each training participant’s model update to a certain value and then add noise. In this case, adaptive clipping uses a quartile value of the updated norm distribution instead of a fixed norm value.
By applying these and referencing previous works [35,55,56,57,58,59,60,61], our algorithm for differentially private federated learning on the Internet of Vehicles is shown in Algorithm 1. We note that symbols are borrowed from conventional local differential privacy equations. In the algorithm, let m be the number of participating vehicles and γ [ 0 , 1 ] be the target quantile of the Gaussian distribution at which we want to clip. For each round t, Q t denotes a set of vehicles sampled, C t is a clipping threshold, and  η C is a learning rate. Each vehicle i sends a usual delta update Δ i t and additionally a bit b i t that is necessary for adjusting the clipping threshold, where b i t = I | | Δ i t | | 2 C t . When denoting an average value, b ¯ t = 1 m i Q t b i t , our algorithm follows a geometric update rule: C C · e x p ( η C ( b ¯ γ ) )  [57]. To perturb sensitive information b ¯ t , Gaussian noise is added to the sum, obtaining b ˜ t = 1 m ( i Q t b i t + N ( O , σ b 2 ) ) . The training procedure initializes two parameters: it sets a model to θ 0 and a clipping threshold to C 0 . It also takes two noise parameters: a noise standard deviation on clipped counts b i t ( σ b ) and a noise multiplier on the vector sum Δ i t ( z Δ ), where z Δ = ( z 2 ( 2 σ b ) 2 ) 1 2 . The combined averaging algorithm on the coordinator first limits the updated parameter values of each vehicle to not exceed the clipping critical value before integrating the values. Then, it adds Gaussian noise to the sum, which is used to update the global parameters and clipping values. The major updates are summarized below:
Algorithm 1 Mobility-based differentially private federated learning with adaptive clipping for Internet of Vehicles
 1:
procedure Training(m, γ , η c , η s , η C , z Δ , σ b , β )
 2:
    for each round t = 0, 1, 2,... do
 3:
         Q t a set of vehicles sampled in round t
 4:
        for each vehicle i Q t  do
 5:
           ( Δ i t , b i t ) ← FedAvg(i, θ t , η c , C t )
 6:
        end for
 7:
         σ Δ z Δ C t
 8:
         Δ ˜ t = 1 m ( i Q t Δ i t + N ( 0 , I σ Δ 2 ) )
 9:
         Δ ¯ t = β Δ ¯ t 1 + ( 1 β ) Δ ˜ t
10:
         θ t + 1 θ t + η s Δ ¯ t
11:
         b ˜ t = 1 m ( i Q t b i t + N ( O , σ b 2 ) )
12:
         C t + 1 C t · e x p ( η C ( b ˜ t γ ) )
13:
    end for
14:
end procedure
 
15:
procedure FedAvg(i, θ 0 , η , C)
16:
     θ θ 0
17:
     S vehicle i’s local data split into batches
18:
    for batch s S  do
19:
         θ θ η L ( θ ; s )
20:
    end for
21:
     Δ θ θ 0
22:
     b I | | Δ | | C
23:
     Δ Δ · m i n ( 1 , C | | Δ | | )
24:
    return  ( Δ , b )
25:
end procedure
  • The model update values of vehicles participating in the training are clipped before transmission to the federated learning coordinator to limit the individual effects of each vehicle.
  • To obscure the effects of individual vehicles, it adds sufficient noise values to the total values updated before using the combined averaging algorithm.
  • To determine the value to be clipped, adaptive clipping is applied and Gaussian is used for noise.

4.3. Mobility Pattern

This work considers two mobility patterns: a random waypoint (RWP) model [62] and a reference point group mobility (RPGM) model [63]. In RWP, mobile nodes (vehicles) move freely and randomly without restrictions. That is, the direction, speed, and destination are all chosen randomly, independent of other nodes. Its simplicity and wide availability make it one of the most popular mobility models. Its movement includes pause times between changes in destination and speed. A mobile node begins by staying in a location for a fixed number of seconds (pause time). Upon timeout, the node chooses a random destination in a simulation area and a random speed between min_speed and max_speed, where min_speed is not zero. It then travels to the destination at this speed. Upon arrival, the node stays again for a fixed period before choosing another random destination and speed. It repeats this process until the end of the simulation. This work adopts the RWP model to simulate events that do not occur frequently but are important for training, such as traffic accidents [64]. In the model, a rare-event generation mechanism is used for sensing training data input. With respect to differential privacy, a noise multiplier is applied when updating model parameter values for adaptive clipping.
In RPGM, mobile nodes are divided into groups, and group members move together. Each group has a logical center that travels through a random path following the RWP model. The motion of the center characterizes the movement of member nodes, including direction and speed. Each member node randomly moves around its own predefined reference point that is in the vicinity of the center and whose location is continuously updated according to the center’s mobility. The node’s random motion within the group also follows the RWP model. This way, the node follows the center’s mobility closely, with some deviation. Figure 4 illustrates the movement of three mobile nodes with the RPGM model. At time t, three black dots represent reference points, R P ( t ) , for the three mobile nodes (denoted as big white circles). The motion of the logical center is represented by a group motion vector, G M ¯ , that is used to calculate new locations of the reference points, R P ( t + 1 ) , at time t + 1 . Then, the new position of a node is calculated by summing the new reference point with a random motion vector, R M ¯ . The direction of R M ¯ is uniformly distributed between 0 and 2 π , and its length is also uniformly distributed within a predefined radius centered at R P ( t + 1 ) . The RPGM model is one of the most popular group mobility models to represent vehicle platooning, and thus it is adopted in this work to represent vehicles’ mobility in non-accident situations [65]. The model uses a grid-based observation mechanism as a training input data sensing method in a federated learning setting and an adaptive clipping method for differential privacy.

5. Experimental Setup

The goal of the experiments and evaluation is to assess the accuracy and privacy performance of the proposed, differentially private, mobility-based federated learning model. For the experiments, we use digit images in the Federated Extended MNIST (FEMNIST) dataset that partitions the EMNIST dataset based on writers [66]. As such, the data replicate non-IID data distribution for federated learning. The mobility-based federated learning model (named mobile FL in the resulting graphs) is implemented by using the tensorflow-federated and tensorflow-privacy libraries. For performance comparison, we also implement a conventional federated learning model (named vanilla FL), where a coordinator selects participants randomly.
Experiments are run with 750 vehicles in total deployed in a one-square-kilometer area for one hour. Each vehicle is a candidate for a participant who executes a training process and is assumed to be connected to the coordinator at all times. The coordinator selects participating vehicles for each learning; it takes different selection strategies in different learning models and mobility patterns. Experiments are run in two different scenarios that differ in mobility models, mechanisms for data sensing, and differential privacy algorithms. In the first scenario with the RPGM model, each vehicle periodically detects its location and training data every 10 s for the mobile FL model. In the vanilla FL model, a coordinator clusters vehicles into groups based on their location and selects vehicles (as training nodes) from each group that detect the most recent training data. The second scenario with the RWP model sets a speed range of 5–20 m/s, a pause probability of 0.5, and a maximum pause duration of 5 min. Events are randomly generated following a Poisson distribution. An event can occur anywhere inside the simulation area, and once it occurs, it stays active for a limited amount of time. An active event is represented as a small area. Once a vehicle moves into the event area, it detects and records data samples related to the event. To account for the space–time locality, stored data are removed after a certain period of time after recording or when the vehicle drives away from the event area.
Privacy performance is evaluated by a noise value in differential privacy. For experiments, the noise sensitivity of the training models (i.e., the mobile FL model and the vanilla FL model) must be determined by adaptive clipping. To this end, we conduct preliminary experiments after each model is multiplied by noise multiplier (z) values of 0, 0.5, and 0.75. The results identify a noise tolerance level that does not harm the accuracy performance critically at each model with a relatively small number of vehicles (10 vehicles). They can also provide the number of vehicles and noise multiplier values, which is necessary to ensure the accuracy of a model that an application wants, given target δ and ε values. Next, the experiments run final trainings with these numbers. We note that a noise multiplier value of 0 represents that differential privacy is not applied to a training model, whereas values of 0.5 or 0.75 indicate privacy-enhanced training models in this work.

6. Performance Evaluation

This section presents the experimental results in terms of accuracy and loss performance (our preliminary data were presented as a poster in [67]). We show results in the scenario with the RPGM model first, as it represents vehicles mobility in non-accident situations. Then, the experimental results in the case of events (e.g., car accidents) are given.

6.1. Scenario 1. Reference Point Group Mobility

Figure 5 and Figure 6 illustrate performance in terms of model accuracy and training loss in vanilla FL and mobile FL, respectively, in the RPGM mobility scenario. The figures also show the impacts of different noise levels applied to them for privacy. The graphs clearly demonstrate the trade-off between accuracy and privacy. We initially set the performance of models with the noise multiplier of 0.5 as the target performance. The ( δ , ε ) values are set to (2, 1 × 105), which are widely accepted. Then, each model adjusts both the number of vehicles participating in learning per round and the noise value until it achieves the target performance for privacy. Our experiments find that 36 participating vehicles and a noise multiplier value of 1.6 (while keeping the training steps at 200) ensure the required level of privacy and accuracy performance for the models.
Table 1 enumerates and compares the accuracy and loss values of the four training models (vanilla FL, mobile FL, and their two privacy-enhanced models). Mobile FL shows the best performance. The privacy-enhanced mobile FL model performs with 80.28% accuracy, which is degraded by 3.11% when compared to mobile FL. It is interesting to observe that it performs better than vanilla FL; that is, it is able to protect private data and predict results more correctly than conventional FL models. The accuracy performance of the two privacy-enhanced FL models degrades by 3.26% on average compared to non-privacy FL models, but this loss is fully compensated for by their capabilities in protecting vehicles against inference attacks.

6.2. Scenario 2. Random Waypoint

Figure 7 and Figure 8 illustrate performance in terms of model accuracy and training loss in vanilla FL and mobile FL, respectively, in the RWP mobility scenario. The figures also show the impacts of different noise levels applied to them for privacy. In the graphs, a noticeable observation is that performance in vanilla FL drops more significantly than that in mobile FL as the noise multiplier value increases. In the latter, a coordinator makes use of its knowledge base maintaining vehicles’ semantic information and selects vehicles that have more data needed for learning as participants before training. This method is particularly important in this scenario because a small number of vehicles may store data associated with accidents, and most of them are highly likely to be selected as participants, which thus prevents dramatic performance degradation.
Table 2 summarizes the experimental results. Just like in scenario 1, mobile FL performs best, and the privacy-enhanced mobile FL model shows a 2.96% lower accuracy performance than the former. The latter also performs better than vanilla FL by 9.16%, which is a much larger gap than that in scenario 1. The worst performance is from the privacy-enhanced vanilla FL model. Vehicles having no knowledge about accidents are equally likely to be selected as participants, and their update values are noised, which results in a sharp drop in performance, from 73.17% to 41.73%.

6.3. Membership Inference Attack

The last experiment sets up a membership inference attack [6] on the first mobility scenario that can ensure the desired level of privacy preservation by applying the adaptive clipping method. Table 3 shows the attack success rates (ASRs) when the attack is launched on the four training models. The closer the value is to 0.5, the less able the attacker is to distinguish between training data. As shown, the two privacy-enhanced models degrade the rate by around 10%, meaning that information is more protected from the attack. However, since the value itself is greater than 0.5, applying differential privacy alone may not be enough to resolve the attack completely. Using additional technologies such as weight pruning [68] could mitigate its effect, which remains an area that will be addressed in one of our future works.

7. Discussion

Communication costs can be a very important constraint in federated learning, as servers and participants need to communicate model parameters frequently. Existing research has taken three strategies to reduce communication overhead: an optimization algorithm, model compression, and client selection. A rational strategy is to reduce communication frequency and increase the time interval of model aggregation, which can be achieved by improving the performance of local training [9]. The authors also found that increasing local iteration time and reducing global communication rounds can increase a global model’s convergence efficiency while ensuring the performance of federated learning. Compressing network traffic by reducing the overall size of the uploaded model parameters is another useful strategy. Recent works [69,70] investigated Top-k sparsification-based compression techniques in which parameters with the largest variations (i.e., k “important” parameters) are selected and transmitted to a server, instead of uploading an entire local model. The authors in [71,72] employed quantization and encoding schemes to further reduce the sizes of the “value” and “index” parts in the k parameters. Furthermore, client selection methods are logical strategies for improving communication efficiency in federated learning. Liu et al. [73] studied a method that detects whether local model updates from clients match a global model’s update trend to avoid sending irrelevant parameters. Deng et al. [74] investigated a method to evaluate clients’ model quality to select high-quality clients for federated learning. Lai et al. [75] proposed a framework to prioritize client participation using data that provide the greatest utility in improving model accuracy and training speed. Nishio and Yonetani [76] designed a new protocol that predefines deadlines for updating and uploading model parameters so that a server can aggregate them in a given time. The mobility-based federated learning in this work requires vehicles to share their own knowledge with a coordinating server, which adds additional communication cost. However, the server preferentially selects vehicles that have important data for model training, which enables satisfactory accuracy to be achieved in fewer communication rounds. In particular, the selection mechanism in this work is similar to that in [74], demonstrating that their framework can improve accuracy performance by 32–113% after 30 learning rounds and can converge to the global optimum with fewer participants.
The convergence of federated learning has been analyzed in many studies. Several works examined the effects of data heterogeneity on convergence by assuming full client participation [55,77,78]. However, not all clients actually participate in each training round, and thus the authors in [79,80,81] analyzed federated learning with unbiased partial client participation where clients are selected uniformly at random. These works show that the federated averaging (FedAvg) algorithm on non-IID datasets achieves a convergence rate of O (1/T), where T is the number of communication rounds. Since the mobility-based federated learning in this work, especially in non-accident situations, takes unbiased partial client participation, its error convergence is expected to follow the same rate as shown in our experimental results in Section 6.1. In the case of car accident events, on the other hand, our learning model takes a biased client selection strategy, and recent empirical studies have shown that biased selection affects convergence speed. Nishio and Yonetani [76] proposed a client selection method based on the client’s radio resources and hardware to reduce communication costs. The authors in [82,83] discovered that selecting clients with higher local losses can sharply increase the rate of convergence. Our experimental results in Section 6.2 do not deviate from the observation that biased selection affects convergence. An error in vanilla FL almost fails to converge, whereas the curves in mobile FL look stabilized and show a similar convergence rate to that in Section 6.1.

8. Conclusions

This work has presented performance issues in differentially private federated learning models running in different mobility scenarios. The accuracy performance and level of protection have been evaluated by considering three factors: learning models, vehicle mobility, and a privacy algorithm. The accuracy of the privacy-enhanced learning models was mostly affected by vehicles’ mobility and the traffic situations on roads. It was also observed that the conventional learning model widely used in federated learning did not work well in the mobile, vehicular environment, especially when enhanced with a privacy method. This is mainly attributed to the performance of the participant selection mechanism, which is highly affected by vehicles’ mobility. An additional observation is that the adaptive clipping method applied in this work is more useful than prior methods because it guarantees model performance and the target degree of privacy preservation and thus can proceed with model learning.
This study focused on privacy and the associated trade-offs in mobility-based federated learning. We plan to consider a weight pruning method, which is a recently proposed alternative to membership inference attacks, to enable more efficient and accurate learning while also showing improved security results. Future works will also examine how well privacy-enhanced learning models protect data against variants of inference attacks and expand on experiments by considering more mobility scenarios.

Author Contributions

Conceptualization, E.-j.K. and E.-K.L.; Methodology, E.-j.K. and E.-K.L.; Software, E.-j.K.; Validation, E.-j.K.; Investigation, E.-K.L.; Resources, E.-K.L.; Data curation, E.-j.K.; Writing—original draft, E.-j.K. and E.-K.L.; Writing—review & editing, E.-K.L.; Visualization, E.-j.K.; Supervision, E.-K.L.; Project administration, E.-K.L.; Funding acquisition, E.-K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2024-RS-2023-00259061) supervised by the IITP (Institute for Information & Communications Technology Planning & Evaluation).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gerla, M.; Lee, E.K.; Pau, G.; Lee, U. Internet of vehicles: From intelligent grid to autonomous cars and vehicular clouds. In Proceedings of the IEEE World Forum on Internet of Things, Seoul, Republic of Korea, 6–8 March 2014. [Google Scholar]
  2. Chellapandi, V.P.; Yuan, L.; Brinton, C.G.; Żak, S.H.; Wang, Z. Federated Learning for Connected and Automated Vehicles: A Survey of Existing Approaches and Challenges. IEEE Trans. Intell. Veh. 2024, 9, 119–137. [Google Scholar] [CrossRef]
  3. Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings, R.; et al. Advances and Open Problems in Federated Learning. Found. Trends Mach. Learn. 2021, 14, 1–210. [Google Scholar] [CrossRef]
  4. Melis, L.; Song, C.; Cristofaro, E.D.; Shmatikov, V. Exploiting Unintended Feature Leakage in Collaborative Learning. In Proceedings of the IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 20–22 May 2019; pp. 691–706. [Google Scholar]
  5. Carlini, N.; Liu, C.; Erlingsson, Ú.; Kos, J.; Song, D. The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks. In Proceedings of the USENIX Security Symposium, Santa Clara, CA, USA, 14–16 August 2019; pp. 267–284. [Google Scholar]
  6. Shokri, R.; Stronati, M.; Song, C.; Shmatikov, V. Membership Inference Attacks against Machine Learning Models. In Proceedings of the IEEE Symposium on Security and Privacy, San Jose, CA, USA, 22–24 May 2017; pp. 3–18. [Google Scholar]
  7. Dwork, C. Differential Privacy: A Survey of Results. In Proceedings of the Theory and Applications of Models of Computation, Xian, China, 25–29 April 2008; pp. 1–19. [Google Scholar]
  8. Li, G.; Zhu, P.; Li, J.; Yang, Z.; Cao, N.; Chen, Z. Security Matters: A Survey on Adversarial Machine Learning. arXiv 2018, arXiv:1810.07339. [Google Scholar]
  9. McMahan, H.B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Ft. Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  10. Cho, H.; Mathur, A.; Kawsar, F. Device or User: Rethinking Federated Learning in Personal-Scale Multi-Device Environments. In Proceedings of the ACM Conference on Embedded Networked Sensor Systems, Coimbra, Portugal, 15–17 November 2021; pp. 446–452. [Google Scholar]
  11. Ouyang, X.; Xie, Z.; Zhou, J.; Xing, G.; Huang, J. ClusterFL: A Clustering-based Federated Learning System for Human Activity Recognition. ACM Trans. Sens. Netw. 2022, 19, 17. [Google Scholar] [CrossRef]
  12. Gong, Q.; Ruan, H.; Chen, Y.; Su, X. CloudyFL: A cloudlet-based federated learning framework for sensing user behavior using wearable devices. In Proceedings of the ACM International Workshop on Embedded and Mobile Deep Learning, Portland, OR, USA, 1 July 2022; pp. 13–18. [Google Scholar]
  13. Bu, C.; Zhang, L.; Cui, H.; Cheng, D.; Wu, H.; Song, A. Learn From Others and Be Yourself in Federated Human Activity Recognition via Attention-Based Pairwise Collaborations. IEEE Trans. Instrum. Meas. 2024, 73, 2506415. [Google Scholar] [CrossRef]
  14. Xu, S.; Zhang, L.; Tang, Y.; Han, C.; Wu, H.; Song, A. Channel Attention for Sensor-Based Activity Recognition: Embedding Features into all Frequencies in DCT Domain. IEEE Trans. Knowl. Data Eng. 2023, 35, 12497–12512. [Google Scholar] [CrossRef]
  15. Cheng, D.; Zhang, L.; Bu, C.; Wang, X.; Wu, H.; Song, A. ProtoHAR: Prototype Guided Personalized Federated Learning for Human Activity Recognition. IEEE J. Biomed. Health Inform. 2023, 27, 3900–3911. [Google Scholar] [CrossRef]
  16. Feng, J.; Rong, C.; Sun, F.; Guo, D.; Li, Y. PMF: A Privacy-preserving Human Mobility Prediction Framework via Federated Learning. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 1–21. [Google Scholar] [CrossRef]
  17. Fan, Z.; Song, X.; Jiang, R.; Chen, Q.; Shibasaki, R. Decentralized Attention-based Personalized Human Mobility Prediction. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 1–26. [Google Scholar] [CrossRef]
  18. Liu, Y.; Yu, J.J.Q.; Kang, J.; Niyato, D.; Zhang, S. Privacy-Preserving Traffic Flow Prediction: A Federated Learning Approach. IEEE Internet Things J. 2020, 7, 7751–7763. [Google Scholar] [CrossRef]
  19. Guo, Y.; Liu, F.; Cai, Z.; Zheng, H.; Chen, L.; Zhou, T.; Xiao, N. PREFER: Point-of-interest REcommendation with efficiency and privacy-preservation via Federated Edge leaRning. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2021, 5, 1–25. [Google Scholar] [CrossRef]
  20. Li, W.; Zhang, C.; Tanaka, Y. Pseudo Label-Driven Federated Learning-Based Decentralized Indoor Localization via Mobile Crowdsourcing. IEEE Sens. J. 2020, 20, 11556–11565. [Google Scholar] [CrossRef]
  21. Wu, Z.; Wu, X.; Long, Y. Prediction Based Semi-Supervised Online Personalized Federated Learning for Indoor Localization. IEEE Sens. J. 2022, 22, 10640–10654. [Google Scholar] [CrossRef]
  22. Brik, B.; Messaadia, M.; Sahnoun, M.; Bettayeb, B.; Benatia, M.A. Fog-supported Low-latency Monitoring of System Disruptions in Industry 4.0: A Federated Learning Approach. ACM Trans. Cyber-Phys. Syst. 2022, 6, 14. [Google Scholar] [CrossRef]
  23. Zeng, T.; Semiari, O.; Chen, M.; Saad, W.; Bennis, M. Federated Learning for Collaborative Controller Design of Connected and Autonomous Vehicles. In Proceedings of the IEEE Conference on Decision and Control, Austin, TX, USA, 13–17 December 2021; pp. 5033–5038. [Google Scholar]
  24. Chhikara, P.; Tekchandani, R.; Kumar, N.; Tanwar, S. Federated learning-based aerial image segmentation for collision-free movement and landing. In Proceedings of the ACM MobiCom Workshop on Drone Assisted Wireless Communications for 5G and Beyond, Virtual Event, 29 October 2021; pp. 13–18. [Google Scholar]
  25. Hammedi, W.; Brik, B.; Senouci, S.M. Toward Optimal MEC-Based Collision Avoidance System for Cooperative. IEEE Trans. Intell. Transp. Syst. 2023, 24, 2525–2537. [Google Scholar]
  26. Nuding, F.; Mayer, R. Data Poisoning in Sequential and Parallel Federated Learning. In Proceedings of the ACM International Workshop on Security and Privacy Analytics, Baltimore, MD, USA, 25–27 April 2022; pp. 24–34. [Google Scholar]
  27. Ye, D.; Yu, R.; Pan, M.; Han, Z. Federated Learning in Vehicular Edge Computing: A Selective Model Aggregation Approach. IEEE Access 2020, 8, 23920–23935. [Google Scholar] [CrossRef]
  28. Saputra, Y.M.; Hoang, D.T.; Nguyen, D.N.; Dutkiewicz, E.; Mueck, M.D.; Srikanteswara, S. Energy Demand Prediction with Federated Learning for Electric Vehicle Networks. In Proceedings of the IEEE Global Communications Conference, Waikoloa, HI, USA, 9–13 December 2019. [Google Scholar]
  29. Thorgeirsson, A.T.; Scheubner, S.; Fünfgeld, S.; Gauterin, F. Probabilistic Prediction of Energy Demand and Driving Range for Electric Vehicles With Federated Learning. IEEE Open J. Veh. Technol. 2021, 2, 151–161. [Google Scholar] [CrossRef]
  30. Zhu, L.; Liu, Z.; Han, S. Deep Leakage from Gradients. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32, pp. 14774–14784. [Google Scholar]
  31. Phong, L.T.; Aono, Y.; Hayashi, T.; Wang, L.; Moriai, S. Privacy-Preserving Deep Learning via Additively Homomorphic Encryption. IEEE Trans. Inf. Forensics Secur. 2018, 13, 1333–1345. [Google Scholar] [CrossRef]
  32. Nasr, M.; Shokri, R.; Houmansadr, A. Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning. In Proceedings of the IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 20–22 May 2019; pp. 739–753. [Google Scholar]
  33. Lyu, L.; Yu, H.; Ma, X.; Chen, C.; Sun, L.; Zhao, J.; Yang, Q.; Yu, P.S. Privacy and Robustness in Federated Learning: Attacks and Defenses. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–21. [Google Scholar] [CrossRef]
  34. Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep Learning with Differential Privacy. In Proceedings of the ACM SIGSAC Conference on Computer and Communications Security, Vienna, Austria, 24–28 October 2016; pp. 308–318. [Google Scholar]
  35. McMahan, H.B.; Ramage, D.; Talwar, K.; Zhang, L. Learning Differentially Private Recurrent Language Models. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–14. [Google Scholar]
  36. Geyer, R.C.; Klein, T.J.; Nabi, M. Differentially Private Federated Learning: A Client Level Perspective. arXiv 2017, arXiv:1712.07557. [Google Scholar]
  37. Bhowmick, A.; Duchi, J.; Freudiger, J.; Kapoor, G.; Rogers, R. Protection Against Reconstruction and Its Applications in Private Federated Learning. arXiv 2018, arXiv:1812.00984. [Google Scholar]
  38. Zhao, Y.; Zhao, J.; Yang, M.; Wang, T.; Wang, N.; Lyu, L.; Niyato, D.; Lam, K.Y. Local Differential Privacy based Federated Learning for Internet of Things. IEEE Internet Things J. 2021, 8, 8836–8853. [Google Scholar] [CrossRef]
  39. Sun, L.; Qian, J.; Chen, X. LDP-FL: Practical Private Aggregation in Federated Learning with Local Differential Privacy. In Proceedings of the International Joint Conference on Artificial Intelligence, Montreal, QC, Canada, 19–26 August 2021; pp. 1571–1578. [Google Scholar]
  40. Agarwal, N.; Suresh, A.T.; Yu, F.; Kumar, S.; Mcmahan, H.B. cpSGD: Communication-efficient and differentially-private distributed SGD. In Proceedings of the International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 2–8 December 2018; pp. 7564–7575. [Google Scholar]
  41. Truex, S.; Baracaldo, N.; Anwar, A.; Steinke, T.; Ludwig, H.; Zhang, R.; Zhou, Y. A Hybrid Approach to Privacy-Preserving Federated Learning. In Proceedings of the ACM Workshop on Artificial Intelligence and Security, London, UK, 15 November 2019; pp. 1–11. [Google Scholar]
  42. Lu, N.; Cheng, N.; Zhang, N.; Shen, X.; Mark, J.W. Connected Vehicles: Solutions and Challenges. IEEE Internet Things J. 2014, 1, 289–299. [Google Scholar] [CrossRef]
  43. Deveaux, D.; Higuchi, T.; Uçar, S.; Wang, C.H.; Härri, J.; Altintas, O. On the Orchestration of Federated Learning Through Vehicular Knowledge Networking. In Proceedings of the IEEE Vehicular Networkng Conference, Virtual Conference, 16–18 December 2020; pp. 1–8. [Google Scholar]
  44. Deveaux, D.; Higuchi, T.; Uçar, S.; Härri, J.; Altintas, O. A Definition and Framework for Vehicular Knowledge Networking: An Application of Knowledge-Centric Networking. IEEE Veh. Technol. Mag. 2021, 16, 57–67. [Google Scholar] [CrossRef]
  45. Zhang, X.; Wang, J.; Zhang, H.; Li, L.; Pan, M.; Han, Z. Data-Driven Transportation Network Company Vehicle Scheduling With Users’ Location Differential Privacy Preservation. IEEE Trans. Mob. Comput. 2023, 22, 813–823. [Google Scholar] [CrossRef]
  46. Qian, Y.; Ma, Y.; Chen, J.; Wu, D.; Tian, D.; Hwang, K. Optimal Location Privacy Preserving and Service Quality Guaranteed Task Allocation in Vehicle-Based Crowdsensing Networks. IEEE Trans. Intell. Transp. Syst. 2021, 22, 4367–4375. [Google Scholar] [CrossRef]
  47. Ma, Z.; Zhang, T.; Liu, X.; Li, X.; Ren, K. Real-Time Privacy-Preserving Data Release Over Vehicle Trajectory. IEEE Trans. Veh. Technol. 2019, 68, 8091–8102. [Google Scholar] [CrossRef]
  48. Cai, S.; Lyu, X.; Li, X.; Ban, D.; Zeng, T. A Trajectory Released Scheme for the Internet of Vehicles Based on Differential Privacy. IEEE Trans. Intell. Transp. Syst. 2022, 23, 16534–16547. [Google Scholar] [CrossRef]
  49. Li, Y.; Zhang, P.; Wang, Y. The Location Privacy Protection of Electric Vehicles with Differential Privacy in V2G Networks. Energies 2018, 11, 2625. [Google Scholar] [CrossRef]
  50. Sun, X.; Ye, Q.; Hu, H.; Wang, Y.; Huang, K.; Wo, T.; Xu, J. Synthesizing Realistic Trajectory Data With Differential Privacy. IEEE Trans. Intell. Transp. Syst. 2023, 24, 5502–5515. [Google Scholar] [CrossRef]
  51. Batool, H.; Anjum, A.; Khan, A.; Izzo, S.; Mazzocca, C.; Jeon, G. A secure and privacy preserved infrastructure for VANETs based on federated learning with local differential privacy. Inf. Sci. 2024, 652, 119717. [Google Scholar] [CrossRef]
  52. Olowononi, F.O.; Rawat, D.B.; Liu, C. Federated Learning with Differential Privacy for Resilient Vehicular Cyber Physical Systems. In Proceedings of the IEEE Annual Consumer Communications and Networking Conference, Virtual Conference, 9–12 January 2021; pp. 1–5. [Google Scholar]
  53. Khaliq, A.A.; Anjum, A.; Ajmal, A.B.; Webber, J.L.; Mehbodniya, A.; Khan, S. A Secure and Privacy Preserved Parking Recommender System Using Elliptic Curve Cryptography and Local Differential Privacy. IEEE Access 2022, 10, 56410–56426. [Google Scholar] [CrossRef]
  54. Sani, A.R.; Hassan, M.U.; Chen, J. Privacy Preserving Machine Learning for Electric Vehicles: A Survey. arXiv 2022, arXiv:2205.08462. [Google Scholar]
  55. Reddi, S.; Charles, Z.; Zaheer, M.; Garrett, Z.; Rush, K.; Konečný, J.; Kumar, S.; McMahan, H.B. Adaptive Federated Optimization. arXiv 2021, arXiv:2003.00295. [Google Scholar]
  56. Wang, Y.X.; Balle, B.; Kasiviswanathan, S.P. Subsampled Renyi Differential Privacy and Analytical Moments Accountant. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Okinawa, Japan, 16–18 April 2019; Proceedings of Machine Learning Research. pp. 1226–1235. [Google Scholar]
  57. Andrew, G.; Thakkar, O.; McMahan, B.; Ramaswamy, S. Differentially Private Learning with Adaptive Clipping. In Proceedings of the Advances in Neural Information Processing Systems, Virtual Conference, 6–14 December 2021; Volume 34, pp. 17455–17466. [Google Scholar]
  58. Tramèr, F.; Boneh, D. Differentially Private Learning Needs Better Features (or Much More Data). arXiv 2021, arXiv:2011.11660. [Google Scholar]
  59. Liu, K.; Hu, S.; Wu, S.Z.; Smith, V. On Privacy and Personalization in Cross-Silo Federated Learning. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022; Volume 35, pp. 5925–5940. [Google Scholar]
  60. Noble, M.; Bellet, A.; Dieuleveut, A. Differentially Private Federated Learning on Heterogeneous Data. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Virtual Conference, 28–30 March 2022; Volume 151, pp. 10110–10145. [Google Scholar]
  61. Cheng, A.; Wang, P.; Zhang, X.S.; Cheng, J. Differentially Private Federated Learning With Local Regularization and Sparsification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 10122–10131. [Google Scholar]
  62. Johnson, D.B.; Maltz, D.A. Dynamic Source Routing in Ad Hoc Wireless Networks. In Mobile Computing; Imielinski, T., Korth, H.F., Eds.; Springer: Boston, MA, USA, 1996; Volume 353, pp. 153–181. [Google Scholar]
  63. Hong, X.; Gerla, M.; Pei, G.; Chiang, C.C. A group mobility model for ad hoc wireless networks. In Proceedings of the ACM International Workshop on Modeling and Simulation of Wireless and Mobile Systems, Seattle, WA, USA, 20 August 1999; pp. 53–60. [Google Scholar]
  64. Roychowdhury, S.; Zhao, M.; Wallin, A.; Ohlsson, N.; Jonasson, M. Machine Learning Models for Road Surface and Friction Estimation using Front-Camera Images. In Proceedings of the International Joint Conference on Neural Networks, Rio, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar]
  65. Osman, O.A.; Hajij, M.; Bakhit, P.R.; Ishak, S. Prediction of Near-Crashes from Observed Vehicle Kinematics using Machine Learning. Transp. Res. Rec. 2019, 2673, 463–473. [Google Scholar] [CrossRef]
  66. Caldas, S.; Wu, P.; Li, T.; Konečný, J.; McMahan, H.B.; Smith, V.; Talwalkar, A. LEAF: A Benchmark for Federated Settings. arXiv 2018, arXiv:1812.01097. [Google Scholar]
  67. Kim, E.J.; Lee, E.K. Performance Impact of Differential Privacy on Federated Learning in Vehicular Networks. In Proceedings of the IEEE/IFIP Network Operations and Management Symposium, Budapest, Hungary, 25–29 April 2022. [Google Scholar]
  68. Wang, Y.; Wang, C.; Wang, Z.; Zhou, S.; Liu, H.; Bi, J.; Ding, C.; Rajasekaran, S. Against Membership Inference Attack: Pruning is All You Need. In Proceedings of the International Joint Conference on Artificial Intelligence, Montreal, QC, Canada, 19–26 August 2021; pp. 3141–3147. [Google Scholar]
  69. Zhang, F.; Zhai, J.; Shen, X.; Wang, D.; Chen, Z.; Mutlu, O.; Chen, W.; Du, X. TADOC: Text analytics directly on compression. VLDB J. 2020, 30, 163–188. [Google Scholar] [CrossRef]
  70. Abdi, A.; Fekri, F. Quantized Compressive Sampling of Stochastic Gradients for Efficient Communication in Distributed Deep Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 3105–3112. [Google Scholar]
  71. Wen, W.; Xu, C.; Yan, F.; Wu, C.; Wang, Y.; Chen, Y.; Li, H. TernGrad: Ternary Gradients to Reduce Communication in Distributed Deep Learning. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30, pp. 1509–1519. [Google Scholar]
  72. Sattler, F.; Wiedemann, S.; Müller, K.R.; Samek, W. Robust and Communication-Efficient Federated Learning From Non-i.i.d. Data. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 3400–3413. [Google Scholar] [CrossRef]
  73. Liu, S.; Yu, G.; Yin, R.; Yuan, J.; Shen, L.; Liu, C. Joint Model Pruning and Device Selection for Communication-Efficient Federated Edge Learning. IEEE Trans. Commun. 2022, 70, 231–244. [Google Scholar] [CrossRef]
  74. Deng, Y.; Lyu, F.; Ren, J.; Wu, H.; Zhou, Y.; Zhang, Y.; Shen, X. AUCTION: Automated and Quality-Aware Client Selection Framework for Efficient Federated Learning. IEEE Trans. Parallel Distrib. Syst. 2022, 33, 1996–2009. [Google Scholar] [CrossRef]
  75. Lai, F.; Zhu, X.; Madhyastha, H.V.; Chowdhury, M. Oort: Efficient Federated Learning via Guided Participant Selection. In Proceedings of the USENIX Symposium on Operating Systems Design and Implementation, virtual Event, 14–16 July 2021; pp. 19–35. [Google Scholar]
  76. Nishio, T.; Yonetani, R. Client Selection for Federated Learning with Heterogeneous Resources in Mobile Edge. In Proceedings of the IEEE International Conference on Communications, Shanghai, China, 20–14 May 2019; pp. 1–7. [Google Scholar]
  77. Koloskova, A.; Loizou, N.; Boreiri, S.; Jaggi, M.; Stich, S. A Unified Theory of Decentralized SGD with Changing Topology and Local Updates. In Proceedings of the International Conference on Machine Learning, Virtual Event, 12–18 July 2020; Volume 119, pp. 5381–5393. [Google Scholar]
  78. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated Optimization in Heterogeneous Networks. In Proceedings of the Machine Learning and Systems, Austin, TX, USA, 2–4 March 2020; Volume 2, pp. 429–450. [Google Scholar]
  79. Jhunjhunwala, D.; Sharma, P.; Nagarkatti, A.; Joshi, G. Fedvarp: Tackling the variance due to partial client participation in federated learning. In Proceedings of the Conference on Uncertainty in Artificial Intelligence, Eindhoven, The Netherlands, 1–5 August 2022; Volume 180, pp. 906–916. [Google Scholar]
  80. Yang, H.; Fang, M.; Liu, J. Achieving Linear Speedup with Partial Worker Participation in Non-IID Federated Learning. In Proceedings of the International Conference on Learning Representations, Virtual Conference, 3–7 May 2021. [Google Scholar]
  81. Li, X.; Huang, K.; Yang, W.; Wang, S.; Zhang, Z. On the Convergence of FedAvg on Non-IID Data. In Proceedings of the International Conference on Learning Representations, Virtual Conference, 26 April–1 May 2020. [Google Scholar]
  82. Goetz, J.; Malik, K.; Bui, D.; Moon, S.; Liu, H.; Kumar, A. Active Federated Learning. arXiv 2019, arXiv:1909.12641. [Google Scholar]
  83. Cho, Y.J.; Wang, J.; Joshi, G. Client Selection in Federated Learning: Convergence Analysis and Power-of-Choice Selection Strategies. arXiv 2020, arXiv:2010.01243. [Google Scholar]
Figure 1. Federated learning consists of a server (coordinator) and a list of connected clients (devices). It takes action processes to train a global AI model in a distributed manner.
Figure 1. Federated learning consists of a server (coordinator) and a list of connected clients (devices). It takes action processes to train a global AI model in a distributed manner.
Applsci 14 05245 g001
Figure 2. An inference attack on federated learning.
Figure 2. An inference attack on federated learning.
Applsci 14 05245 g002
Figure 3. Architecture and operations of the mobility-based federated learning.
Figure 3. Architecture and operations of the mobility-based federated learning.
Applsci 14 05245 g003
Figure 4. Movement of three mobile nodes in the RPGM model.
Figure 4. Movement of three mobile nodes in the RPGM model.
Applsci 14 05245 g004
Figure 5. Vanilla FL with RPGM: performance of model accuracy (left) and training loss (right).
Figure 5. Vanilla FL with RPGM: performance of model accuracy (left) and training loss (right).
Applsci 14 05245 g005
Figure 6. Mobile FL with RPGM: performance of model accuracy (left) and training loss (right).
Figure 6. Mobile FL with RPGM: performance of model accuracy (left) and training loss (right).
Applsci 14 05245 g006
Figure 7. Vanilla FL with RWP: performance of model accuracy (left) and training loss (right).
Figure 7. Vanilla FL with RWP: performance of model accuracy (left) and training loss (right).
Applsci 14 05245 g007
Figure 8. Mobile FL with RWP: performance of model accuracy (left) and training loss (right).
Figure 8. Mobile FL with RWP: performance of model accuracy (left) and training loss (right).
Applsci 14 05245 g008
Table 1. Enumerated model accuracy and loss in the RPGM model.
Table 1. Enumerated model accuracy and loss in the RPGM model.
ModelAccuracy (%)Loss
Vanilla FL, z = 080.070.6301
Vanilla FL, z = 0.5
(privacy-enhanced)
77.350.6970
Mobile FL, z = 082.860.5542
Mobile FL, z = 0.5
(privacy-enhanced)
80.280.6327
Table 2. Enumerated model accuracy and loss in the RWP model.
Table 2. Enumerated model accuracy and loss in the RWP model.
ModelAccuracy (%)Loss
Vanilla FL, z = 073.170.8369
Vanilla FL, z = 0.5
(privacy-enhanced)
41.731.7769
Mobile FL, z = 082.260.5691
Mobile FL, z = 0.5
(privacy-enhanced)
79.870.6388
Table 3. Success rate of membership inference attack in four training models in the RPGM scenario.
Table 3. Success rate of membership inference attack in four training models in the RPGM scenario.
ModelAttack Success Rate
Vanilla FL, z = 00.77
Vanilla FL, z = 0.5 (privacy-enhanced)0.68
Mobile FL, z = 00.75
Mobile FL, z = 0.5 (privacy-enhanced)0.64
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, E.-j.; Lee, E.-K. Evaluating the Impact of Mobility on Differentially Private Federated Learning. Appl. Sci. 2024, 14, 5245. https://doi.org/10.3390/app14125245

AMA Style

Kim E-j, Lee E-K. Evaluating the Impact of Mobility on Differentially Private Federated Learning. Applied Sciences. 2024; 14(12):5245. https://doi.org/10.3390/app14125245

Chicago/Turabian Style

Kim, Eun-ji, and Eun-Kyu Lee. 2024. "Evaluating the Impact of Mobility on Differentially Private Federated Learning" Applied Sciences 14, no. 12: 5245. https://doi.org/10.3390/app14125245

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop