Next Article in Journal
PharmiTech: Addressing Polypharmacy Challenges through AI-Driven Solutions
Previous Article in Journal
Physicochemical Properties of Betacyclodextrin-Assisted Extracts of Green Rooibos (Aspalathus linearis)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Federated Learning for Mitigating Advanced Persistent Threats in Cyber-Physical Systems

1
Department of Electrical and Computer Engineering, University of Windsor, Windsor, ON N9B 3P4, Canada
2
Faculty of Computer Science, University of New Brunswick, Fredericton, NB E3B 5A3, Canada
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(19), 8840; https://doi.org/10.3390/app14198840
Submission received: 30 August 2024 / Revised: 25 September 2024 / Accepted: 27 September 2024 / Published: 1 October 2024

Abstract

:
Malware triage is essential for the security of cyber-physical systems, particularly against Advanced Persistent Threats (APTs). Proper data for this task, however, are hard to come by, as organizations are often reluctant to share their network data due to security concerns. To tackle this issue, this paper presents a secure and distributed framework for the collaborative training of a global model for APT triage without compromising privacy. Using this framework, organizations can share knowledge of APTs without disclosing private data. Moreover, the proposed design employs robust aggregation protocols to safeguard the global model against potential adversaries. The proposed framework is evaluated using real-world data with 15 different APT mechanisms. To make the simulations more challenging, we assume that edge nodes have partial knowledge of APTs. The obtained results demonstrate that participants in the proposed framework can privately share their knowledge, resulting in a robust global model that accurately detects APTs with significant improvement across different model architectures. Under optimal conditions, the designed framework detects almost all APT scenarios with an accuracy of over 90 percent.

1. Introduction

The security of the network layer in cyber-physical systems (CPSs) is of paramount importance in various domains [1,2,3]. While conventional Intrusion Detection Systems (IDSs) are effective at flagging common cyber-attacks in network traffic, they often struggle to differentiate more advanced threats that use a sequence of different attack mechanisms to achieve malicious objectives [4,5]. This limitation prevents security experts from fully understanding the true objectives of the attacks once they are detected.
Advanced Persistent Threats (APTs) are a prime example of these sophisticated attacks that jeopardize network security in CPSs [4,5,6]. In contrast to commonly known cyber-attacks, APTs are continuous attacks that take place over time. These targeted attacks are designed to remain undetected in the system for the duration of the attack. Mitigating APTs requires more complex solutions, as conventional defense mechanisms often fall short in detecting these attacks [7]. Failure to flag APT attacks can lead to catastrophic consequences, including financial problems and loss of information. For example, in 2009, an APT attack named Stuxnet hacked into the Supervisory Control and Data Acquisition (SCADA) system and Programmable Logic Controllers (PLCs) of a nuclear plant, aiming at the physical destruction of this facility [8].
These challenges necessitate tailored solutions to eliminate APTs promptly. Many research efforts are dedicated to the detection of APTs using machine learning techniques [9,10]. Detection models are the core of malware triage, distinguishing sophisticated malware and prioritizing it over regular malware [11,12]. In this process, the characteristics of the detected malware should be analyzed based on past events and the dynamics of the network so that the potential impact of the attack on the system can be estimated. This enables security teams to manage their resources and eliminate the threat more effectively. By doing so, organizations can identify and respond to APTs promptly, which, in turn, reduces the risk of long-term infiltration and damage [13]. In other words, utilizing this approach facilitates streamlining the response process, allowing security teams to focus on the most critical threats first, such as APTs, while efficiently handling less severe malware incidents.
Existing APT detection models can be categorized into anomaly detection and signature-based detection models [7]. Signature-based models usually employ supervised learning to distinguish between benign and malicious traffic [14,15]. Nonetheless, given the small proportion of APT samples in network traffic, signature-based models may fail during operation despite the high accuracy they exhibit during training (i.e., the accuracy paradox) [16]. Moreover, signature-based methods may struggle to detect unknown APT samples, as their supervised models usually recognize data distributions similar to those of their training data [17]. Conversely, anomaly detection methods are designed to detect abnormal patterns that may indicate an adversary in the network [10,14]. However, this approach usually creates an excessive workload for the security team when the false alarm rate increases. As a solution, rule mining can be used to rank the detected anomalies and assist in triaging them [10].
The similarity between the signature-based approach and the anomaly detection approach stems from their need for large amounts of training data [7,14]. Since these models depend on centralized data offerings, their scope of knowledge is quite limited, and hence the effectiveness of detection is adversely affected, especially against new APTs. Similar to other data-driven methods, building an automated system for the detection and triage of APTs requires ample high-quality data. Nevertheless, such data are hard to come by for several reasons. Firstly, due to the complexity and long-term nature of APTs, simulating these attacks in a lab setting is very challenging and time-consuming. Secondly, real-world data associated with real-world cases of APTs are very rare, as organizations are reluctant to share their network data publicly due to security concerns. One potential solution to this problem is using federated learning (FL), which enables collaborative learning via parameter sharing and data decentralization [18]. Although FL can be used as a solution for data decentralization in distributed training, its privacy and security are still concerning, as the FL surface itself may be subjected to adversaries, as investigated in the literature [19,20,21].
To address the aforementioned challenges, this paper makes the following contributions:
  • To tackle the issue of data scarcity and secure knowledge sharing, we design a secure detection framework through which different organizations can jointly train a global model for malware triage without sharing their private data, thereby adhering to privacy regulations.
  • In addition, adaptive clipping is used to safeguard the global model against adversaries that may degrade detection performance.
  • To ensure the detection model does not experience the accuracy paradox, the detection framework is evaluated using highly imbalanced and non-i.i.d. datasets across different nodes.
  • Experiments are conducted using public records associated with 15 APT cases to ensure the practicality of the designed framework for real-world applications.
The remainder of this paper is organized as follows. Section 2 states the targeted problem and briefly reviews the required background for this paper. Section 3 presents the proposed methodology. Section 4 illustrates and analyzes the experimental results. Finally, this paper concludes in Section 5.

2. Background

The contribution of this paper lies at the intersection of APT defense and FL. This section briefly reviews each of these topics to clarify the connection between these fields in this work. Then, the data utilized for simulating a case study is introduced at the end of this section.

2.1. Advanced Persistent Threats

The integration of physical components, a network layer, and an application layer in modern CPSs makes them susceptible to cyber-attacks [22,23,24]. An APT is one of the most concerning types of these threats [25]. These sophisticated attacks have very specific objectives and are executed gradually, with a tendency to maintain long-term access to the system. Moreover, the attackers operating an APT often orchestrate the attack themselves rather than using automated tools and codes. They are skilled and well resourced, combining several tools and techniques to carry out each step of the APT. Furthermore, the continuous and persistent nature of APTs makes them hard to detect by security teams, as they cannot definitively categorize an attack as an APT due to a lack of information. Thus, the security team must promptly detect any APTs in the system so that enough resources and manpower can be allocated to eliminate such threats.
The detection of APT attacks is often performed as part of the malware triage process, where different adversaries are ranked based on their degree of severity to set priorities for the security team. This data-driven process, however, is heavily reliant on the availability of data and historical patterns associated with critical threats such as APTs [5,15]. Several research endeavors have presented solutions for detecting APTs [14,26,27,28]. For instance, a risk management approach can be used to analyze the system’s expected state [29]. Multi-label learning has been shown to be effective at categorizing each malicious sample into different groups to provide extended insight into the data [30]. A dynamic quarantine and recovery scheme was designed in [6] to minimize the impact of APTs when they are present in a system. Nonetheless, despite the scarcity of data for effectively combating APTs, there are not many studies that address this problem. Privacy and security concerns associated with revealing network data prevent organizations from sharing such data. This paper tackles this issue by employing FL.

2.2. Federated Learning

FL is a machine learning paradigm that collaboratively trains a global model over several nodes or edge devices while preserving privacy [18,31,32]. User privacy in FL is ensured through data decentralization, as each edge device forbids access to the local data it contains [33,34,35]. Instead, nodes train a model on their local data and communicate the locally optimized parameters to an aggregator server. The aggregator receives these updates from edge devices and aggregates them into a global model [36]. Then, the parameters of this newly obtained model are sent to the edge devices so that they can update their local models to reflect the latest state of the global model. As this cycle continues, edge devices share the knowledge obtained from the data they contain and can improve their local models based on the information gathered from other nodes through parameter aggregation.
In the context of CPSs, FL can be used to build a global model to detect intrusions and cyber-attacks without the need to share private data [37,38,39]. Each organization or node of the network that has a local malware detection model can constantly improve its model based on the information that other nodes have acquired from their data, without having access to such data. However, employing FL creates an extra communication surface between the nodes of the network. As a result, if a node in this network becomes rogue, it can jeopardize the network through different adversaries [19,21,40,41,42]. This problem, however, is addressed by employing robust aggregation principles in our network, as explained in Section 3.

2.3. Case Study

To simulate real-world APTs in computer networks, we considered the data introduced in [11]. These data were created based on known APTs with publicly available reports. The dataset contains 2086 APT samples obtained from 15 different APTs, as listed in Table 1. A total of 9021 non-APT malware samples are also included in the data. The objective of this case study is to distinguish between APT and non-APT attacks so that malware triage can be performed accordingly. Table 1 indicates the population of samples associated with each APT mechanism. To minimize latency, samples were recorded using static features that could be extracted with minimal delay. Initially, 4000 features were extracted using the PEFrame tool. These features consisted of optional headers, MS-DOS headers, file headers, obfuscated string statistics, Mutex, Packer, buckets, and imported APIs. As investigated in [12], fewer than 300 of these features carry the required information for the task at hand. As a result, we selected 264 of the originally extracted features. Feature selection was performed based on the importance score given by random forest, as explained in [12].

3. Methodology

Figure 1 depicts the overall structure of the APT detection framework. Considering a set of nodes using a neural network model M in the FL network, each node contains a local dataset D = { X , Y } , where X = { x 1 , x 2 , , x m } and Y = { y 1 , y 2 , , y m } denote the set of samples and labels, respectively. We assume that X R n and Y N . Furthermore, the data are non-i.i.d. across different nodes. In a security context, X denotes the samples recorded from network traffic, whereas Y denotes the labels specifying the attack mechanisms with which the recorded samples are associated.
At time step t, a model is trained at node k on D k t and optimizes the parameters, which in our case are the weights of the network W k t = { w 1 , w 2 , w l } , where l is the number of layers. W k t is then communicated to the aggregation server. The server receives W k t for N nodes and commences the aggregation process. As a common standard, the server can zero out abnormally large values with a threshold T 0 to eliminate the effect of potential data poisoning attacks. Assuming that w i j indicates the j-th element of w i W k t , then w i j W k t , 1 k N :
Clip ( w i j , T 0 ) = w i j if | w i j | T 0 0 otherwise
If a client becomes rogue and acts maliciously, they can disrupt the global model by communicating a set of poisoned parameters. While such attacks may vary in their objectives, the adversarial task of a malicious client is generally to maximize the loss of the global model as follows:
max ξ k L ( W ¯ k t , ξ k ) ,
where L is a loss function that returns the global model error given poisoned updates obtained by perturbation ξ k . Depending on the attack mechanism (e.g., model poisoning, data poisoning), ξ k can be applied to the local data as D k t + ξ k or directly to the resulting gradients as W k t + ξ k .
The server then uses the clipped weights to aggregate the updates. Given N nodes in the network, Federated Averaging (FedAvg) [18] is defined as follows:
FedAvg { W k } k = 1 N = 1 N k = 1 N W k t
The server continuously receives W k t from nodes and computes Clip ( W k t , T 0 t ) to obtain the clipped weights Δ ^ k t . Δ ^ k t is then passed as FedAvg ( { Δ ^ k t } k = 1 N ) to update the weights of the global model W ¯ t + 1 for the next iteration t + 1 (see line 10 in Algorithm 1).
Nonetheless, finding a fixed threshold for T 0 that optimally works for all nodes and datasets is not feasible in most cases. As a solution, one can dynamically adjust the threshold T. We follow the adaptive clipping approach proposed in [43] and employ an adaptive update rule. T is adjusted to find an approximate value at a specified quantile of the update norm distribution. Assuming that γ [ 0 , 1 ] is the targeted quantile, for any T, the Clip ( · , · ) function can be reformulated to the loss function J with the following definition:
J ( X , T ) = ( 1 γ ) ( T X ) if X T γ ( X T ) otherwise
The targeted quantile is selected based on the expected risk of adversaries in the system. Similar to most anomaly detection-based defense mechanisms in federated learning, adjusting this parameter involves a trade-off between security and performance. For large values of γ , gradients are aggressively clipped, which results in highly secure aggregation. Conversely, choosing small values of γ results in a more accurate model with a higher risk of being compromised.
Algorithm 1: FedAvg with Adaptive Clipping
Applsci 14 08840 i001
Algorithm 2: Client Update
Applsci 14 08840 i002
Aiming to find an update rule that leads to the optimal T for a specified value of γ , we treat the above formulation as an optimization problem. To do so, the average derivative of J for timestep t is calculated as follows:
J ¯ ( X , T t ) = 1 m i = 1 m ( 1 γ ) if x i T t γ if x i > T t = 1 m ( 1 γ ) i = 1 m I x i T t γ i = 1 m I x i > T t 1 m i = 1 m I x i T t γ
The update rule for estimating T t + 1 can be obtained as follows:
T t + 1 T t η T 1 m i = 1 m I x i T t γ ,
where η T is the learning rate used for the threshold estimation. Equation (5) enhances the model’s performance and robustness by dynamically adapting the threshold to changes in the update distribution.
In the obtained update rule in (6), the value of T linearly changes with a maximum of η T in each iteration. This can be problematic when the desired T is significantly larger or smaller than the current value. To address this problem, a geometric form of this update rule is employed to speed up and improve the convergence of the function, as follows:
T t + 1 T t . exp η T 1 m i = 1 m I x i T t γ
Each node communicates both W k t and T k t to the server. FedAvg with adaptive clipping is defined in Algorithm 1, which details the aggregation operation on the server side. Client-side operations are defined in Algorithm 2 based on the aforementioned formulations.

4. Experimental Results

Figure 2 depicts the flow of experiments performed under centralized and decentralized settings, which are described as follows:
  • Centralized training: Initially, an algorithm was selected to train ten different models on different subsets of training data. Each model was then evaluated on the test dataset to analyze its performance in a centralized setting.
  • Decentralized training: In a decentralized setting, the designed framework was used to collaboratively train a model using ten different nodes that contained the same training subsets. Once the training was completed, the global state of the model, which was accessible by all nodes, was evaluated on the test dataset.
Each experiment was repeated ten times to ensure statistical reliability; consequently, the results were averaged to report the overall performance.

4.1. Experimental Settings

Three neural network models—the One-Dimensional Convolutional Neural Network (1D CNN) [44], Denoising Autoencoder (DAE) [45], and Multi-Layer Perceptron (MLP)—were used as the base learners. All models used the Adam optimizer [46]. These techniques and the proposed FL framework were implemented in TensorFlow using Python 3.10.14 The FL framework was executed for 100 training rounds. The learning rate for adaptive clipping was empirically set to η T = 0.2 . The parameter settings for the local models are reported in Table 2.
To create non-i.i.d. datasets for different nodes of the FL network, ten datasets were considered, as shown in Table 3. All datasets contained non-APT malware along with five other APT classes that were randomly selected, with the condition that each APT class was included in at least one of the datasets. A test subset was created by selecting 20 percent of the samples from each class, as shown in Figure 2. The training datasets listed in Table 3 were drawn from the remaining 80 percent.

4.2. Results

In order to study the effect of FL on constructing an accurate global model for APT detection, we first analyzed the results for the MLP, CNN, and 1D CNN in ten different scenarios. In each scenario, a model was trained on only one of the datasets listed in Table 3. The trained model was then evaluated on the test subset, which contained 20 percent of the samples from each class.
Figure 3 shows the obtained results using a heatmap for each algorithm. The datasets used for training the model in each experiment are indicated on the vertical axis, whereas the APT classes are specified along the horizontal axis. The color spectrum on the right side of each box indicates the range of colors used to differentiate between smaller and larger values. Figure 3a illustrates that the MLP was not able to detect APT classes with acceptable accuracy when trained on a subset of data with only a few APT classes included. Given that the majority of samples were non-APT malware, the model exhibited better performance in detecting malware. Among the datasets used for training, datasets six to eight resulted in better overall performance across all classes. Comparing the spectrum of colors in Figure 3b with the other panels in this figure, it can be concluded that the 1D CNN generally outperformed the MLP and DAE when trained on partial data. Nevertheless, this model seemed to exhibit poorer performance in detecting samples of the APT28 class than its competitors. Interestingly, the 1D CNN’s performance was more consistent across different experiments and scenarios. On the other hand, the DAE exhibited poorer detection performance than the MLP and 1D CNN, as the color spectrum shown in Figure 3c is generally darker than those in Figure 3a,b. Regardless, the achieved results showed less variance than those of the MLP. Although using a more advanced structure with more hidden layers could probably boost the performance of the MLP, 1D CNN, and DAE, we relied on simple structures to study the effect of using FL on boosting detection performance.
By integrating the selected classifiers into the designed FL structure, different nodes of the network can collaboratively train a global detection model to detect APT samples. Here, the FL network used ten nodes, where each node trained a local model on only one of the datasets listed in Table 3. Moreover, all of these nodes used a common network structure. Figure 4 depicts the cross-entropy loss of the global model through the iterations of the collaborative training. As nodes shared their parameters and updated their local weights in each round, the global loss decreased on the training data. This ensured that the global model was a perfect fit that successfully captured the APT distributions of each node. Furthermore, the global model converged rapidly and reached the optimal loss in fewer than 50 rounds of training. This implies the practicality of using FL in the collaborative training of a global APT detection model in CPSs. This application is vital to more data-sensitive CPSs, such as power systems and industrial control systems, where sensitive operational parameters and communication protocols can be targeted by APTs.
The obtained results shown in Figure 5 indicate a significant improvement in the detection performance of both the APT classes and malware. Regardless of the choice of model architecture at the edge of the network, the proposed FL framework outperformed the stand-alone models. While all three models exhibited outstanding results in detecting the APT classes, the 1D CNN outperformed the DAE and MLP combinations with the designed framework, as depicted in Figure 5b. The only class whose detection accuracy fell below 90 percent was Sandwork. Nevertheless, this pattern can also be seen for both the MLP and DAE in Figure 5a,c.
The observed improvements in the previous analysis indicate the significance of FL application across different CPSs. To begin with, accurate APT detection can prevent significant operational disruptions in industrial control systems. Another notable application of FL is in smart grids, which rely on distributed communication and control. These systems are often vulnerable to malicious activities and can benefit from the timely and accurate identification of adversaries.

5. Conclusions

Aiming to mitigate APT attacks in CPSs, this work designed a framework for the collaborative training of detection models used for the task of malware triage. Moreover, the proposed method tackles the issue of data scarcity for the task at hand by replacing data sharing with parameter sharing, thereby eliminating the privacy and security concerns associated with sharing private data between organizations. While the proposed method is based on FL, it also employs adaptive norm clipping to mitigate potential adversaries that may target the federated network during the collaborative training session. This adds an extra layer of security to the FL structure and safeguards the system against a wide range of adversaries. The proposed method was evaluated on real-world data for APT triage through several scenarios. Finally, the obtained results showed that the designed framework effectively builds a global model that successfully detects all 15 APT classes and the non-APT malware considered in this work. For future work, extending the design to employ privacy-preserving mechanisms can further enhance the model’s robustness. In addition, more APT variants can be considered in the simulations to evaluate the framework’s adaptability and generalization to new threats.

Author Contributions

Conceptualization, E.H. and R.R.-F.; methodology, E.H. and R.R.-F.; software, E.H.; validation, E.H.; formal analysis, E.H.; investigation, E.H.; resources, R.R.-F. and M.S.; data curation, E.H.; writing—original draft preparation, E.H.; writing—review and editing, R.R.-F. and M.S.; visualization, E.H.; supervision, R.R.-F. and M.S.; project administration, R.R.-F. and M.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Sciences and Engineering Research Council of Canada (NSERC) under funding reference numbers CGSD3-569341-2022 and RGPIN-2021-02968.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://github.com/GiuseppeLaurenza/I_F_Identifier/blob/master/dataset.tar.gz (accessed on 26 September 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kim, S.; Park, K.J.; Lu, C. A Survey on Network Security for Cyber–Physical Systems: From Threats to Resilient Design. IEEE Commun. Surv. Tutor. 2022, 24, 1534–1573. [Google Scholar] [CrossRef]
  2. Ahmed Jamal, A.; Mustafa Majid, A.A.; Konev, A.; Kosachenko, T.; Shelupanov, A. A review on security analysis of cyber physical systems using Machine learning. Mater. Today Proc. 2023, 80, 2302–2306. [Google Scholar] [CrossRef]
  3. Humayed, A.; Lin, J.; Li, F.; Luo, B. Cyber-Physical Systems Security—A Survey. IEEE Internet Things J. 2017, 4, 1802–1831. [Google Scholar] [CrossRef]
  4. Huang, L.; Zhu, Q. A dynamic games approach to proactive defense strategies against Advanced Persistent Threats in cyber-physical systems. Comput. Secur. 2020, 89, 101660. [Google Scholar] [CrossRef]
  5. Rahman, Z.; Yi, X.; Khalil, I. Blockchain-Based AI-Enabled Industry 4.0 CPS Protection Against Advanced Persistent Threat. IEEE Internet Things J. 2023, 10, 6769–6778. [Google Scholar] [CrossRef]
  6. Yang, L.X.; Li, P.; Yang, X.; Xiang, Y.; Jiang, F.; Zhou, W. Effective Quarantine and Recovery Scheme Against Advanced Persistent Threat. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 5977–5991. [Google Scholar] [CrossRef]
  7. Alshamrani, A.; Myneni, S.; Chowdhary, A.; Huang, D. A Survey on Advanced Persistent Threats: Techniques, Solutions, Challenges, and Research Opportunities. IEEE Commun. Surv. Tutor. 2019, 21, 1851–1877. [Google Scholar] [CrossRef]
  8. Langner, R. Stuxnet: Dissecting a Cyberwarfare Weapon. IEEE Secur. Priv. 2011, 9, 49–51. [Google Scholar] [CrossRef]
  9. Jia, Z.; Xiong, Y.; Nan, Y.; Zhang, Y.; Zhao, J.; Wen, M. MAGIC: Detecting Advanced Persistent Threats via Masked Graph Representation Learning. In Proceedings of the 33rd USENIX Security Symposium (USENIX Security 24), Philadelphia, PA, USA, 14–16 August 2024; pp. 5197–5214. [Google Scholar]
  10. Dong, F.; Wang, L.; Nie, X.; Shao, F.; Wang, H.; Li, D.; Luo, X.; Xiao, X. DISTDET: A Cost-Effective Distributed Cyber Threat Detection System. In Proceedings of the 32nd USENIX Security Symposium (USENIX Security 23), Anaheim, CA, USA, 9–11 August 2023; pp. 6575–6592. [Google Scholar]
  11. Laurenza, G.; Aniello, L.; Lazzeretti, R.; Baldoni, R. Malware Triage Based on Static Features and Public APT Reports. In Proceedings of the Cyber Security Cryptography and Machine Learning, Cham, Switzerland, 29–30 June 2017; pp. 288–305. [Google Scholar]
  12. Laurenza, G.; Lazzeretti, R.; Mazzotti, L. Malware Triage for Early Identification of Advanced Persistent Threat Activities. Digit. Threat. 2020, 1, 16. [Google Scholar] [CrossRef]
  13. Sharma, A.; Gupta, B.B.; Singh, A.K.; Saraswat, V.K. Advanced Persistent Threats (APT): Evolution, anatomy, attribution and countermeasures. J. Ambient. Intell. Humaniz. Comput. 2023, 14, 9355–9381. [Google Scholar] [CrossRef]
  14. Abu Talib, M.; Nasir, Q.; Bou Nassif, A.; Mokhamed, T.; Ahmed, N.; Mahfood, B. APT beaconing detection: A systematic review. Comput. Secur. 2022, 122, 102875. [Google Scholar] [CrossRef]
  15. Akbar, K.A.; Wang, Y.; Ayoade, G.; Gao, Y.; Singhal, A.; Khan, L.; Thuraisingham, B.; Jee, K. Advanced Persistent Threat Detection Using Data Provenance and Metric Learning. IEEE Trans. Dependable Secur. Comput. 2023, 20, 3957–3969. [Google Scholar] [CrossRef]
  16. Thomas, C.; Balakrishnan, N. Improvement in minority attack detection with skewness in network traffic. In Proceedings of the Data Mining, Intrusion Detection, Information Assurance, and Data Networks Security, Orlando, FL, USA, 17–18 March 2008; Volume 6973, pp. 226–237. [Google Scholar]
  17. Chen, P.; Desmet, L.; Huygens, C. A Study on Advanced Persistent Threats. In Proceedings of the Communications and Multimedia Security, Aveiro, Portugal, 25–26 September 2014; pp. 63–72. [Google Scholar]
  18. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; Arcas, B.A.y. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, Ft. Lauderdale, FL, USA, 20–22 April 2017; Volume 54, pp. 1273–1282. [Google Scholar]
  19. Hallaji, E.; Razavi-Far, R.; Saif, M.; Wang, B.; Yang, Q. Decentralized Federated Learning: A Survey on Security and Privacy. IEEE Trans. Big Data 2024, 10, 194–213. [Google Scholar] [CrossRef]
  20. Bhagoji, A.N.; Chakraborty, S.; Mittal, P.; Calo, S. Analyzing Federated Learning through an Adversarial Lens. In Proceedings of the 36th International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 634–643. [Google Scholar]
  21. Hallaji, E.; Razavi-Far, R.; Saif, M. Federated and Transfer Learning: A Survey on Adversaries and Defense Mechanisms; Springer: Berlin, Germany, 2022; pp. 29–55. [Google Scholar]
  22. Han, S.; Xie, M.; Chen, H.H.; Ling, Y. Intrusion Detection in Cyber-Physical Systems: Techniques and Challenges. IEEE Syst. J. 2014, 8, 1052–1062. [Google Scholar]
  23. Farajzadeh-Zanjani, M.; Hallaji, E.; Razavi-Far, R.; Saif, M. Generative-Adversarial Class-Imbalance Learning for Classifying Cyber-Attacks and Faults—A Cyber-Physical Power System. IEEE Trans. Dependable Secur. Comput. 2022, 19, 4068–4081. [Google Scholar] [CrossRef]
  24. Zhu, T.; Ye, D.; Cheng, Z.; Zhou, W.; Yu, P.S. Learning Games for Defending Advanced Persistent Threats in Cyber Systems. IEEE Trans. Syst. Man Cybern. Syst. 2023, 53, 2410–2422. [Google Scholar] [CrossRef]
  25. Kayan, H.; Nunes, M.; Rana, O.; Burnap, P.; Perera, C. Cybersecurity of Industrial Cyber-Physical Systems: A Review. ACM Comput. Surv. 2022, 54, 1–35. [Google Scholar] [CrossRef]
  26. Zhu, T.; Yu, J.; Xiong, C.; Cheng, W.; Yuan, Q.; Ying, J.; Chen, T.; Zhang, J.; Lv, M.; Chen, Y.; et al. APTSHIELD: A Stable, Efficient and Real-Time APT Detection System for Linux Hosts. IEEE Trans. Dependable Secur. Comput. 2023, 20, 5247–5264. [Google Scholar] [CrossRef]
  27. Liang, H.; Li, C.; Li, X.; Jiang, S. APT Malware Classification Method Based on Feature Fusion. In Proceedings of the International Conference on Computer Information Science and Artificial Intelligence (CISAI), Kunming, China, 17–19 September 2021; pp. 456–462. [Google Scholar]
  28. Pitolli, G.; Laurenza, G.; Aniello, L.; Querzoni, L.; Baldoni, R. MalFamAware: Automatic family identification and malware classification through online clustering. Int. J. Inf. Secur. 2021, 20, 371–386. [Google Scholar] [CrossRef]
  29. Yang, L.X.; Li, P.; Yang, X.; Tang, Y.Y. A Risk Management Approach to Defending Against the Advanced Persistent Threat. IEEE Trans. Dependable Secur. Comput. 2020, 17, 1163–1172. [Google Scholar] [CrossRef]
  30. Hallaji, E.; Razavi-Far, R.; Saif, M. Expanding analytical capabilities in intrusion detection through ensemble-based multi-label classification. Comput. Secur. 2024, 139, 103730. [Google Scholar] [CrossRef]
  31. Razavi-Far, R.; Wang, B.; Taylor, M.E.; Yang, Q. An Introduction to Federated and Transfer Learning. In Federated and Transfer Learning; Springer International Publishing: Cham, Switzerland, 2023; pp. 1–6. [Google Scholar]
  32. Zhang, C.; Xie, Y.; Bai, H.; Yu, B.; Li, W.; Gao, Y. A survey on federated learning. Knowl. Based Syst. 2021, 216, 106775. [Google Scholar] [CrossRef]
  33. Preuveneers, D.; Rimmer, V.; Tsingenopoulos, I.; Spooren, J.; Joosen, W.; Ilie-Zudor, E. Chained Anomaly Detection Models for Federated Learning: An Intrusion Detection Case Study. Appl. Sci. 2018, 8, 2663. [Google Scholar] [CrossRef]
  34. Asad, M.; Moustafa, A.; Ito, T. FedOpt: Towards Communication Efficiency and Privacy Preservation in Federated Learning. Appl. Sci. 2020, 10, 2864. [Google Scholar] [CrossRef]
  35. Mothukuri, V.; Parizi, R.M.; Pouriyeh, S.; Huang, Y.; Dehghantanha, A.; Srivastava, G. A survey on security and privacy of federated learning. Future Gener. Comput. Syst. 2021, 115, 619–640. [Google Scholar] [CrossRef]
  36. Qi, P.; Chiaro, D.; Guzzo, A.; Ianni, M.; Fortino, G.; Piccialli, F. Model aggregation techniques in federated learning: A comprehensive survey. Future Gener. Comput. Syst. 2024, 150, 272–293. [Google Scholar] [CrossRef]
  37. Agrawal, S.; Sarkar, S.; Aouedi, O.; Yenduri, G.; Piamrat, K.; Alazab, M.; Bhattacharya, S.; Maddikunta, P.K.R.; Gadekallu, T.R. Federated Learning for intrusion detection system: Concepts, challenges and future directions. Comput. Commun. 2022, 195, 346–361. [Google Scholar] [CrossRef]
  38. Hallaji, E.; Razavi-Far, R.; Saif, M.; Herrera-Viedma, E. Label noise analysis meets adversarial training: A defense against label poisoning in federated learning. Knowl. Based Syst. 2023, 266, 110384. [Google Scholar] [CrossRef]
  39. Rahman, S.A.; Tout, H.; Talhi, C.; Mourad, A. Internet of Things Intrusion Detection: Centralized, On-Device, or Federated Learning? IEEE Netw. 2020, 34, 310–317. [Google Scholar] [CrossRef]
  40. Blanchard, P.; El Mhamdi, E.M.; Guerraoui, R.; Stainer, J. Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  41. Fang, M.; Cao, X.; Jia, J.; Gong, N. Local Model Poisoning Attacks to Byzantine-Robust Federated Learning. In Proceedings of the 29th USENIX Security Symposium. USENIX Association, Berkeley, CA, USA, 12–14 August 2020; pp. 1605–1622. [Google Scholar]
  42. Reisizadeh, A.; Farnia, F.; Pedarsani, R.; Jadbabaie, A. Robust Federated Learning: The Case of Affine Distribution Shifts. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–12 December 2020; Curran Associates, Inc.: New York, NY, USA, 2020; Volume 33, pp. 21554–21565. [Google Scholar]
  43. Andrew, G.; Thakkar, O.; McMahan, B.; Ramaswamy, S. Differentially Private Learning with Adaptive Clipping. In Proceedings of the International Conference on Neural Information Processing Systems, Online, 6–14 December 2021; Volume 34, pp. 17455–17466. [Google Scholar]
  44. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  45. Vincent, P.; Larochelle, H.; Bengio, Y.; Manzagol, P.A. Extracting and composing robust features with denoising autoencoders. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 1096–1103. [Google Scholar]
  46. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2017, arXiv:1412.6980. [Google Scholar]
Figure 1. Block diagram of the APT detection framework. Algorithms 1 and 2 are used for robust aggregation and client updates, respectively.
Figure 1. Block diagram of the APT detection framework. Algorithms 1 and 2 are used for robust aggregation and client updates, respectively.
Applsci 14 08840 g001
Figure 2. Block diagram of the experimental flow. Centralized and decentralized experiments are shown on the left and right sides of the diagram, respectively. The experimental dataset, consisting of APT and non-APT samples, was divided into training and test subsets to ensure that both centralized and decentralized experiments used the same datasets.
Figure 2. Block diagram of the experimental flow. Centralized and decentralized experiments are shown on the left and right sides of the diagram, respectively. The experimental dataset, consisting of APT and non-APT samples, was divided into training and test subsets to ensure that both centralized and decentralized experiments used the same datasets.
Applsci 14 08840 g002
Figure 3. Success rates of detecting different APT classes without using FL. In each experiment, the indicated model is trained on only one dataset and evaluated on the test subset. (ac) show the results of MLP, 1D-CNN, and DAE, respectively.
Figure 3. Success rates of detecting different APT classes without using FL. In each experiment, the indicated model is trained on only one dataset and evaluated on the test subset. (ac) show the results of MLP, 1D-CNN, and DAE, respectively.
Applsci 14 08840 g003
Figure 4. Cross-entropy loss of training a global model using FL with adaptive clipping. Ten nodes participate in the FL training process.
Figure 4. Cross-entropy loss of training a global model using FL with adaptive clipping. Ten nodes participate in the FL training process.
Applsci 14 08840 g004
Figure 5. Detection performance under centralized and decentralized settings. (ac) show the results of MLP, 1D-CNN, and DAE, respectively.
Figure 5. Detection performance under centralized and decentralized settings. (ac) show the results of MLP, 1D-CNN, and DAE, respectively.
Applsci 14 08840 g005
Table 1. APT mechanisms included in the dataset [11]. Class populations are indicated accordingly.
Table 1. APT mechanisms included in the dataset [11]. Class populations are indicated accordingly.
LabelClassNo. Samples
1APT2868
2APT29205
3APT30101
4Carbanak105
5Desert Falcon45
6Hurricane Panda315
7Lazarus Group58
8Mirage54
9Patchwork559
10Sandwork44
11Shiqiang31
12Transparent Tribe267
13Violin Panda23
14Volatile Cedar35
15Winnti Group176
16Non-APT Malware9021
Table 2. Parameter settings used for edge local models.
Table 2. Parameter settings used for edge local models.
Parameter1D CNNDAEMLP
Epochs50010010
Batch size6425632
No. hidden layers221
No. neurons64, 64128, 256128, 64
Learning rate0.0010.0010.002
Noise factor-0.5-
Table 3. Scenarios used to create training datasets for each experiment. Each FL node uses only one of these datasets.
Table 3. Scenarios used to create training datasets for each experiment. Each FL node uses only one of these datasets.
DatasetClassesNo. Samples
Dataset 1Malware, Transparent Tribe, Winnti Group, Carbanak, Desert Falcon, Violin Panda7710
Dataset 2Malware, APT29, Lazarus Group, Mirage, Sandwork, Violin Panda7523
Dataset 3Malware, Patchwork, Transparent Tribe, Carbanak, APT28, Volatile Cedar8043
Dataset 4Malware, Carbanak, APT28, Mirage, Shiqiang, Violin Panda7441
Dataset 5Malware, Winnti Group, Carbanak, APT30, Volatile Cedar, Violin Panda7569
Dataset 6Malware, Patchwork, Winnti Group, Lazarus Group, Mirage, Volatile Cedar7921
Dataset 7Malware, Hurricane Panda, Transparent Tribe, APT29, APT30, Shiqiang7952
Dataset 8Malware, Patchwork, Transparent Tribe, APT29, APT28, Lazarus Group8141
Dataset 9Malware, Transparent Tribe, Winnti Group, Mirage, Shiqiang, Violin Panda7658
Dataset 10Malware, Carbanak, APT30, Lazarus Group, Mirage, Desert Falcon7506
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hallaji, E.; Razavi-Far, R.; Saif, M. Robust Federated Learning for Mitigating Advanced Persistent Threats in Cyber-Physical Systems. Appl. Sci. 2024, 14, 8840. https://doi.org/10.3390/app14198840

AMA Style

Hallaji E, Razavi-Far R, Saif M. Robust Federated Learning for Mitigating Advanced Persistent Threats in Cyber-Physical Systems. Applied Sciences. 2024; 14(19):8840. https://doi.org/10.3390/app14198840

Chicago/Turabian Style

Hallaji, Ehsan, Roozbeh Razavi-Far, and Mehrdad Saif. 2024. "Robust Federated Learning for Mitigating Advanced Persistent Threats in Cyber-Physical Systems" Applied Sciences 14, no. 19: 8840. https://doi.org/10.3390/app14198840

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop