Next Article in Journal
Hidden Abstract Stack Markov Models with Learning Process
Previous Article in Journal
Federated Transfer Fault Diagnosis Method Based on Variational Auto-Encoding with Few-Shot Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Novel Dynamic Defense Strategies in Networked Control Systems under Stochastic Jamming Attacks

Laboratory of Systems Communications Sys’Com, National Engineering School of Tunis, University of Tunis El Manar, Tunis 7000, Tunisia
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(13), 2143; https://doi.org/10.3390/math12132143
Submission received: 8 June 2024 / Revised: 4 July 2024 / Accepted: 5 July 2024 / Published: 8 July 2024

Abstract

:
In contemporary networked control systems (NCSs), ensuring robust and adaptive security measures against dynamic threats like jamming attacks is crucial. These attacks can disrupt the control signals, leading to degraded performance or even catastrophic failures. This paper introduces a novel approach to enhance NCS security by applying stochastic game theory to model and resolve interactions between a defender and a jammer. We develop a two-player zero-sum game where the defender employs mixed strategies to minimize the expected cost of maintaining system stability and control effectiveness in the face of potential jamming. Our model discretizes the state space and employs backward induction to dynamically update the value functions associated with various system states, reflecting the ongoing adjustment of strategies in response to the adversary’s actions. Utilizing linear programming in MATLAB, we optimize the defender’s mixed strategies to systematically mitigate the impact of jamming. The results from extensive simulations demonstrate the efficacy of our proposed strategies in attack scenarios, indicating a substantial enhancement in the resilience and performance of NCSs against jamming attacks. Specifically, the proposed method improved network state stability by 75 % , reducing the fluctuation range by over 50 % compared with systems without defense mechanisms. This study not only advances the theoretical framework for security in NCSs but also provides practical insights for the design of resilient control systems under uncertainty.

1. Introduction

Networked control systems (NCSs) are pivotal in modern infrastructure, managing everything from transportation networks to power grids [1,2]. The integration of computing and communication technologies with control systems has significantly enhanced their efficiency and flexibility. However, this integration also exposes NCSs to various cyber threats, among which jamming attacks represent a severe risk. Jamming attacks deliberately disrupt communication channels, thereby degrading the performance of control systems or even causing total system failure [3]. The dynamic and unpredictable nature of these attacks requires equally adaptive and robust defense mechanisms. Traditional security measures often focus on static defenses or reactive protocols that might not suffice in the face of sophisticated, evolving threats. Thus, there is a pressing need for strategies that anticipate and mitigate potential attacks in real time. Stochastic game theory [4,5] provides a promising framework to address these challenges.
By modeling the interactions between defenders and attackers as a game with probabilistic outcomes, we can develop strategies that adapt to changes in the environment and the attacker’s behavior.
This paper applies stochastic game theory to the specific case of jamming in NCS, proposing a two-player zero-sum game where the defender aims to minimize a cost function that represents the risk and impact of potential disruptions.
Our approach involves discretizing the state space of the system, allowing for a detailed analysis of possible states and transitions. Backward induction is used to iteratively calculate value functions, representing the minimum expected future costs from any given state, under the assumption of optimal play by both the defender and the attacker. This dynamic programming method [6,7] not only accommodates the stochastic nature of the threat but also provides a framework for updating defense strategies as the game progresses. The introduction of linear programming to optimize the defender’s mixed strategies further enhances the model’s practicality by providing a method to determine the optimal defensive actions quickly and efficiently. This integration of stochastic game theory with linear optimization offers new insights into developing proactive defense mechanisms in NCSs, aiming to keep the systems resilient against unpredictable adversarial behaviors [8].
In the literature, there are many algorithms to tackle the challenges of the security of NCSs. To better understand the strengths and weaknesses of various algorithms in the context of networked control system (NCS) security, Table 1 summarizes the advantages and disadvantages of static defenses, reactive protocols, stochastic game theory, linear programming optimization, machine learning approaches, and backward induction.
This article aims to propose a game-theoretical approach to address security issues in NCSs under intelligent jamming attacks. The major contributions of the proposed model are underlined as follows:
  • This research introduces a refined model of stochastic game theory specifically adapted for handling security issues in NCSs under jamming attacks. It extends traditional game-theoretic approaches by incorporating dynamic and stochastic elements that more accurately reflect the complexities encountered in real-world scenarios.
  • The use of backward induction to iteratively update value functions offers a robust framework for real-time decision making. This method allows for the strategic adjustments of defense based on evolving system states and attacker strategies, providing a continuous adaptation mechanism that is vital in dealing with dynamic threats.
  • The application of linear programming to determine the defender’s mixed strategies is a practical contribution that enhances the computational feasibility of implementing game-theoretic solutions in real systems. This optimization technique allows for quick recalculations, which is essential for systems requiring immediate responses to security breaches.

2. Existing Literature

The security of networked control systems (NCSs) against cyber-physical attacks, particularly jamming, has garnered increasing attention due to the critical roles these systems play in modern infrastructure. This section reviews pertinent literature on the application of game theory to NCS security, the use of stochastic models in cybersecurity, and existing methods for real-time adaptive strategy development.
Early works, such as those by Zhu and Basar in [9], explored game theory as a tool to model the interactions between attackers and defenders in cyber-physical systems. They demonstrated how game theory could provide strategic insights into optimal defensive actions. However, many of these models assumed static game settings and did not account for the dynamic nature of attack vectors.
More recent studies have begun to incorporate dynamic game elements. For instance, Nguyen et al. in [10] extended zero-sum game models to include time-varying strategies and demonstrated their effectiveness in simulated environments. While these studies advanced the application of game theory in cybersecurity, they often neglected the stochastic variations in the state transitions and the real-time recalibration of strategies.
Stochastic models have been effectively used to describe uncertainties in cyber-attacks. The authors in [11] employed Markov decision processes (MDPs) to assess risk and devise countermeasures in real time. Their work provided foundational methodologies for understanding probabilistic attack impacts but did not directly address jamming in NCSs.
The adaptation of stochastic models specifically to jamming attacks can be seen in the work cited in [12], which used stochastic games to analyze the effectiveness of various jamming techniques against wireless communication systems. These contributions are crucial but often require further refinement to integrate real-time defense adaptations in NCSs.
The challenge of developing adaptive strategies that can respond in real time to evolving threats has led to innovative solutions, such as those proposed by Yu et al. in [13]. They introduced a learning-based algorithm that adjusts defensive tactics based on ongoing attack patterns, showcasing significant improvements in system resilience.
While the work of the authors in [14] considers that the attack signal depends on the results of the attack–defense game between the jammer and the transmitter, here, the state of the attack signal follows a random process to model the interaction between the players as close as possible to real-world scenarios.
Wang et al. [15] explored remote state estimation under DoS attacks in CPSs with arbitrary tree topology using a Bayesian Stackelberg game approach, highlighting the robustness of strategic defense mechanisms in complex network structures. Zhou et al., in [16], provided a comprehensive review of the cybersecurity landscape on remote state estimation, offering a broad perspective on the challenges and advancements in this field. The authors in [17] examined remote state estimation with a strategic sensor using a Stackelberg game framework, contributing to the understanding of strategic interactions in cybersecurity. In [18], Wu et al. presented a game-theoretic approach to remote state estimation in the presence of a DoS attacker, which emphasizes the relevance of game theory in designing resilient estimation strategies and finally, Yang et al., in [19], discussed cyber attacks on remote state estimation in cyber-physical systems using a game-theoretic approach, demonstrating the applicability of these models in real-world scenarios. In summary, while substantial research has explored various aspects of NCS security, the integration of game theory with stochastic modeling and linear programming for dynamic and real-time strategy optimization remains underexplored. Our work contributes to this niche by providing a comprehensive framework that not only anticipates potential attack scenarios but also adapts to them dynamically, ensuring sustained system performance and resilience.

3. Proposed Methodology

This section provides a comprehensive explanation of the mathematical frameworks and computational techniques used in our study to enhance the security of networked control systems against stochastic jamming attacks. The methodology is built around the formulation of a stochastic game model, the discretization of state spaces, the definition of cost functions reflecting system penalties under attack, and the employment of linear programming for optimizing the defender’s mixed strategies.

3.1. Modeling an NCS under Jamming Attacks

Networked control systems (NCSs) are particularly susceptible to physical attacks due to the way in which feedback measurements are transmitted [20]. In these systems, as shown in Figure 1, feedback data, crucial for the operation and control of physical processes, are often sent over wireless communication networks. This mode of transmission exposes the data to various security threats, including jamming attacks where a jammer sends noise signals that are greater than the transmitting signals of the transmitter, hence disrupting the system functionality and losing the feedback measurements [21]. Without these measurements, the NCS becomes an open-loop system, which leads to instability.
The dynamics of the NCS are modeled using a linear state transition equation influenced by potential jamming attacks, represented as stochastic disturbances in the control input. The state transition is described by
x k + 1 = A x k + γ k B K u k
where the following hold:
  • x k is the state vector at time step k;
  • A is the state transition matrix describing the system dynamics under normal operation;
  • γ k is a binary variable that represents the presence (1) or absence (0) of a jamming attack at time step k;
  • B is the control input matrix;
  • K is the feedback gain matrix.
The matrix A + γ k B K thereby encapsulates the controlled system response, dynamically adjusting to reflect the impact of jamming on the control signals.
In this paper, we formulate a two-player zero-sum stochastic game for NCSs subjected to jamming attacks. The primary objective of our work is to design an optimal defense mechanism for the sensor transmitter, which operates within the constraints of limited network resources.

3.2. Formulation of the Zero-Sum Stochastic Game in NCSs

To this end, we formulate a zero-sum stochastic game [22] framework that captures the essence of this adversarial interaction, where the gains of one player are exactly balanced by the losses of the other. This section describes the foundation elements of the game’s formulation, including the game’s players, state space, actions, payoffs, and the stochastic nature of the game dynamics.

3.2.1. Game Players

The game involves two players with opposing objectives:
  • Defender (D): The sensor transmitter, whose goal is to send feedback measurements to the plant.
  • Attacker (A): The jammer, who aims to degrade the system’s performance by disrupting communications.

3.2.2. State Space

The state space of the game, denoted by S, comprises all possible conditions of the NCS that can be affected by both the defender’s actions and the presence of jamming. The state at any time k is represented as x k = [ x 1 1 , x 2 2 , . . , x k i . . x k m ] , which represent system performance metrics constituting the state variables of the NCS and the operational status variable γ k , which indicates the state of the wireless network. The set of γ k is { 0 , 1 } .
In dynamic programming, discretizing [23,24] the state space based on given ranges for state variables is a fundamental step. Dynamic programming algorithms typically require looping over all possible states and actions to compute the optimal policy. When state and action spaces are continuous, it becomes computationally infeasible to evaluate every possible state or action. Discretizing these spaces converts them into a finite set, making the computations manageable. Therefore, in this paper, each state variable x k i is considered to be within the range [ l ̲ , l ¯ ) , and every range is divided into n number of discrete intervals, represented by discrete numbers.

3.2.3. Actions

Both players have a set of available actions depending on their objective. The defender can choose from a variety of control strategies or countermeasures designed to mitigate the impact of jamming and maintain the NCS’s stability. The attacker selects from strategies designed to maximize disruption and degrade the dynamic performance of the NCS.
In our case, the actions for D and A at time slot k are power levels to transmit signals denoted by a k A and a k D . Each player has two actions, a k A = { N A , A } and a k D = { N D , D } , which means either Attack (A), Defend (D), or do nothing (NA, ND) at any time slot k.

3.2.4. Transition Probabilities

The game’s dynamics are governed by stochastic elements, primarily the uncertainty in the attacker’s actions and the probabilistic effects of those actions on the state transitions. The transition probabilities are defined by the function P ( x k + 1 | x k , a D , a A ) . These probabilities encapsulate the likelihood of moving from one state to another, given the current state and the actions taken by both players. The Monte Carlo method [25,26] is very useful for estimating the probabilities of transitioning from one state to another in a stochastic game. Given the current state and the actions taken by both players, first ensure that the model correctly reflects the dynamics of the stochastic game, second, run multiple simulations of the game to estimate probabilities, and finally, refine the results by increasing the number of simulations.
The Monte Carlo method is very powerful, especially in game theory, where exact calculations may be cumbersome or require an impractical amount of computational resources.

3.2.5. Cost Function and Payoff

The payoffs in this zero-sum game are primarily defined through a cost function, C ( x k , a D , a A ) that quantifies the immediate cost to the defender based on the current state and the actions of both players. This cost function reflects not only the direct impact of jamming on system performance but also the operational costs associated with implementing various defensive strategies. For the attacker, the payoff is the negative of the defender’s cost, emphasizing the zero-sum nature of the game [27].
For each state and potential jamming scenario, the cost function is defined as
C ( x k , a D , a A ) = c 1 + c 2 f ( x k )
C ( x k , a D , a A ) = c 1 + c 2 ( x k T Q x k )
where the following hold:
  • c 1 and c 2 are cost coefficients that weight the importance of different aspects of the state impact;
  • f ( x k ) represents the quadratic state error that models the degraded dynamic performance of the NCS;
  • Q is a weighting matrix that scales the quadratic cost associated with the state vector.
This cost function provides a basis for evaluating the efficacy of different defensive strategies in terms of minimizing expected costs arising from state transitions influenced by jamming.

3.3. Solving the Game

The strategic solution to this game involves finding a Nash equilibrium through minimax strategies [28]. The defender seeks to minimize the maximum possible expected loss, considering the worst-case actions of the attacker. Conversely, the attacker aims to maximize the minimum expected gain, exploiting the most vulnerable aspects of the system. Solving this game typically involves calculating value functions and employing strategies such as dynamic programming or linear programming to optimize the expected payoffs over a horizon. At each time step and for each state, a linear programming problem is formulated to minimize the expected cost, given the current state and the probability of attack scenarios [29].
In our analysis, the strategies deployed by both players are state-dependent and involve mixed strategies. Mixed strategies are critical in ensuring that each stage of the game reaches a saddle-point equilibrium, a concept crucial for stability in game theory.
Let s S denote a specific state in the state space. Each player selects actions from the set of possible actions { a D } for the defender and { a A } for the attacker.
  • The defender’s mixed strategy at state s, denoted as p D ( s ) , is a probability distribution over the defender’s actions set. This strategy is represented by
    p D ( s ) : = [ p 1 D ( s ) , p 2 D ( s ) ]
    where p 1 D ( s ) and p 2 D ( s ) are the probabilities of choosing the first and second actions, respectively, at state s. These probabilities satisfy the following normalization condition:
    i = 1 2 p i D ( s ) = 1 .
  • Similarly, the attacker’s mixed strategy at state s, denoted as q A ( s ) , is also a probability distribution over the attacker’s actions set. This strategy is represented by
    q D ( s ) : = [ q 1 A ( s ) , q 2 A ( s ) ]
    where q 1 A ( s ) and q 2 A ( s ) are the probabilities of choosing the first and second actions, respectively, at state s. These probabilities must also adhere to the following normalization condition:
    i = 1 2 q i A ( s ) = 1 .
Given the zero-sum nature of the game, the sum of outcomes for the defender and the attacker for any action pair at any state is zero. This means that the gains of one player are exactly offset by the losses of the other. Consequently, in zero-sum games, it suffices to consider only one set of game tables to understand the dynamics and outcomes, as the matrices for the opposing player can be directly inferred.
In this study, the defender aims to minimize his expected total cost, while the attacker wants to maximize his expected payoff. We give details on how to solve the game only from the defender’s point of view; thus, the defender game table has the structure presented in Table 2.
In this paper, we are considering an N-stage stochastic game [30], where outcomes depend not only on the current actions of the players but also on the state of the system, which itself changes in response to these actions and random attack effects. The payoff for the defender at each stage depends on the current state, the actions chosen by both players and the resultant next state. The total payoff to the defender is typically the sum of the payoffs over all stages, discounted by a factor that accounts for the strategic preferences.
The one-stage value of the game is derived using the defender’s mixed strategy p D ( s ) and the game table in Table 2 as follows:
c ( s ) = p 1 D ( s ) q 1 A ( s ) ( c 1 + c 2 x k T Q x k ) + p 1 D ( s ) q 2 A ( s ) ( c 1 + c 2 x k T Q x k ) + p 2 D ( s ) q 1 A ( s ) ( c 1 + c 2 x k T Q x k ) + 0 .
The expected total value of this game over all possible actions at time instant k, considering the one stage value c i ( s ) , is
V k ( s ) = E i = k N α c i ( s )
At any given stage k, we define the function Q k ( a , d , s ) represented by Equation (10) as the expected discounted cost when the attacker takes action a and the defender takes action d, when the current state is s.
The function Q k ( a , d , s ) represents a Bellman equation [31], which is frequently used in stochastic games. It involves determining the optimal policy by maximizing (or minimizing) the expected sum of current rewards (or costs) plus the discounted future rewards (or costs).
Q k ( a , d , s ) = c k ( a , d , s ) + α s S p ( s | s , a , d ) V k + 1 ( s )
where c k ( a , d , s ) is the immediate reward at the kth stage, V k + 1 ( s ) is the total expected cost from the next state with states s , p ( s | s , a , d ) are the transition probabilities described in Section 3.2.4, and α is the discount factor.
Since the defender’s objective is to minimize their expected total cost in opposition to the attacker’s objective, the minimax principle is used to determine optimal strategies for the defender; hence, the value of the stochastic game for the defender can be described as
V k ( s ) = min p D ( s ) max q A ( s ) q A ( s ) T Q k ( a , d , s ) p D ( s )
To solve the problem in Equation (11), we define a variable z = m a x i [ Q k ( a , d , s ) ] , and the problem comes to solve the linear programming problem described as follows:
min p D ( s ) z s . t . Q k ( a , d , s ) p D ( s ) i z , p D ( s ) 0 , 1 T p D ( s ) = 1 .
The linear programming problem for the attacker is as follows:
max q A ( s ) y s . t . Q k ( a , d , s ) q A ( s ) i y , q A ( s ) 0 , 1 T q A ( s ) = 1 .
where y = m i n i [ Q k ( a , d , s ) ]
The following lines resume all the steps to model and solve the zero-sum game for an attack–defense approach in NCSs in order to find optimal defense strategies.
  • Define the state space: Discretize the state space based on the given ranges for the state variables;
  • Define the action space: Define actions for the defender and the jammer;
  • Define the system dynamics: Use the provided state space model to compute the transition probabilities;
  • Define the cost function: Use the provided cost function to calculate the immediate costs;
  • Perform value function iteration: iterate to convergence to find the value function;
  • Use linear programming: Formulate the LP problem to find the optimal policy for the Defender/Attacker;
  • Solve the LP problem: Using MATLAB’s optimization toolbox to solve the LP problem [32].

4. Simulations and Discussion

This section provides a practical example to show the effectiveness of the proposed optimal scheme and evaluate the dynamic performance of the NCS under jamming attacks. We consider a linear system for a robotic arm as the plant of the NCS.
The system parameters and configurations for this NCS are presented as follows:
A = 0.75 0 1 0 ,   B = 9.35 0 ,   K = 0.9777 , 0.9533
where the matrices A and B define the fundamental dynamics of the system.
The matrix A is designed to represent the natural dynamics of the robotic arm. The chosen values reflect the system’s physical properties described by the following model:
m q ¨ ( t ) + c q ˙ ( t ) = τ ( t )
where q ( t ) and q ˙ ( t ) are the actuator motor position and velocity, respectively, m is the inertia ( m = 0.107 kg · m 2 ), and c is the Coriolis force parameter ( c = 0.08 Nms ). The matrix B is selected to capture the influence of control inputs on the robotic arm’s states. The values in B ensure that the control signals appropriately affect the system, allowing for effective state regulation. The parameters are chosen based on standard modeling practices for robotic arms, which dictates how control efforts translate into state changes. The feedback gain matrix K is calculated using the Linear Quadratic Regulator (LQR) method. This method is widely used in control theory to design controllers that minimize a cost function representing the trade-off between state deviations and control effort. The specific values in K are obtained by solving the Riccati equation for the given system dynamics. This ensures that the controller is optimized for performance in the absence of jamming attacks, providing a baseline for evaluating the proposed scheme’s effectiveness under attack conditions.
The coefficients c 1 = 0.85 and c 2 = 0.08 are used to compute the cost associated with different states and actions within the game, and the discount factor α is set at 0.9 .
The Q matrix representing the quadratic costs associated with the state variables, emphasizing the penalization of deviations from desired states, is chosen as an identity matrix of size two.
The state space of the system is discretized through x 1 r a n g e and x 2 r a n g e , which define the possible values of two state variables. x 1 r a n g e progresses from 0.1 to 1 in increments of 0.1, while x 2 r a n g e includes integer values from 1 to 10. This discretization creates a grid over which the game’s dynamics are computed, allowing for the approximation of continuous state spaces typically found in control problems. This setup is essential for applying numerical methods to solve the game, as it confines the potentially infinite state space to a manageable, finite set of states over which the value function and strategies can be optimized. Because of the discretization of states, a linear interpolation is necessary when system states are between values at the two nearest quantized states [32]. In this paper, we used linear interpolation using i n t e r p 2 , which is a two-dimensional interpolation of matrices in MATLAB. This function is used to estimate the values of the value function at nongrid points based on the known grid values.
In this study, two different matrices are used to calculate cost under different conditions to account for different scenarios related to the presence or absence of a jamming attack, indicated by the value of γ k . These matrices represent different cost structures that the defender faces depending on whether or not a jamming attack is active at any given time step k.
  • Matrix table when γ k = 0 : Represented by Table 3 and its corresponding Transition probabilities represented by P γ 0 .
    P γ 0 = 0.2 , 0.8 ; 0.3 , 0.7 ; 0.9 , 0.1 ; 0.5 , 0.5
  • Matrix table when γ k = 1 : Represented by Table 4 and its corresponding Transition probabilities represented by P γ 1 .
    P γ 1 = 0.1 , 0.9 ; 0.2 , 0.8 ; 0.8 , 0.2 ; 0.5 , 0.5
Table 3. Defender’s Game Table when γ k = 0 .
Table 3. Defender’s Game Table when γ k = 0 .
ANA
D 5 c 1 + c 2 x k T Q x k 2 c 1 + c 2 x k T Q x k
ND c 1 + c 2 x k T Q x k 0
Table 4. Defender’s Game Table when γ k = 1 .
Table 4. Defender’s Game Table when γ k = 1 .
ANA
D 3 c 1 + c 2 x k T Q x k c 1 + c 2 x k T Q x k
ND 2 c 1 + c 2 x k T Q x k 0
According to the proposed method for solving the stochastic game in the previous section, we can obtain the optimal defense and attack strategies for all the possible states.
Figure 2 highlights the dynamic nature of the defender’s strategy in response to when γ k = 1 , which means there is no attack. Initially, the defender committed to the strategy Defend, then the probability of defending gradually decreased from 1 to around 0.3, which suggests that over time steps, the defender starts to evaluate and shift their strategy in response to the absence of attack. Then, the probability of defense drops to zero, realizing that defending is no longer viable. γ k remains equal to 1.
Figure 3 highlights the opposite case when γ k = 0 , which means there is an attack, and we can clearly see that the defender increases their defense probability until it converges to 1.
As explained previously in this paper, the attack signal represented by γ k is chosen to be random, as in real-time scenarios. The effectiveness of our approach relies on the response of the defender strategy to the attack signal depicted in Figure 4.
The fluctuations suggest that the defender’s strategy effectiveness varies significantly with the randomness of γ k . These ups and downs indicate periods where the defense strategy successfully adapts to random changes in the value of the attack signal.
To better illustrate the effectiveness of our approach, Figure 5 shows the correlation between the defender’s strategy and the value of γ k . Comparing both plots, we observe that when γ k is 1, the defender’s probability of defense tends to be lower (absence of attack). Conversely, when γ k is 0, the probability of defense increases (presence of attack). Finally, to evaluate the effect of the proposed defense mechanism on the communication channel state, in terms of packet loss rate, we simulate our program 100 times in Matlab R2021b environment.
The results of these simulations are shown in Figure 6; the first plot indicates that without any defense mechanism, the network is highly volatile and possibly vulnerable to external disruptions.
With defense mechanisms in place, the network state is significantly more stable, indicating effective mitigation of attacks or disruptions as shown in the second plot of Figure 6.

5. Limitations and Future Research Directions

While the proposed stochastic game-based defense mechanism significantly improves the security and stability of networked control systems (NCSs) against jamming attacks, several limitations need to be acknowledged.
  • Computational Complexity: The implementation of the stochastic game framework requires significant computational resources. The iterative nature of the N-stage stochastic games and the need for real-time updates can lead to high computational costs, particularly for large-scale systems with numerous states and actions.
  • Scalability Issues: The scalability of the proposed method is a concern, especially for very large NCSs with complex topologies. As the number of states and control actions increases, the complexity of solving the stochastic game also increases, potentially making it impractical for very large systems.
  • Multiple-Attacker Scenarios: The current approach primarily focuses on single-attacker scenarios. In real-world applications, NCSs may face coordinated attacks from multiple adversaries. Extending the framework to handle multiple-attacker scenarios is a necessary step for comprehensive security.
  • Impact on Network Performance: While the defense mechanism improves security, it may also impact the overall performance of the network. The balance between maintaining security and ensuring optimal network performance needs further exploration to minimize any adverse effects
The proposed stochastic game approach provides a robust defense mechanism against jamming attacks in NCSs, offering significant improvements in system stability and resilience. However, addressing the aforementioned limitations is crucial for enhancing the applicability and effectiveness of the method in real-world scenarios. Future research should focus on optimizing computational efficiency, improving modeling accuracy, and extending the framework to handle more complex attack scenarios.

6. Conclusions

This research has systematically explored the application of stochastic game theory to the security of networked control systems (NCSs) under conditions of potential jamming attacks. By modeling the interactions between defenders and attackers within a framework of N-stage stochastic games, we derived strategic insights that not only enhance the theoretical understanding of security dynamics but also provide practical guidance for the design and operation of more resilient control systems.
Our analysis demonstrates that the proposed defense mechanism significantly enhances the stability and security of NCS. The experimental results, illustrated in the figure, clearly show the following:
  • State Stability Improvement: Without defense, the network state fluctuates significantly, indicating high instability. The state fluctuates between 0 and 1 frequently, demonstrating a lack of resilience against jamming attacks, while with defense, the network state remains consistently stable with minimal fluctuations. This indicates that the defense mechanism effectively mitigates the impact of jamming attacks.
  • Percentage of Time Steps in Stable State: Without defense, the network state remains within a stable range (0.1 to 0.9) approximately 20 % of the time, while with defense, the network state remains within a stable range approximately 95 % of the time, showing a significant improvement in stability.
These results indicate that the proposed defense mechanism improves the network’s stability by 75 percentage points, making it substantially more resilient to jamming attacks.
The methodologies and strategies developed can be directly applied to the design of security protocols for real-world NCSs. By integrating these adaptive strategies into their operational frameworks, systems can better anticipate and react to cyber threats in a manner that minimizes disruption and damage.

Author Contributions

Conceptualization, H.M. and T.E.; methodology, H.M.; software, H.M. and T.E.; validation, H.M. and T.E.; formal analysis, H.M.; investigation, H.M. and T.E.; resources, H.M. and T.E.; data curation, H.M.; writing—original draft preparation, H.M.; writing—review and editing, H.M. and T.E.; visualization, H.M. and T.E.; supervision, H.M. and T.E.; project administration, H.M. and T.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bemporad, A.; Heemels, M.; Johansson, M. Networked Control Systems; Springer: London, UK, 2010; Volune 406. [Google Scholar]
  2. Wang, F.Y.; Liu, D. Networked Control Systems; Springer: London, UK, 2008; pp. 153–196. [Google Scholar]
  3. Sandberg, H.; Gupta, V.; Johansson, K.H. Secure networked control systems. Annu. Rev. Control Robot. Auton. Syst. 2022, 5, 445–464. [Google Scholar] [CrossRef]
  4. Marsili, M.; Zhang, Y.C. Stochastic dynamics in game theory. arXiv 1998, arXiv:9801309. [Google Scholar]
  5. Neyman, A.; Sorin, S. (Eds.) Stochastic Games and Applications; Springer Science and Business Media: Berlin/Heidelberg, Germany, 2003; Volume 570. [Google Scholar]
  6. Ross, S.M. Introduction to Stochastic Dynamic Programming; Academic Press: Cambridge, MA, USA, 2014. [Google Scholar]
  7. Bertsekas, D. Abstract Dynamic Programming; Athena Scientific: Mumbai, India, 2022. [Google Scholar]
  8. Wang, X.; Sun, Y.; Ding, D. Adaptive dynamic programming for networked control systems under communication constraints: A survey of trends and techniques. Int. J. Netw. Dyn. Intell. 2022, 2022, 85–98. [Google Scholar] [CrossRef]
  9. Zhu, Q.; Basar, T. Game-theoretic methods for robustness, security, and resilience of cyberphysical control systems: Games-in-games principle for optimal cross-layer resilient control systems. IEEE Control Syst. Mag. 2015, 35, 46–65. [Google Scholar]
  10. Nguyen, A.T.; Teixeira, A.M.; Medvedev, A. A single-adversary-single-detector zero-sum game in networked control systems. IFAC-PapersOnLine 2022, 55, 49–54. [Google Scholar] [CrossRef]
  11. Malik, S.U.; Anjum, A.; Moqurrab, S.A.; Srivastava, G. Towards enhanced threat modelling and analysis using a Markov Decision Process. Comput. Commun. 2022, 194, 282–291. [Google Scholar] [CrossRef]
  12. Adlamani, S.; Eksioglu, B.; Medal, H.; Nandi, A. Jamming attacks on wireless networks: A taxonomic survey. Int. J. Prod. Econ. 2016, 172, 76–94. [Google Scholar] [CrossRef]
  13. Yu, Y.; Liu, G.P.; Huang, Y.; Hu, W. Optimal Resilient Tracking Control for Networked Systems with Multi-Channels under Attacks. IEEE Trans. Ind. Electron. 2023, 71, 3001–3011. [Google Scholar] [CrossRef]
  14. Liu, S.; Liu, P.X.; El Saddik, A. A stochastic game approach to the security issue of networked control systems under jamming attacks. J. Frankl. Inst. 2014, 351, 4570–4583. [Google Scholar] [CrossRef]
  15. Wang, Y.; Xing, W.; Zhang, J.; Liu, L.; Zhao, X. Remote State Estimation under DoS Attacks in CPSs with Arbitrary Tree Topology: A Bayesian Stackelberg Game Approach. IEEE Trans. Signal Inf. Process. Over Netw. 2024, 10, 527–538. [Google Scholar] [CrossRef]
  16. Zhou, J.; Shang, J.; Chen, T. Cybersecurity Landscape on Remote State Estimation: A Comprehensive Review. IEEE/CAA J. Autom. Sin. 2024, 11, 851–865. [Google Scholar] [CrossRef]
  17. Ni, Y.; Ren, X.; Dey, S.; Shi, L. Remote state estimation with a strategic sensor using a Stackelberg game framework. IEEE Trans. Control. Netw. Syst. 2021, 8, 1613–1623. [Google Scholar] [CrossRef]
  18. Wu, Y.; Li, Y.; Shi, L. A game-theoretic approach to remote state estimation in presence of a dos attacker. IFAC-PapersOnLine 2017, 50, 2595–2600. [Google Scholar] [CrossRef]
  19. Yang, G.; Sun, X.; Liu, L.; Li, X. Cyber Attacks on Remote State Estimation in Cyber-Physical Systems: A Game-Theoretic Approach. In Advances in Applied Nonlinear Dynamics, Vibration and Control-2021: The Proceedings of 2021 International Conference on Applied Nonlinear Dynamics, Vibration and Control (ICANDVC2021) 1; Springer: Singapore, 2022; pp. 29–42. [Google Scholar]
  20. Park, P.; Araújo, J.; Johansson, K.H. Wireless networked control system co-design. In Proceedings of the 2011 International Conference on Networking, Sensing and Control 2011, Delft, The Netherlands, 11–13 April 2011; pp. 486–491. [Google Scholar]
  21. Yang, H.; Shi, M.; Xia, Y.; Zhang, P. Security research on wireless networked control systems subject to jamming attacks. IEEE Trans. Cybern. 2018, 49, 2022–2031. [Google Scholar] [CrossRef]
  22. Renault, J. A tutorial on zero-sum stochastic games. arXiv 2019, arXiv:1905.06577. [Google Scholar]
  23. Tang, M.; Ji, B.; Fang, X.; Yu, S.S. Discretization-strategy-based solution for berth allocation and quay crane assignment problem. J. Mar. Sci. Eng. 2022, 10, 495. [Google Scholar] [CrossRef]
  24. Almubarak, H.; Stachowicz, K.; Sadegh, N.; Theodorou, E.A. Safety embedded differential dynamic programming using discrete barrier states. IEEE Robot. Autom. Lett. 2022, 7, 2755–2762. [Google Scholar] [CrossRef]
  25. Barbu, A.; Zhu, S.C. Monte Carlo Methods; Springer: Singapore, 2020; Volume 35, p. 36. [Google Scholar]
  26. Younesi, A.; Shayeghi, H.; Safari, A.; Siano, P. Assessing the resilience of multi microgrid based widespread power systems against natural disasters using Monte Carlo Simulation. Energy 2020, 207, 118220. [Google Scholar] [CrossRef]
  27. Hattori, M.; Satoh, A.; Tanaka, Y. Minimax theorem and Nash equilibrium of symmetric multi-players zero-sum game with two strategic variables. arXiv 2018, arXiv:1806.07203. [Google Scholar]
  28. Sadi, Y.; Ergen, S.C.; Park, P. Minimum energy data transmission for wireless networked control systems. IEEE Trans. Wirel. Commun. 2014, 13, 2163–2175. [Google Scholar] [CrossRef]
  29. Liu, S.; Liu, P.X.; El Saddik, A. A stochastic security game for Kalman filtering in networked control systems (NCSs) under denial of service (DoS) attacks. Ifac Proc. Vol. 2013, 46, 106–111. [Google Scholar] [CrossRef]
  30. Kappen, H.J. Optimal Control Theory and the Linear Bellman Equation. 2011. Available online: https://www.ma.imperial.ac.uk/~nsjones/timeseriesbook.pdf (accessed on 3 March 2024).
  31. Ponce-Ortega, J.M.; Ochoa-Barragán, R.; Ramírez-Márquez, C. Optimization Using the Software MATLAB. In Optimization of Chemical Processes: A Sustainable Perspective; Springer Nature: Cham, Switzerland, 2024; pp. 439–463. [Google Scholar]
  32. Sullivan, P. Interpolating Value Functions in Discrete Choice Dynamic Programming Models. 2006. Available online: https://mpra.ub.uni-muenchen.de/864/ (accessed on 10 February 2024).
Figure 1. The networked control system with a jamming attacker.
Figure 1. The networked control system with a jamming attacker.
Mathematics 12 02143 g001
Figure 2. Probability of defending as a strategy when γ k = 0 .
Figure 2. Probability of defending as a strategy when γ k = 0 .
Mathematics 12 02143 g002
Figure 3. Probability of defending as a strategy when γ k = 1 .
Figure 3. Probability of defending as a strategy when γ k = 1 .
Mathematics 12 02143 g003
Figure 4. Defender’s strategy response to random γ k .
Figure 4. Defender’s strategy response to random γ k .
Mathematics 12 02143 g004
Figure 5. Defender’s strategy response to random γ k .
Figure 5. Defender’s strategy response to random γ k .
Mathematics 12 02143 g005
Figure 6. Network state without vs with the defense mechanism.
Figure 6. Network state without vs with the defense mechanism.
Mathematics 12 02143 g006
Table 1. Advantages and Disadvantages of Different Algorithms.
Table 1. Advantages and Disadvantages of Different Algorithms.
Algorithm TypeAdvantagesDisadvantages
Static Defense MechanismSimplicity, low computational costIneffectiveness against dynamic and adaptive attacks
Reactive ProtocolsQuick response to detected threatsDelayed detection, high false positive rate
Stochastic Game theoryAdaptive to changing threats, strategic decision makingHigh computational complexity, requires accurate modeling
Linear Programming OptimizationEfficient calculation of optimal strategies, scalabilitySensitive to model inaccuracies, initial setup complexity
Machine Learning ApproachesCan learn from evolving threats, improved accuracyRequires large datasets, high computational resources
Backward InductionProvides optimal strategy updates, adapts to changing scenariosComputationally intensive, requires precise state transition odeling
Table 2. Defender’s Game Table.
Table 2. Defender’s Game Table.
ANA
D c 1 + c 2 x k T Q x k c 1 + c 2 x k T Q x k
ND c 1 + c 2 x k T Q x k 0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mejdi, H.; Ezzedine, T. Novel Dynamic Defense Strategies in Networked Control Systems under Stochastic Jamming Attacks. Mathematics 2024, 12, 2143. https://doi.org/10.3390/math12132143

AMA Style

Mejdi H, Ezzedine T. Novel Dynamic Defense Strategies in Networked Control Systems under Stochastic Jamming Attacks. Mathematics. 2024; 12(13):2143. https://doi.org/10.3390/math12132143

Chicago/Turabian Style

Mejdi, Hana, and Tahar Ezzedine. 2024. "Novel Dynamic Defense Strategies in Networked Control Systems under Stochastic Jamming Attacks" Mathematics 12, no. 13: 2143. https://doi.org/10.3390/math12132143

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop