Next Article in Journal
Deciding on the Regularity of a Planar Coons Map
Previous Article in Journal
Characterization of the Appointment’s Reasons for “P—Psychological” on the ICPC-2 Scale in Primary Health Care Services
Previous Article in Special Issue
Optimizing Power Flow and Stability in Hybrid AC/DC Microgrids: AC, DC, and Combined Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Deceptive Tactics for Defense and Attack in Bayesian–Markov Stackelberg Security Games

by
Julio B. Clempner
Escuela Superior de Física y Matemáticas, Instituto Politécnico Nacional School of Physics and Mathematics, National Polytechnic Institute, Edificio 9 U.P. Adolfo Lopez Mateos, Col. San Pedro Zacatenco, Mexico City 07730, Mexico
Math. Comput. Appl. 2025, 30(2), 29; https://doi.org/10.3390/mca30020029
Submission received: 1 January 2025 / Revised: 23 February 2025 / Accepted: 10 March 2025 / Published: 17 March 2025
(This article belongs to the Special Issue Applied Optimization in Automatic Control and Systems Engineering)

Abstract

:
In this paper, we address the challenges posed by limited knowledge in security games by proposing a novel system grounded in Bayesian–Markov Stackelberg security games (SSGs). These SSGs involve multiple defenders and attackers and serve as a framework for managing incomplete information effectively. To tackle the complexity inherent in these games, we introduce an iterative proximal-gradient approach to compute the Bayesian Equilibrium, which captures the optimal strategies of both defenders and attackers. This method enables us to navigate the intricacies of the game dynamics, even when the specifics of the Markov games are unknown. Moreover, our research emphasizes the importance of Bayesian approaches in solving the reinforcement learning (RL) algorithm, particularly in addressing the exploration–exploitation trade-off. By leveraging Bayesian techniques, we aim to minimize the expected total discounted costs, thus optimizing decision-making in the security domain. In pursuit of effective security game implementation, we propose a novel random walk approach tailored to fulfill the requirements of the scenario. This innovative methodology enhances the adaptability and responsiveness of defenders and attackers, thereby improving overall security outcomes. To validate the efficacy of our proposed strategy, we provide a numerical example that demonstrates its benefits in practice. Through this example, we showcase how our approach can effectively address the challenges posed by limited knowledge, leading to more robust and efficient security solutions. Overall, our paper contributes to advancing the understanding and implementation of security strategies in scenarios characterized by incomplete information. By combining Bayesian and Markov Stackelberg games, reinforcement learning algorithms, and innovative random walk techniques, we offer a comprehensive framework for enhancing security measures in real-world applications.

1. Introduction

1.1. Brief Review

In safeguarding vital infrastructures, strategically allocating limited security resources is crucial. The primary objective is to prevent attacks from unpredictable assailants targeting critical assets. However, in selecting defense techniques, the balance between defenders’ private information disclosure and attackers’ observational capabilities must be carefully considered. This equilibrium ensures effective protection against potential threats while minimizing vulnerabilities arising from asymmetrical information.
Game theory offers a powerful mathematical framework for optimizing decision-making in strategic scenarios, which is particularly evident in SSGs tackling high-complexity problems. SSGs encapsulate situations where defenders (leaders) and attackers (followers) strategically engage in sequential decision-making to achieve their objectives. In this model, defenders are represented by the leaders, while attackers are represented by the followers, highlighting the strategic interplay between protective and offensive roles in the system. Multiplayer game theory presents significant complexity and challenges in finding solutions due to the interaction of multiple agents with varying strategies and objectives. The dynamic nature and interdependencies of players’ actions create a highly intricate environment, requiring sophisticated approaches to model, analyze, and achieve equilibrium in these games. Initially, defenders commit to a strategy, which attackers then observe before determining their own course of action [1,2]. Defenders employ random techniques to safeguard potential targets, anticipating attackers’ objectives within the game’s dynamics. Meanwhile, attackers adopt best-reply strategies while cognizant of these defensive tactics. Mathematically, the defenders’ advantage lies in revealing their random plans, yet this overlooks potential knowledge loss due to incomplete observations. For instance, when attackers disguise themselves as civilians to infiltrate a target, defenders may struggle to ascertain the available security resources accurately. In such scenarios, revealing complete information isn’t always advantageous for defenders. Instead, strategic disclosure of incomplete information can be more beneficial [3,4]. This approach acknowledges the complexity of real-world security dilemmas, where optimal decision-making hinges on dynamically balancing information disclosure and strategic concealment. By leveraging game theory’s mathematical rigor, SSGs provide a framework for navigating these strategic challenges, ultimately enhancing security decision-making in complex and dynamic environments.

1.2. Related Work

Various game-theoretical frameworks have been employed to model the interaction between defenders and attackers in security scenarios, with Bayesian Stackelberg games emerging as a particularly successful solution [5]. Conitzer and Sandholm [6] adapted the Bayesian Stackelberg game into a normal-form representation using Harsanyi’s approach [5], determining game equilibrium by evaluating each follower’s plan to ascertain if it constitutes an optimal reaction. Paruchuri et al. [7] proposed a mixed-integer linear programming (MILP) approach to solve Bayesian Stackelberg games, leveraging optimization techniques to identify optimal strategies for both defenders and attackers. Jain et al. [8] introduced a strategy combining hierarchical decomposition with branch and bound techniques, aiming to efficiently explore the solution space and achieve equilibrium in the game. Yin and Tambe [9] suggested a hybrid technique merging best-first search with heuristic branching rules to solve Bayesian Stackelberg games effectively. This approach enhances the computational efficiency of finding equilibrium strategies by guiding the search process through the solution space. Overall, these methodologies contribute to advancing our understanding of security game dynamics and provide practical solutions for effectively addressing security challenges in real-world scenarios, highlighting the versatility and adaptability of game-theoretical approaches in security analysis and decision-making.
For further exploration, information can be found in Wilczynski’s comprehensive survey [10]. Sayed Ahmed [11] introduced a deception-based Stackelberg game anti-jamming mechanism. Gan et al. delved into security games involving uncoordinated defenders cooperating to protect targets, with each defender optimizing their resource allocation selfishly for utility maximization [12]. These studies underscore the breadth of applications and innovative approaches within the realm of SSGs, showcasing their relevance in modern security paradigms.
SSGs find diverse applications in everyday scenarios, contributing significantly to the enhancement of security measures across various domains. Wilczynski [10] provided a comprehensive survey highlighting the breadth of SSG applications. Basilico et al. [13] determined the minimum number of robots required to patrol a given environment, compute optimal patrolling strategies across various coordination dimensions, and experimentally evaluate the proposed techniques, demonstrating their effectiveness in ensuring comprehensive coverage and the efficient use of robotic resources. Clempner [14] proposed a security model implemented with a temporal difference method, and it incorporates prior information to effectively address security issues. By leveraging continuous learning and adapting to evolving threats, this model aims to enhance the robustness and resilience of security systems. Albarran et al. [15] applied SSGs to distribute security resources across airport terminals, leveraging partially observed Markov game settings to address complex security challenges. In urban settings, Trejo et al. [16] employed reinforcement learning within SSG frameworks to adapt attacker and defender strategies, improving security measures at geographically dispersed shopping malls. Solis et al. [17] extended SSG applications to maritime security, utilizing a multiplayer ship differential SSG to model pursuit–evasion scenarios in continuous time. Alcantara et al. [18] incorporated topographical information into SSGs to generate realistic patrol plans for defenders in major cities, enhancing target identification and security deployment strategies. Sayed Ahmed [11] proposed a deception-based SSG anti-jamming solution, addressing vulnerabilities in communication systems. Gan et al. [12] explored SSGs where uncoordinated defenders collaborate to protect targets, with each defender optimizing their resource allocation selfishly. Clempner and Poznyak [19] introduced an attacker–defender SSG system, leveraging ergodic Markov models and non-decreasing Lyapunov-like functions to represent solutions. Rahman and Oh [20] investigated online patrolling robot route-planning tasks using SSG frameworks, enhancing security surveillance capabilities. Wang et al. [21] developed the DeDOL method based on deep reinforcement learning within SSG models, offering advanced solutions for security optimization. Li et al. [22] proposed a Bayesian Stackelberg Markov game for adversarial federated learning, utilizing meta-RL-based pre-training and adaptation to combat diverse attacks, achieving robust, adaptive, and efficient federated learning defense strategies. Sengupta and Kambhampati [23] framed adversarial federated learning as a Bayesian Stackelberg Markov game, using RL-based pre-training and meta-RL defenses to combat adaptive attacks, achieving robust and efficient federated learning defense strategies. Shukla et al. [24] presented a cybersecurity game for networked control systems, where an attacker disrupts communication and a defender protects key nodes. A cost-based Stackelberg equilibrium and robust defense method optimize security. Genetic algorithms enhance strategies for large power systems.
The versatility and effectiveness of SSGs highlight their importance in enhancing security strategies and protecting critical assets across various contexts. Whether applied to scheduling security patrols at airports, optimizing resource allocation in urban settings, or addressing vulnerabilities in communication systems, SSGs offer valuable insights and solutions. By integrating game-theoretical frameworks with real-world security challenges, SSGs enable proactive and adaptive approaches to security management, ultimately contributing to safer environments and mitigating potential threats effectively.

1.3. Main Results

Improving the existing security framework involves addressing the challenges posed by incomplete information. This is achieved through several key approaches. Firstly, an iterative proximal-gradient approach is proposed to compute the Bayesian Stackelberg equilibrium, enabling effective decision-making despite uncertainty. Additionally, a Bayesian reinforcement learning (BRL) algorithm is introduced, allowing for adaptive strategies that learn from past experiences and observations. Furthermore, a random walk approach is developed to implement SSGs, offering a dynamic and flexible method for responding to evolving threats. By integrating these techniques, the security framework gains the ability to adapt and respond effectively in the face of incomplete information, thereby enhancing overall resilience and effectiveness in mitigating security risks. These methods enable decision-makers to make more informed choices and optimize security strategies in complex and uncertain environments. A numerical example is used to provide useful recommendations for defenders’ resource allocation against attackers.

1.4. Organization of the Paper

The paper follows a structured layout, with Section 2 providing preliminary information. In this section, foundational concepts and background knowledge relevant to the study are presented to establish a basis for subsequent discussions. Section 3 elaborates on the problems under consideration and introduces a Bayesian Stackelberg game model, which serves as the theoretical framework for analyzing strategic interactions between defenders and attackers. This section outlines the key components of the model and discusses its applicability to the specific security scenario under investigation. In Section 4, the paper proposes a learning algorithm tailored to the context of the Bayesian Stackelberg game, aiming to optimize decision-making strategies for both defenders and attackers. The algorithm is designed to adapt and improve over time based on feedback and observed outcomes. Section 5 introduces a novel random walk model, offering an alternative approach to addressing security challenges within the Bayesian Stackelberg game framework. This model presents a new perspective on strategic decision-making in dynamic security environments. A numerical example illustrating the application of the proposed methodologies is provided in Section 6, demonstrating their effectiveness and practical relevance in real-world scenarios. Finally, Section 7 concludes the paper by summarizing key findings, drawing conclusions based on the analysis conducted, and discussing potential avenues for future research and development in the field of SSGs.

2. Preliminaries

In a discrete-time and finite-horizon framework, we consider an environment characterized by private and independent values, as described in prior works [3,25,26]. In this setup, players denoted by l L , where L = 1 , 2 , , n , receive rewards in discrete time periods t T , with T N . These rewards are contingent upon the current physical allocation a t l A l and the type Θ l Θ l of the player. Here, A l represents the feasible set of allocations available to player l at time t, while Θ l denotes the set of possible types for player l. The reward obtained by player l at time t is determined by their chosen allocation and their specific type. This framework captures the dynamics of decision-making in scenarios where players have private information and their actions impact the outcomes over a finite time horizon. Such models find applications in various domains, including economics, game theory, and decision science, where understanding the interplay between private information and strategic decision-making is crucial.
In period t, the type vector is denoted by θ t = ( θ t 1 , , θ t n ) Θ , where Θ = × l L Θ l . The set of feasible allocations in period t may be contingent on the vector of past allocations a t = ( a t 1 , , a t n ) A , where A = × l L A l . We denote Θ l as × i L , i l Θ i , and Δ ( A ) as the set of all probability distributions over A. Additionally, we assume that A l and Θ l are finite sets, ensuring a manageable and well-defined problem space. This setup enables the modeling of dynamic interactions and decision-making processes among multiple players with diverse types and feasible allocation options.
Each player l has a common prior distribution P l ( θ 0 l ) . The type Θ l and action a t l determine a probability distribution for the variables θ t + 1 l on Θ l , which is denoted by p l ( θ t + 1 l | θ t l a t l ) . This distribution captures the probabilistic relationship between a player’s current type and action and their subsequent type in the next period, facilitating the modeling of dynamic decision-making processes within the game.
A Markov chain is described by the transition matrix p l θ t + 1 l | θ t l a t l and the common prior distribution P l ( θ 0 l ) where P l Θ l Δ Θ l and Δ Θ l establishes the probability distributions over Θ l . We assume that each chain ( P l , p ( θ t + 1 l | θ t l , a t l ) ) is ergodic, ensuring that the system will eventually reach a steady state regardless of its initial conditions.
The asymmetric information is determined by the private observation of θ t l . We determine that by the priors distributions P l ( θ 0 l ) and the transition matrices p l ( θ t + 1 l | θ t l a t l ) for each player l, the information of player l, given by θ t + 1 l , does not depend on θ t i , for i l . Following that, participants transmitted m t l messages at the same time, and the message profile was made public m t l .
A (behavioral) strategy σ l ( m t l | θ t l ) for player l is mapping σ l : M l × Θ l Δ ( M l ) , which represents the likelihood with which a player l believes that a message m t l is of type Θ l . The (behavioral) strategy set is given by
S a d m l = σ l ( m t l | θ t l ) 0 m t l Θ l σ l ( m t l | θ t l ) = 1 , θ t l Θ l .
A strategy is a sequence π l ( a t l | m t l ) such that, for each period t, π l ( a t l | m t l ) is a stochastic kernel on A given history H n . The set of all the admissible strategies is
Π a d m l = π l ( a t l | m t l ) 0 a t l A l π l ( a t l | m t l ) = 1 , m t l M l
Remark 1.
A strategy π is typically defined as a policy or plan of action for a player that depends only on the information they possess. In this case, π depends solely on the player’s own message, and this suggests that π is a strategy based on private information or the part of the information that is most relevant to the player. The assumption here might be that each player has access to their own private message but does not directly depend on the messages of other players in their strategy formulation. This can simplify the analysis in certain games. A strategy σ, on the other hand, often refers to a more general form of strategy that could depend on all available information, such as public messages or even the outcomes of communication between players. σ involves communication, and it might be labeled as a communication strategy, where players decide how to share or process information.
Let us introduce the cost function v l ( a t l θ t l m t l ) , which characterizes the losses incurred by players l when they take action a t l based on the message m t l generated under type θ t l .
The average cost [ 0 , T ] of the player l is
V T l ( π , σ ) = t T θ t l Θ l m t l M l a t l A l v l ( a t l θ t l m t l ) p l ( θ t + 1 l | θ t l a t l ) i L π i ( a t i | m t i ) σ i ( m t i | θ t i ) P i ( θ t i ) = t T θ t l Θ l m t l M l a t l A l W l ( a t l θ t l m t l ) i L π i ( a t i | m t i ) σ i ( m t i | θ t i ) P i ( θ t i ) .
where
W l ( a t l θ t l m t l ) = v l ( a t l θ t l m t l ) p l ( θ t + 1 l | θ t l a t l ) .
We assume that players know their payoffs. The strategy σ l ( m t l | θ t l ) minimizes the weighted cost function V l ( π , σ ) , realizing the rule, which is given by
( π , σ ) A r g min σ Π a d m min σ S a d m l L V l ( π , σ ) .
The policy π and the strategy σ is said to satisfy the Bayesian–Nash equilibrium if, for all π Π a d m and σ S a d m , the following condition holds:
V l ( π , σ ) V l ( π , σ l , σ l ) .
where σ l = σ 1 , , σ l 1 , σ l + 1 , , σ n .
Let us introduce the auxiliary ξ variable as follows:
ξ l ( θ t l m t l a t l ) : = π ( a t l | m t l ) σ l ( m t l | θ t l ) P l ( θ t l ) .
such that
Ξ a d m l : = ξ l ( θ t l m t l a t l ) m t l M l θ t l Θ l a t l A l ξ l ( θ t l m t l a t l ) = 1 , m t l M l a t l A l ξ l ( θ t l m t l a t l ) = P l ( θ t l ) > 0 , m t l M l a t l A l θ t l Θ l [ δ θ t l , θ t + 1 l p l ( θ t + 1 l | θ t l a t l ) ] ξ l ( θ t l m t l a t l ) = 0 , θ t + 1 l Θ l .
where δ θ t l , θ t + 1 l is Kroneker’s symbol. It should be noted that the following relations hold:
m t l M l σ ( m t | θ t ) = 1 , θ t l Θ l P l ( θ t l ) = 1 .
We obtain that ξ l Δ l , where
Δ l : = ξ l ( θ t l m t l a t l ) m t l M l θ t l Θ l a t l A l ξ l ( θ t l m t l a t l ) = 1 , m t l M l a t l A l ξ l ( θ t l m t l a t l ) = P l ( θ t l ) > 0 .

3. Security Game Model

In the game involving multiple players, we consider two distinct groups: the defenders, denoted by L = 1 , , l , and the attackers, denoted by F = 1 , , f . Utilizing conventional game-theoretic notation, we denote the player index set as I . The joint action profile of all participants, excluding agent q, is represented as x q = ( x h ) h I / q . In the context of our Stackelberg game, I = L F , where players are indexed by ι = 1 , l ¯ for the defenders and r = 1 , f ¯ for the attackers. This indexing scheme allows for clear differentiation between the defenders and attackers, facilitating the analysis of strategic interactions and decision-making dynamics within the game framework.
Let us say the defenders’ strategies are denoted by x ι X ι , where X is a convex and compact set and
x ι : = c o l ξ i k ι , X ι : = Ξ a d m ι , X : = ι = 1 l X ι
such that c o l is a column-converting operator for the matrix ξ i k ι . The joint strategy of the players is denoted by x = ( x 1 , , x n ) X , while the complement strategy of the players is denoted by x ^ = x ι . Here,
x ι : = x 1 , , x ι 1 , x ι + 1 , , x l X ι : = h = 1 , h ι l X ι
such that x = ( x ι , x ι ^ ) .
Similarly, if we let y r Y r r = 1 , f ¯ represent the attackers’ strategies and Y be a convex and compact set where
y r : = c o l ξ i k ) r , Y r : = Ξ a d m r r = 1 , f ¯ , Y : = r = 1 f y r
is denoted by y = ( y 1 , , y f ) Y : = r = 1 f Y r , then the joint strategy of the attackers and y r ^ = y r are strategies of the complement of the players adjoint to y r , i.e.,
y r : = y 1 , , y r 1 , y r + 1 , , y f Y r : = r = 1 , r f f Y r
such that y = ( y r , y r ^ ) .
In the scenario we investigate, defenders and attackers engage in a Nash game within the framework of a simultaneous play game that is restricted to a Stackelberg game. In simultaneous play games, the Nash equilibrium serves as the notion of equilibrium. Here, each player independently selects their strategy, aiming to maximize their own utility given the strategies chosen by other players. A Nash equilibrium is reached when no player can unilaterally deviate from their strategy to achieve a better outcome. In contrast, hierarchical play games, such as Stackelberg games, involve sequential decision-making, where one player (the defender) commits to a strategy first, and the other player (the attacker) observes this strategy before determining their own. The equilibrium concept in hierarchical play games differs from the Nash equilibrium, as it involves optimizing strategies in anticipation of the actions of other players in the sequence of play. Formalizing the Stackelberg game within this framework involves defining the roles of defenders and attackers, specifying their strategies, and analyzing the equilibrium outcomes under the sequential decision-making structure.
Let W = X Y . If a player q I has a cost function φ q w q , w q , then we have the following:
Definition 1.
The joint strategy w W is a Nash equilibrium if, for each q I ,
φ q w φ q w q , w q , w q W q .
In the game progression, defenders forecast attackers behavior by playing non-cooperatively, anticipating attackers’ actions at a Stackelberg equilibrium. This strategic anticipation enables defenders to make informed decisions when committing to their strategies, considering the likely responses of attackers within the framework of the Nash equilibrium. By strategically forecasting attackers’ behavior, defenders aim to optimize their own outcomes in the game. To achieve the game’s aim, defenders must first identify a strategy x = x 1 , , x l X that is satisfying for any admissible x ι X ι and any ι = 1 , l ¯ .
Λ x : = ι = 1 l min x ι X ι φ ι x ι , x ι φ ι x ι , x ι .
The cost function of the defenders ι who play the strategy x ι X ι and the complement of the defenders who play the strategy x ι X ι is φ ι x ι , x ι .
Taking the utopia point as
x ¯ ι : = arg min x ι X ι φ ι x ι , x ι
then one can describe Equation (6) in the following manner:
Λ x : = ι = 1 l φ ι x ¯ ι , x ι φ ι x ι , x ι .
The functions φ ι x ι , x ι ι = 1 , l ¯ are supposed to be convex in all their arguments.
The function Λ x fulfills Nash’s condition
max x X g x = ι = 1 l φ ι x ¯ ι , x ι φ ι x ι , x ι 0
for any x ι x ι and all ι = 1 , l ¯
A strategy x X is a Nash equilibrium if
x A r g min x X Λ x .
If Λ x is strictly convex, then
x = arg min x X Λ x .
In addition, in this process, the attackers attempt to reach some of the Nash equilibria and try to find a joint strategy y = y 1 , , y f Y that satisfies any admissible y r y r and any r = 1 , f ¯
Ψ y : = r = 1 f min y r Y r φ r y r , y r φ r y r , y r .
The cost function of the attackers m who play the strategy y r y r and the complement of the defenders who play the strategy y r y r is φ r y r , y r .
Bearing in mind the utopia point
y ¯ r : = arg min y r Y r φ r y r , y r
it is possible to rewrite Equation (6) as follows:
Ψ y : = r = 1 f φ r v ¯ r , y r φ r y r , y r .
The functions φ r y r , y r r = 1 , f ¯ are supposed to be convex in all their arguments.
The function Ψ y satisfies the Nash condition
max y r y r w y = r = 1 f φ r y ¯ r , y r φ r y r , y r 0
for any y r y r and all r = 1 , f ¯ .
The defenders’ goal is to find a solution to the optimization challenge given by the following definition:
Definition 2.
A Stackelberg game is a game with l defenders and f attackers; it is called a Stackelberg–Nash game if the defenders strive to solve the issue presented by
min x X { Λ x | y | y A r g min v Y Ψ v | x }
and the attackers try to solve the problem
min y Y Ψ y | x
such that
Λ x | y : = ι = 1 l φ ι x ¯ ι , x ι | y φ l x ι , x ι | y
and
Ψ y | x : = r = 1 f φ r y ¯ r , y r | x φ r y r , y r | x .
Then, the equilibrium notion in games is the Nash equilibrium in simultaneous play games and the Stackelberg equilibrium in hierarchical play games.
Definition 3.
(Stackelberg equilibrium) In a game with l defenders, the strategy x X is said to be a Stackelberg–Nash equilibrium strategy for the defenders if
max x ρ ( x ) Λ x | y max x ρ ( x ) Λ x | y
such that
ρ ( x ) = { y Y | Ψ y | x Ψ v | x v Y }
is the best reply set of the attackers.
The definition of Stackelberg equilibrium given above can be redefined for the set of the attackers when the set ρ ( x ) is substituted with the set of Nash equilibria, considering that the defenders play the strategy x and then the attackers’ best reply is a Nash equilibrium.
The general format iterative version n = 0 , 1 , of the proximal-gradient method for computing the Stackelberg equilibrium is as follows:
1. The first half-step (prediction):
v ^ n = arg min w ˜ X × Y 1 2 w ˜ v ˜ n 2 + γ Ψ ( w ˜ , v ˜ n )
2. The second (basic) half-step:
v ˜ n + 1 = arg min w ˜ X × Y 1 2 w ˜ v ˜ n 2 + γ Ψ ( w ˜ , v ^ n )

4. Learning

Reinforcement learning (RL) addresses the challenge of learning optimal actions in an unknown Markov setting through interaction [27]. Players aim to devise strategies that minimize their expected costs. In a discrete-time setting, at each time step t, players observe a cost v i , where i denotes the player. The overarching objective is to minimize the average cost V over time, guiding decision-making towards more favorable outcomes. By iteratively adjusting actions based on observed costs and environmental feedback, RL algorithms seek to discover optimal strategies that lead to the most advantageous long-term outcomes in dynamic and uncertain environments.
We consider RL problems where the underlying environment is a Bayesian–Markov decision process. Specifically, in the Stackelberg game, players indexed by ι = 1 , l ¯ are the defenders and players indexed by r = 1 , f ¯ are the attackers. We will develop the results in general for a player i I ( I = L F ) and we specify defenders and attackers when it will be necessary. We propose a setting based on experiences, which is calculated by adding the number χ of experiences given as follows: let h t H t ( t T ) be the history at time t and ( θ t , a t ) Θ × A . Let, for each ( θ t + 1 , θ t , a t ) Θ × Θ × A ,
χ i ( θ ˜ t + 1 i θ ˜ t i a t i ) = t T 1 ( θ ˜ t + 1 i θ ˜ t i a t i ) = E 1 ( θ ˜ t + 1 i ) 1 ( θ ˜ t i ) 1 ( a t i )
denote the experimentally observed absolute average number of transitions from type θ ^ t i applying action a t i , We obtain the normally distributed and asymptotically unbiased maximum likelihood estimator given by p i θ ˜ t + 1 i | θ ˜ t i a t i
p ˜ i θ ˜ t + 1 i | θ ˜ t i a t i = χ i ( θ ˜ t + 1 i θ ˜ t i a t i ) χ i ( θ ˜ t i a t i ) .
Therefore, the cost of player l L is
v ˜ i θ ˜ t + 1 i θ ˜ t i a t i = t T α i ( θ ˜ t + 1 i θ ˜ t i a t i ) χ i ( θ ˜ t + 1 i θ ˜ t i a t i ) χ i ( θ ˜ t + 1 i θ ˜ t i a t i )
where
α i ( θ ˜ t + 1 i θ ˜ t i a t i ) = v i ( θ ˜ t + 1 i θ ˜ t i a t i ) + α i r
and α i u i such that r randomly takes the values between 1 or 1.
We assume that the det P i ( θ t i | m t i ) 0 where Σ = P i ( θ t i | m t i ) 1 exists (the inverse of P i ( θ t i | m t i ) ). We suggest a framework from experiences that is calculated counting the number χ of unobserved experiences recursively as follows:
χ i ( m t i m t i a t i ) = t T 1 ( m t i | m t i a t i ) = E 1 ( m t i ) 1 ( m t i ) 1 ( a t i )
where χ i ( m t i a t i ) is the number of visits in the state m t i and χ i ( m t i m t i a t i ) is the total number of times that the system changes from m t i to m t i applying a t i . We obtain that χ i ( m t i a t i ) = m t i M t i χ i ( m t l m t i a t i ) .
The estimated transition matrix p ˜ i ( θ ˜ t + 1 | θ ˜ t x t ) is given by
p ˜ i θ ˜ t + 1 i | θ ˜ t i a t i = m t i M i m t i M i Σ i ( m t i θ ˜ t i ) Σ i ( m t i θ ˜ t + 1 i ) χ i ( m t i m t i a t i ) θ ^ t + 1 i Θ t i m t i M i m t i M i Σ i ( m t i θ ˜ t i ) Σ i ( m t i θ ˜ t + 1 i ) χ i ( m t i m t i a t i ) .
In order to recover the variables of interest, we calculate
σ ι ( m t ι | θ t ι ) = a t ι A ι ξ ι ( θ t ι m t ι a t ι ) m t ι M ι a t ι A ι ξ ι ( θ t ι m t i a t ι ) .
π ι ( a t ι | m t ι ) = 1 Θ ι θ t ι Θ ι ξ ι ( θ t ι m t ι a t ι ) a t ι A ι ξ ι ( θ t ι m t ι a t ι ) .
P ι ( θ t ι ) = m t ι M ι a t ι A ι ξ ι ( θ t ι m t ι a t ι ) > 0 .
σ r ( m t r | θ t r ) = a t r A r ξ r ( θ t r m t r a t r ) m t r M r a t r A r ξ r ( θ t r m t j a t r ) .
π r ( a t r | m t r ) = 1 Θ r θ t r Θ r ξ r ( θ t r m t r a t r ) a t r A r ξ r ( θ t r m t r a t r ) .
P j ( θ t r ) = m t r M r a t r A r ξ r ( θ t r m t j a t r ) > 0 .
The BRL algorithm for the SSG is outlined in Algorithm 1 and Algorithm 2. At time t = 0 , for each defender ι = 1 , , l and each attacker r = 1 , , f , we initialize the parameters. Specifically, the initial belief states for both the defender and the attackers are represented as θ ˜ 0 ι P ι ( θ 0 ι ) and θ ˜ 0 r P r ( θ 0 r ) , where P ι and P r are probability distributions over the possible states θ 0 of the defender and attackers, respectively. These beliefs are updated throughout the game as players take action and observe the outcomes.
Algorithm 1: BRL algorithm for the SSG
ι = 1 , l ¯ and r = 1 , f ¯ .
t = 1 .
θ ˜ 0 ι P ι ( θ 0 ι ) and θ ˜ 0 r P r ( θ 0 r ) .
p ˜ ι = p ι and p ˜ r = p r .
ε > 0 .
e t ι Error( p ^ t 1 ι , p ^ t ι ) and e t r Error( p ^ t 1 r , p ^ t r )
while ( ε > e t ι and ε > e t r )
(i) x min x X Λ x | y | y arg min α Y Ψ α | x and y min y Y Ψ y | x .
(ii) recover π Equations (14)–(17) and σ Equations (13)–(16)
(iii) m t ι σ ι ( m t ι | θ ˜ t ι ) and m t r σ r ( m t r | θ ˜ t r ) .
(iv) a t ι π ι ( a t ι | m t ι ) and a t r π r ( a t r | m t r ) .
(v) θ ˜ t + 1 ι p ^ ι ( θ t + 1 ι | θ t ι a t ι ) and θ ˜ t + 1 r p ^ r ( θ t + 1 r | θ t r a t r ) .
(vi) χ ι ( θ ˜ t + 1 i θ ˜ a t ι ) + = 1 , χ ι ( θ ˜ ι a t ι ) + = 1 and χ r ( θ ˜ t + 1 i θ ˜ a t r ) + = 1 , χ r ( θ ˜ r a t r ) + = 1 .
(vii) v ˜ ι θ ˜ t + 1 ι θ ˜ t ι a t ι Equation (11) and v ˜ r θ ˜ t + 1 r θ ˜ t r a t r Equation (11).
(viii) p ˜ ι θ ˜ t + 1 ι | θ ˜ t ι a t ι Equation (12) and p ˜ r θ ˜ t + 1 r | θ ˜ t r a t r Equation (12).
(ix) e t ι Error( p ^ t 1 ι , p ^ t ι ) and e t r Error( p ^ t 1 r , p ^ t r ).
(x) θ ˜ t ι = θ ˜ t + 1 ι , θ ˜ t r = θ ˜ t + 1 r p ι = p ^ ι , p r = p ^ r , v ι = v ˜ ι , v r = v ˜ r and t = t + 1 .
end while
Algorithm 2: Computation of the error
function Error ( p ^ t 1 , p ^ t )
e = a t A t r ( ( p ^ t 1 p ^ t ) ( p ^ t 1 p ^ t ) ) .
return e.
end function
The transition probabilities for each player’s states are defined as p ˜ ι = p ι ( θ ˜ 1 ι | θ ˜ 0 ι a 0 ι ) and p ˜ r = p r ( θ ˜ 1 r | θ ˜ 0 r a 0 r ) , where p ˜ ι and p ˜ r represent the likelihood of transitioning from one state to another given the actions a 0 ι and a 0 r taken by the defenders and the attackers. These transition probabilities are essential for predicting the future states of the game based on current strategies.
An important aspect of the algorithm is to account for estimation errors in the parameters of the model. We denote the allowable error in the estimated parameters by ε > 0 , which sets a threshold for how much deviation is acceptable. During each step of the algorithm, the actual error e t ι for the defenders and e t r for the attackers is computed to track the accuracy of the estimated parameters.
At each time step t, the BRL algorithm computes the optimal strategy σ ( m t l | Θ l ) and the corresponding policy π ( a t l | m t l ) for both the defenders and the attackers. The strategy σ determines the probability distribution over the possible messages m t l given the state Θ l , while the policy π dictates the probability of taking a particular action a t l given the message m t l . These strategies and policies are central to the Bayesian Stackelberg equilibrium, where players optimize their decisions based on their beliefs about the other players’ actions.
For each step, random messages m t ι and m t r are drawn from the respective optimal strategies σ ι ( m t ι | θ ˜ t ι ) and σ r ( m t r | θ ˜ t r ) . Based on these messages, actions are chosen randomly according to the policies π ι ( a t ι | m t ι ) for the defenders and π r ( a t r | m t r ) for the attackers.
Once the actions are selected and executed, the game transitions to the next state. The updated states θ ˜ t + 1 ι and θ ˜ t + 1 r are determined by the transition probabilities p ^ ι ( θ t + 1 ι | θ t ι a t ι ) and p ^ r ( θ t + 1 r | θ t r a t r ) , respectively.
After transitioning to the new states, we update the counts χ ι ( θ ˜ t + 1 ι , θ ˜ t ι , a t ι ) and χ r ( θ ˜ t + 1 r θ ˜ t r a t r ) , which track how frequently each state–action pair is encountered. These counts are critical for refining the players’ estimates of the value functions v ˜ ι ( θ ˜ t + 1 ι θ ˜ t ι a t ι ) and v ˜ r ( θ ˜ t + 1 r θ ˜ t r a t r ) . These value functions represent the expected reward for each player given their current and next states and the actions taken.
As the game progresses, the algorithm updates the transition probabilities p ˜ ι ( θ ˜ t + 1 ι | θ ˜ t ι a t ι ) and p ˜ r ( θ ˜ t + 1 r | θ ˜ t r a t r ) based on observed outcomes. At each step, the mean-square errors e t ι and e t r are calculated to measure the difference between the estimated and actual outcomes. The algorithm continues to iterate as long as these errors exceed the predefined threshold ε , ensuring that the estimates converge to an acceptable level of accuracy.
Once the errors e t ι and e t r fall below the threshold ε , the resulting strategy profile σ ( m t | θ t ) and policy π ( a t | m t ) form a Bayesian Stackelberg equilibrium. This equilibrium represents the optimal solution for the game, where the defenders and attackers have maximized their respective utilities given the available information.

5. Random Walk Algorithm

Random walks are basic models in probability theory with deep mathematical features and a wide range of applications in SSGs [16,28]. The long-term asymptotic behavior of the defenders to catch the attackers is a key question for these models.
The random walk process for the SSG is described by Algorithm 3 as follows. We will look at a game in which four players compete: defenders and attackers. Both sides have the option of using a randomized approach. The defenders and attackers can both move to any state at the same time. The defenders’ goal is to catch the attackers in the fewest possible steps. The attackers, on the other hand, choose a strategy that maximizes the amount of time it takes for the defenders to catch the attackers. If one of the defenders moves to a state where an attacker is, they are said to be caught. The game is over when the defenders have caught the attackers. We refer to a random walk as a discrete-time Markov process ( X t , t 0 ) in a finite-type space Θ . The random walk is assumed to be time-homogeneous. This means that the X t + 1 distribution given ( X 0 , , X t ) is solely dependent on X t and not on time t. To ensure that the random walk does not become ’stuck’ in some place of the state space, we also require some type of irreducibility: the set Θ is finite.
Given a type θ t , a message m t is chosen randomly by the defenders and the attackers from the behavior strategies m t ι σ ι ( m t ι | θ t ι ) and m t r σ r ( m t r | θ t r ) . They then choose actions a t ι π ι ( a t ι | m t ι ) and a r π r ( a t r | m t r ) . To minimize and maximize the chance of damage, the transition matrix p ι θ t + 1 ι | θ t ι a t ι and p r θ t + 1 r | θ t r a t r are used to select the next state in the process. The defenders and attackers aim to finish the SSG the next time ( t + 1 ). The process continues until the fulfillment of the SSG meets the game-over or capture statuses provided by the following:
ι L r F θ t + 1 ι Θ t ι θ t + 1 r Θ t r 1 ( θ t + 1 ι ) 1 ( θ t + 1 r )
Algorithm 3: Random walk for the SSG
1. Choose randomly an initial type for the defenders θ 0 ι P ι ( θ 0 ι ) and attackers θ 0 r P r ( θ 0 r ) .
2. Let π ι ( a t ι | m t ι ) and π r ( a t r | m t r ) be the resulting policies and, σ m t | θ t ι and σ m t | θ t r the resulting strategies.
3. For each attacker r and defender ι Do
4. Choose a type θ t ι σ m t | θ t ι and θ t r σ r | i ι
5. Choose actions a t ι π ι ( a t ι | m t ι ) and a r π r ( a t r | m t r ) .
6. From p ι θ t + 1 ι | θ t ι a t ι select θ t + 1 ι considering p ι · | θ t ι a t ι and from p r θ t + 1 r | θ t r a t r select θ t + 1 r considering p r · | θ t r a t r .
8. Update the original values by including θ t + 1 ι and θ t + 1 r to the random walk process.
9. Repeat steps 3–8 until the status expressed in Equation (19) is met.

6. Numerical Example

We will employ three players to simulate the SSG: one defender ( ι = 1 ) and two attackers ( r = 1 , 2 ). The primary objective of the defender is to minimize or halt the damage inflicted by the attackers, while the attackers aim to maximize the expected damage they can cause to a set of targets. This dynamic reflects a typical security scenario, where an agent (the defender) must safeguard resources or locations, while adversarial players (attackers) try to exploit weaknesses.
The game is structured in a sequential manner and is commonly modeled as a Stackelberg game. The defender acts as the leader, committing to a strategy first, which is observed by the attackers. Once the defender’s strategy is known, the attackers, acting as followers, select their strategies in response. This turn-based nature highlights the strategic advantage of the defender, as their actions guide the attackers’ responses.
Each player in the game has a set of possible states and actions. In this particular setup, we assume that each player has θ = 4 states, representing different possible situations or configurations in which they can find themselves during the game. Additionally, each player has a = 2 possible actions they can take in any given state, representing the limited yet strategic decisions available at each step of the game.
The key variables that drive the analysis of the SSG are the policies ( π ), behavior strategies ( σ ), and distribution vectors (P) for both the defender and attackers. The policy ( π ) defines the overall plan of action for each player, essentially specifying the probabilities with which they select each available action in a given state. Behavior strategies ( σ ) represent the specific actions taken by players as they progress through the game, reflecting how their strategies evolve over time. The distribution vectors (P) capture the likelihood of players being in different states throughout the course of the game.
These variables will be recovered analytically, meaning that through mathematical derivations and strategic analysis, we will determine the optimal policies, strategies, and distributions for both the defender and the attackers. This will allow us to predict the outcomes of the game under various scenarios, providing insight into how effective the defender’s strategy is in mitigating the attackers’ impact and how the attackers can best exploit potential vulnerabilities.
The generated policies are supplied by using the proposed approach as follows:
π 1 ( a 1 | m 1 ) = 0.3925 0.6075 0.3925 0.6075 0.3925 0.6075 0.3961 0.6039 π 2 ( a 2 | m 2 ) = 0.4748 0.5252 0.4748 0.5252 0.4748 0.5252 0.2492 0.7508 π 3 ( a 3 | m 3 ) = 0.7903 0.2097 0.7904 0.2096 0.7904 0.2096 0.6700 0.3300
The defender’s resultant (behavior) strategies are provided by
σ 1 ( m 1 | θ 1 ) = 0.3204 0.3205 0.3204 0.0387 0.0223 0.0223 0.0223 0.9332 0.2253 0.2253 0.2253 0.3240 0.2500 0.2500 0.2500 0.2500
and the (behavior) strategies of the attackers are
σ 2 ( m 2 | θ 2 ) = 0.0919 0.0919 0.0919 0.7243 0.0573 0.0573 0.0573 0.8280 0.2493 0.2495 0.2494 0.2518 0.2499 0.2501 0.2500 0.2500 σ 3 ( m 3 | θ 3 ) = 0.0287 0.0287 0.0287 0.9139 0.2501 0.2503 0.2501 0.2495 0.2491 0.2494 0.2498 0.2517 0.2498 0.2498 0.2500 0.2504
The distribution vectors are as follows:
P 1 ( θ 1 ) = 0.4657 0.4492 0.0448 0.0403 P 2 ( θ 2 ) = 0.1097 0.1752 0.5679 0.1472 P 3 ( θ 3 ) = 0.3492 0.2401 0.1065 0.3041
The algorithm ensures the convergence of the strategies shown in Figure 1, Figure 2 and Figure 3, optimizing decision-making and adapting efficiently to dynamic threats in real-time environments. Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9 demonstrate the convergence of the error in the reinforcement learning algorithm, illustrating its efficiency and effectiveness in optimizing strategies and adapting to dynamic environments over time. By integrating prior information and utilizing a continuous-time approach, the model enhances its robustness, ensuring that the security system remains resilient against evolving threats. These visualizations confirm the model’s capability to maintain optimal performance and adaptiveness.
The primary goal of patrol scheduling is to efficiently allocate security teams to safeguard fixed targets, taking into account limited workforce resources. To address this challenge, Algorithm 3 is utilized to illustrate the realization of an SSG, where target visitations are determined based on the resulting game strategy, denoted as π k | m . This algorithm aids in the planning and deployment of patrols to maximize the protection of critical assets.
Figure 10 provides a visual representation of an instance of the SSG, showcasing the outcomes of engagements between attackers and defenders. In this particular scenario, attacker 2 is apprehended at state 1 by defender 1 after 10 time steps, indicating the successful interception of the threat. Additionally, attacker 2 is captured at state 1, this time after 12 time steps by defender 1. These outcomes demonstrate the effectiveness of the patrol planning process in thwarting attacks and protecting the designated targets. By employing Algorithm 2 and visualizing the SSG outcomes in Figure 1, security planners can gain insights into the effectiveness of their patrol scheduling strategies. This approach allows for the optimization of patrol routes and deployment strategies, ultimately enhancing the security posture despite workforce limitations.
In the alternate realization depicted in Figure 11, a different sequence of events unfolds within the SSG. Here, attacker 2 faces a swift apprehension at state 1 by defender 1 after three time steps. Simultaneously, attacker 1 is intercepted at state 1, albeit after a longer duration of nine time steps, by defender 1. Despite the varying durations and defender involvement, both attackers are ultimately captured. The prompt resolution of attacker 2 by defender 1 underscores the importance of timely responses in security operations. This rapid intervention effectively neutralizes the threat before it can progress further. Meanwhile, the prolonged engagement leading to the capture of attacker 1 highlights the complexities and challenges inherent in security patrolling and target defense.
With the apprehension of both attackers, the game reaches its conclusion. The successful outcomes achieved by the defenders validate the effectiveness of the patrol scheduling strategies implemented. Such realizations provide valuable insights for refining future security tactics, emphasizing the importance of adaptability and resource allocation in mitigating security risks effectively.

7. Conclusions

This research introduces a novel mathematical framework that enhances security measures by integrating optimization techniques to mitigate existing risks while addressing real-time threats. The framework is based on a Bayesian–Markov Stackelberg game model, designed to capture adversarial interactions in security scenarios with incomplete information, where players have limited knowledge of each other’s strategies.
A key feature of this approach is the use of an optimization strategy grounded in the proximal-gradient method, which significantly improves computational efficiency compared to traditional Bayesian–Markov Stackelberg solutions. By leveraging this method, the framework effectively reduces security risks associated with persistent threats. Notably, it incorporates a unique reinforcement learning algorithm that derives rewards from a prior Bayesian distribution. This contribution is particularly significant, as it integrates structured historical data into decision-making, enabling more adaptive and informed responses to evolving security threats.
The framework’s practical effectiveness is demonstrated through numerical examples, showcasing its ability to utilize past information to guide present decisions. These results provide both theoretical insights and empirical evidence of its applicability in real-world security scenarios. Overall, this research presents a comprehensive security optimization approach that bridges the gap between mitigating existing risks and countering ongoing threats in real time. By combining advanced optimization techniques with an innovative reinforcement learning paradigm, this offers a promising avenue for strengthening security in complex, dynamic environments.
Looking ahead, several challenges remain. A key technical endeavor is the implementation of a game-theoretic approach using novel reinforcement learning methods, where capture conditions depend on states and incomplete information, enhancing the realism of patrolling schedules. Additionally, applying this methodology in fortification games and conducting controlled trials to evaluate player responses to game-theoretic scheduling in real-world settings represent significant challenges. These trials can yield valuable insights into the practical implications of game theory for strategic decision-making and scheduling in uncertain environments, ultimately ensuring more robust and adaptable security measures.

Funding

This research received no external funding.

Data Availability Statement

All data required for this article are included within this article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Trejo, K.K.; Clempner, J.B.; Poznyak, A.S. A Stackelberg security game with random strategies based on the extraproximal theoretic approach. Eng. Appl. Artif. Intell. 2015, 37, 145–153. [Google Scholar] [CrossRef]
  2. Solis, C.U.; Clempner, J.B.; Poznyak, A.S. Modeling Multi-Leader-Follower Non-Cooperative Stackelberg Games. Cybern. Syst. 2016, 47, 650–673. [Google Scholar] [CrossRef]
  3. Clempner, J.B.; Poznyak, A.S. A nucleus for Bayesian Partially Observable Markov Games: Joint observer and mechanism design. Eng. Appl. Artif. Intell. 2020, 95, 103876. [Google Scholar] [CrossRef]
  4. Clempner, J.B.; Poznyak, A.S. Optimization and Games for Controllable Markov Chains: Numerical Methods with Application to Finance and Engineering; Springer: Cham, Switzerland, 2023. [Google Scholar]
  5. Harsanyi, J.C.; Selten, R. A generalized Nash solution for two-person bargaining games with incomplete information. Manag. Sci. 1972, 18, 80–106. [Google Scholar] [CrossRef]
  6. Conitzer, V.; Sandholm, T. Computing the Optimal Strategy to Commit to. In Proceedings of the 7th ACM Conference on Electronic Commerce, Ann Arbor, MI, USA, 11–15 June 2006; pp. 82–90. [Google Scholar]
  7. Paruchuri, P.J.; Pearce, P.; Marecki, J.; Tambe, M.; Ordonez, F.; Kraus, S. Playing Games for Security: An Efficient Exact Algorithm for Solving Bayesian Stackelberg Games. In Proceedings of the 7th International Joint Conference on Autonomous Agents and Multiagent Systems, Estoril, Portugal, 12–16 May 2008; Volume 2, pp. 895–902. [Google Scholar]
  8. Jain, M.; Kiekintveld, C.; Tambe, M. Quality-Bounded Solutions for Finite Bayesian Stackelberg Games: Scaling up. In Proceedings of the 10th International Conference on Autonomous Agents and Multiagent Systems, Taipei, Taiwan, 2–6 May 2011; Volume 3, pp. 997–1004. [Google Scholar]
  9. Yin, Z.; Tambe, M. A Unified Method for Handling Discrete and Continuous Uncertainty in Bayesian Stackelberg Games. In Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems, Valencia, Spain, 4–8 June 2012; Volume 2, pp. 855–862. [Google Scholar]
  10. Wilczyński, A.; Jakóbik, A.; Kołodziej, J. Stackelberg Security Games: Models, Applications and Computational Aspects. J. Telecomun. Inf. Technol. 2016, 3, 70–79. [Google Scholar] [CrossRef]
  11. Sayed Ahmed, I. Stackelberg-Based Anti-Jamming Game for Cooperative Cognitive Radio Networks. Ph.D. Thesis, University of Calgary, Calgary, AB, USA, 2017. [Google Scholar]
  12. Gan, J.; Elkind, E.; Wooldridge, M. Stackelberg Security Games with Multiple Uncoordinated Defenders. In Proceedings of the 17th International Conference on Autonomous Agents and Multiagent Systems (AAMAS), Stockholm, Sweden, 10–15 July 2018; pp. 703–711. [Google Scholar]
  13. Basilico, N.; Gatti, N.; Villa, F. Asynchronous multi-robot patrolling against in trusions in arbitrary topologies. In Proceedings of the Conference for the Advancement of Artificial Intelligence (AAAI), Atlanta, GA, USA, 11–15 July 2010; pp. 345–350. [Google Scholar]
  14. Clempner, J.B. Learning attack-defense response in continuous-time discrete-states Stackelberg Security Markov games. J. Exp. Theor. Artif. Intell. 2022. to be published. [Google Scholar] [CrossRef]
  15. Albarran, S.; Clempner, J.B. A Stackelberg security Markov game based on partial information for strategic decision making against unexpected attacks. Eng. Appl. Artif. Intell. 2019, 81, 408–419. [Google Scholar] [CrossRef]
  16. Trejo, K.K.; Clempner, J.B.; Poznyak, A.S. Adapting Attackers and Defenders Preferred Strategies: A Reinforcement Learning Approach in Stackelberg Security Games. J. Comput. Syst. Sci. 2018, 95, 35–54. [Google Scholar] [CrossRef]
  17. Solis, C. Ship differential game approach for multiple players: Stackelberg security games. Optim. Control Appl. Methods 2020, 41, 312–326. [Google Scholar] [CrossRef]
  18. Alcantara-Jiménez, G.; Clempner, J. Repeated Stackelberg security games: Learning with incomplete state information. Reliab. Eng. Syst. Saf. 2020, 195, 106695. [Google Scholar] [CrossRef]
  19. Clempner, J.B.; Poznyak, A.S. Stackelberg Security Games: Computing The Shortest-Path Equilibrium. Expert. Syst. Appl. 2015, 42, 3967–3979. [Google Scholar] [CrossRef]
  20. Rahman, M.; Oh, J. Online Learning for Patrolling Robots Against Active Adversarial Attackers. In Proceedings of the Recent Trends and Future Technology in Applied Intelligence; Lecture Notes in Computer Science; Mouhoub, M., Sadaoui, S., Ait Mohamed, O., Ali, M., Eds.; Springer: Montreal, QC, Canada, 2018; Volume 10868. [Google Scholar]
  21. Wang, Y.; Shi, Z.; Yu, L.; Wu, Y.; Singh, R.; Joppa, L.; Fang, F. Deep Reinforcement Learning for Green Security Games with Real-Time Information. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19), Honolulu, HI, USA, 27 January–1 February 2019; pp. 1401–1408. [Google Scholar]
  22. Li, T.; Li, H.; Pan, Y.; Xu, T.; Zheng, Z.; Zhu, Q. Meta Stackelberg Game: Robust Federated Learning against Adaptive and Mixed Poisoning Attacks. arXiv 2024, arXiv:2410.17431. [Google Scholar]
  23. Sengupta, S.; Kambhampati, S. Multi-agent Reinforcement Learning in Bayesian Stackelberg Markov Games for Adaptive Moving Target Defense. arXiv 2020, arXiv:2007.10457. [Google Scholar]
  24. Shukla, P.; An, L.; Chakrabortty, A.; Duel-Hallen, A. A robust Stackelberg game for cyber-security investment in networked control systems. IEEE Trans. Control Syst. Technol. 2022, 31, 856–871. [Google Scholar] [CrossRef]
  25. Clempner, J.B.; Poznyak, A.S. Analytical Method for Mechanism Design in Partially Observable Markov Games. Mathematics 2021, 9, 321. [Google Scholar] [CrossRef]
  26. Clempner, J.B. A Markovian Stackelberg game approach for computing an optimal dynamic mechanism. Comput. Appl. Math. 2021, 40, 186. [Google Scholar] [CrossRef]
  27. Asiain, E.; Clempner, J.B.; Poznyak, A.S. Controller Exploitation-Exploration: A Reinforcement Learning Architecture. Soft Comput. 2019, 23, 3591–3604. [Google Scholar] [CrossRef]
  28. Benson, A.R.; Gleich, D.F.; Lim, L.K. The Spacey Random Walk: A Stochastic Process for Higher-Order Data. SIAM Rev. 2017, 59, 321–345. [Google Scholar] [CrossRef]
Figure 1. Convergence for defender 1’s strategy ξ 1 .
Figure 1. Convergence for defender 1’s strategy ξ 1 .
Mca 30 00029 g001
Figure 2. Convergence for attacker 2’s strategy ξ 2 .
Figure 2. Convergence for attacker 2’s strategy ξ 2 .
Mca 30 00029 g002
Figure 3. Convergence for attacker 3’s strategy ξ 3 .
Figure 3. Convergence for attacker 3’s strategy ξ 3 .
Mca 30 00029 g003
Figure 4. Convergence for defender 1’s error of p 1 .
Figure 4. Convergence for defender 1’s error of p 1 .
Mca 30 00029 g004
Figure 5. Convergence for attacker 2’s error of p 2 .
Figure 5. Convergence for attacker 2’s error of p 2 .
Mca 30 00029 g005
Figure 6. Convergence for attacker 3’s error p 3 .
Figure 6. Convergence for attacker 3’s error p 3 .
Mca 30 00029 g006
Figure 7. Convergence for defender 1’s error of v 1 .
Figure 7. Convergence for defender 1’s error of v 1 .
Mca 30 00029 g007
Figure 8. Convergence for attacker 2’s error of v 2 .
Figure 8. Convergence for attacker 2’s error of v 2 .
Mca 30 00029 g008
Figure 9. Convergence for attacker 3’s error of v 3 .
Figure 9. Convergence for attacker 3’s error of v 3 .
Mca 30 00029 g009
Figure 10. Game realization 1: the attackers are intercepted by defender 1 (blue).
Figure 10. Game realization 1: the attackers are intercepted by defender 1 (blue).
Mca 30 00029 g010
Figure 11. Game realization 2: the attackers are intercepted by defender 1 (blue).
Figure 11. Game realization 2: the attackers are intercepted by defender 1 (blue).
Mca 30 00029 g011
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Clempner, J.B. Learning Deceptive Tactics for Defense and Attack in Bayesian–Markov Stackelberg Security Games. Math. Comput. Appl. 2025, 30, 29. https://doi.org/10.3390/mca30020029

AMA Style

Clempner JB. Learning Deceptive Tactics for Defense and Attack in Bayesian–Markov Stackelberg Security Games. Mathematical and Computational Applications. 2025; 30(2):29. https://doi.org/10.3390/mca30020029

Chicago/Turabian Style

Clempner, Julio B. 2025. "Learning Deceptive Tactics for Defense and Attack in Bayesian–Markov Stackelberg Security Games" Mathematical and Computational Applications 30, no. 2: 29. https://doi.org/10.3390/mca30020029

APA Style

Clempner, J. B. (2025). Learning Deceptive Tactics for Defense and Attack in Bayesian–Markov Stackelberg Security Games. Mathematical and Computational Applications, 30(2), 29. https://doi.org/10.3390/mca30020029

Article Metrics

Back to TopTop