Next Article in Journal
A Distorted-Image Quality Assessment Algorithm Based on a Sparse Structure and Subjective Perception
Previous Article in Journal
A Numerical Investigation of Activation Energy Impact on MHD Water-Based Fe3O4 and CoFe2O4 Flow between the Rotating Cone and Expanding Disc
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Signaling Security Games with Attack Planner Deception

by
Santing He
1,*,
Mingchu Li
2 and
Runfa Zhang
3
1
School of Mathematical Sciences, Dalian University of Technology, Dalian 116000, China
2
School of Software Technology, Dalian University of Technology, Dalian 116000, China
3
School of Automation and Software Engineering, Shanxi University, Taiyuan 030013, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(16), 2532; https://doi.org/10.3390/math12162532
Submission received: 8 July 2024 / Revised: 9 August 2024 / Accepted: 14 August 2024 / Published: 16 August 2024

Abstract

:
This paper studies a class of attack behavior in which adversaries assume the role of initiators, orchestrating and implementing attacks by hiring executors. We examine the dynamics of strategic attacks, modeling the initiator as an attack planner and constructing the interaction with the defender within a defender–attack planner framework. The individuals tasked with executing the attacks are identified as attackers. To ensure the attackers’ adherence to the planner’s directives, we concurrently consider the interests of each attacker by formulating a multi-objective problem. Furthermore, acknowledging the information asymmetry where defenders have incomplete knowledge of the planners’ payments and the attackers’ profiles, and recognizing the planner’s potential to exploit this for strategic deception, we develop a defender–attack planner model with deception based on signaling games. Subsequently, through the analysis of the interaction between the defender and planner, we refine the model into a tri-level programming problem. To address this, we introduce an effective decomposition algorithm leveraging genetic algorithms. Ultimately, our numerical experiments substantiate that the attack planner’s deceptive strategy indeed yield greater benefits.

1. Introduction

Game theory has found extensive applications in various fields, especially in the Stackelberg games (SGs), which are simple yet powerful models for sequential interaction between two strategic players: a leader and a follower. Specifically, the leader commits to a strategy first, and the follower responds after observing the leader’s commitment. The game was first proposed in 1934, when researchers introduced it to analyze market competition between a large-leader firm and a small-follower firm [1]. Since then, it has been applied to a diverse array of problems, including principal–agent contract design [2], pricing and planning in transportation systems [3], influence maximization [4] and exam design [5], often focusing on the leader’s perspective and with the aim of identifying the optimal strategy for the leader to commit to. Over recent years, with the successful application of SGs in security resource allocation, an increasing number of experts have utilized them to address security problems, known as the Stackelberg Security Games (SSGs) [6]; thus, SSGs have emerged as a significant area of research for SGs. To date, SSGs have been applied to address a multitude of practical issues, such as threat screening games [7], green security games [8,9,10] and crime prevention strategies [11]. In SSGs, the defender, as the leader, implements an optimal mixed (or randomized) defense strategy. This mixed strategy represents a probability distribution across all the possible defense strategies. The attackers, as the followers, observe this and adjust their strategy to optimize their outcome [12]. This has been seen in, for instance, the Los Angeles Sheriff’s Department’s application to the subway system [13] and the US Coast Guard’s initiatives on ports and waterways in Boston and New York City [14], among others. In these applications, a defender aims to protect targets from strategic adversaries by deploying limited resources. The central solution concept in SSGs is the strong Stackelberg equilibrium (SSE) [15,16], which determines the defender’s optimal strategy. However, SSE is predicated on the assumption that both parties have complete knowledge of each other’s information. In numerous real-world scenarios, this assumption often fails to hold true. The defender frequently lacks complete information about the follower. Currently, these issues have been extensively researched, resulting in a multitude of proposed solutions. For example, when the defender is uncertain about the attacker’s payoff but is aware of its range, strategies are explored that are robust to small interval uncertainties [17,18,19]. In scenarios where the defender is only aware of the distribution of the attacker’s payoff, the Bayesian Stackelberg equilibrium is used to maximize the defender’s utility [20,21,22]. Furthermore, as defenders gather information through interactions with attackers, numerous studies have investigated learning from the attackers’ response strategies to refine leadership tactics [17,23,24,25,26]. The essence of these approaches is the defender’s attempt to acquire relevant information about the attacker, assuming the information obtained is accurate. However, when the attacker becomes aware of the defender’s reliance on information, there is a tendency for the attacker to use tactics that obfuscate the defender, intending to provide them with misleading information, which is a form of deceitful behavior.
Currently, the attackers’ deception has attracted significant attention, prompting a variety of studies to focus on this phenomenon. Existing research mainly investigates scenarios where attackers provide false information about a single aspect, such as their payoff [27,28,29] or the degree of rationality [30,31,32]. However, in numerous real-world scenarios, defenders confront a more intricate challenge where attackers have multiple pieces of undisclosed information. Recognizing this gap, broadening the scope of analysis to include situations with followers possessing multiple private information sets is essential. By considering a scenario in which the attacker has two types of private information, we can establish a more comprehensive model. In the existing body of research, the attacker commonly uses strategies that involve either selecting a target to attack [33] or allocating resources for an attack on the targets [34,35]. However, real-world scenarios frequently exhibit a more complex dynamic where the attacker assumes the role of a “leader” who orchestrates an attack and recruits individuals to execute it. In this context, the individual who devises the attack strategy is referred to as the “attack planner”, while the person commissioned to execute the attack is termed the “attacker”. Unlike resources, which can be allocated at will, individuals have independent thoughts and agency. Therefore, it is essential to consider their interests to ensure their adherence to the arranged strategy. Furthermore, the defender is often unaware of the specifics regarding the hired individuals, which we establish as the attack planner’s private information. This can be exploited by the attack planner to mislead defenders by misrepresenting the number of hired individuals.
In summary, this paper introduces a model that delineates the interaction between the attack planner and the defender, then extends this model to account for the attack planner’s deception and the compliance of hired individuals through a multi-objective problem. Subsequently, we propose a tri-level programming framework that aligns with the interactive process. The first level involves signaling, wherein the attack planner seeks to obscure the defender’s comprehension and maximize personal benefits. The second level pertains defense, wherein the defender strategically allocates resources to minimize potential losses. The third level constitutes the attack phase, wherein the attack planner coordinates the execution of the attack, aiming to maximize profit. Simultaneously, to ensure the attackers’ adherence with the plan, the attack strategy formulated by the attack planner must safeguard the attackers’ interests. This model deviates from the conventional tri-level min–max–min programming models [36,37], rendering traditional solving algorithms inapplicable. Consequently, we introduce an efficient, customized algorithm designed to accommodate the unique structure and requirements of the model, ensuring effective resolution and thorough analysis.
Our contributions: In order to deal with the opponent’s deception effectively, this paper analyzes a kind of deception behavior in attack in detail.
  • We model the attack initiator as the attack planner—the individual they hire to carry out the attack is the attacker—and we propose a defender–attack planner (D-AP) model to describe the interaction between them;
  • We account for the interests of each attacker to ensure compliance, framing this as a multi-objective optimization issue;
  • This paper also addresses the defender’s uncertainty to the attack planner’s payment and the particulars of the hired attackers, constructing a defender–attack planner game model with deception (D-APD).
  • We articulate D-APD as a tri-level programming problem and propose a customized decomposition algorithm based on a genetic algorithm (DA-GA) to address it.
The remainder of this paper consists of the following. In Section 2, we summarize the previous related work. In Section 3, we present detailed descriptions of the D-AP and D-APD models. In Section 4, we formulate a tri-level programming problem and propose a solution algorithm. In Section 5, we a conduct further analysis of the model via experimental validation. Finally, Section 6 presents the conclusions of this paper and outlines directions for future research.

2. Related Work

This section provides an overview of the existing literature that is pertinent to our research, focusing on two key aspects: deception behavior within games and the role of the planner.

2.1. Deception in Games

As deception can obfuscate an adversary’s understanding and disrupt their decision-making process, certain research endeavors have explored methods to mitigate or neutralize criminal offensives through deceptive tactics. Within the homeland security domain, studies have formulated models of deceptive strategies wherein the defender employs a range of signals to mislead the attacker [38,39,40]. Ref. [33] examines security games involving the defender’s use of disguised resources. Ref. [41] explores the concept of strategic secrecy in the defender’s approach to infrastructure protection. In addition, when the attacker is unsure of the amount of the defender’s resources, the defender may deceive the attacker by disguising or hiding a portion of the resources, potentially reporting fewer resources than are actually available [33]. In contrast to [33], Ref. [42] studied the phenomenon of bluffing in signaling games, where the defender may claim to possess more resources than are actually available, serving as a deterrent to the attacker and thus preventing the attack. In [43], a problem with multiple attacker types is studied, in which the defender initially commits to a mixed strategy and signaling scheme, identifies the attacker’s type and then acts, and the attacker subsequently responds. In [44], a two-stage game model is presented, in which the first stage is a basic defender–attacker game, and in the second stage, the defender sends a deceptive signal regarding actual resource allocation. Ref. [40] explores the issue of deception in a single-objective multi-period game where multiple types of defenders exist. In each game, they initially select a behavior and a signaling strategy. Another work has discussed the problem of stochastic games with multiple defenders cheating [45]. In [34], researchers used the hypergame framework to study the deception of defenders in a game with multiple attackers. Attackers with multiple types possess distinct payoff functions and resource quantities, so the hypergame framework is used to model each attacker individually. Ref. [46] examines the optimal conditions and methods for defenders to effectively deceive various attackers. Other studies investigated the deception of a defender in a game with ambiguous benefits and multiple attackers [47]. Ref. [48] examines how defenders counter advanced sustainability threats through deceptive behavior.
However, precisely because the deceptive behavior confuses the adversary, more and more attackers will choose to deceive, leading to a deviation in defense strategy and an infliction of more severe damage. In [27], with distinct types characterized by varying payoffs, the attacker deceives the defender by imitating other attacker types. In [28], the attacker deceives the defender by directly presenting a false type. Nguyen et al. studied the deception strategies of attackers in repeated games [29]. Ref. [49] addresses similar problems. Unlike Ref. [29], this study examines the issues within the SSG framework, whereas Ref. [29] focuses on a stochastic one. Ref. [30] studies deception in games with multiple attackers, where multiple attackers pay the same but have different levels of rationality, and perfectly rational attackers adopt a guise of bounded rationality to deceive defenders. However, these studies do not consider the scenario where attackers possess multiple pieces of private information.
In addition, because asymmetric information is a key friction in many economic interactions, deception often appears in the principal–agent problem [2,50,51], which is a prevalent application of the SGs. The principal delegates the management of the output process to the agent. A contract is signed in advance, specifying the terms of an incentive payment. The agent exerts costly effort for managing the output. Subsequently, given the contract offered by the principal, the agent returns an optimal effort response that optimally balances their cost of effort with the proposed compensation. Ultimately, the principal selects the optimal contract to motivate the agent’s effort [52]. Then, a principal may offer a range of contracts to an agent with uncertain characteristics and ask the agent to choose the one that matches their true type. Naturally, the principal’s menu must consider the agent’s potential misrepresentation of their type, meaning that the agent may mislead the principal [53]. However, such a deception is predominantly observed in economic contexts and significantly differs from the deception encountered in the security domain we examine.

2.2. About Planner

Several studies have examined the beneficial impacts of community participation in wildlife preservation [54,55,56,57,58], but these studies lack a mathematical model of the interaction between defenders and attackers; some researchers formulated it using the Stackelberg game framework [59]. Acknowledging the possibility of community members colluding with the attacker, the patrol, acting as the defender, employs community informants to gather intelligence on the attacker’s activities. However, no existing studies have yet considered a scenario where attackers may also hire some people to execute the attack, which significantly diverges from the defender’s recruitment of community members, including the consideration of the interests of the hired attackers. It is in addressing this gap that this paper contributes to the literature.

3. Model

In this section, we initially construct a foundational defender–attack planner game model. Subsequently, considering the attack planner’s private information, we construct a defender–attack planner with a deception model based on signaling games. Table 1 describes the mathematical notation used in this paper.

3.1. Defender–Attack Planner Game Model

In this section, we develop a model that captures the strategic interaction between a defender and an attack planner within the framework of a Stackelberg game, known as the Defender–Attack Planner (D-AP) model.

3.1.1. Model Description

A D-AP game model consists of a defender and an attack planner (hereafter referred to as the planner), where the defender needs to protect a set of N targets, and the planner hires some attackers to attack a subset of targets. The hired attackers are categorized into two types—strong and weak—based on their offensive capabilities (the model is also extensible to scenarios involving multiple attacker types). Their basic salary is different, which is denoted as b s = ( b s 1 , b s 2 ) ; b s 1 and b s 2 correspond to the base salaries of the strong and weak attackers, respectively. The defender also possesses two defense levels: the high-level defense is capable of repelling all attackers, while the low-level defense is effective only against weak attackers. The defender incurs a fixed cost upon protecting a target; c = ( c 1 , c 2 , c 3 ) represents the cost vector, where c 1 = 0 signifies the absence of defense and c 2 and c 3 correspond to the costs of low-level and high-level defense, respectively. The defender has a budget constraint B and the planner also faces budgetary constraints, which restricts the number of attackers they can hire; this is denoted as n = ( n 1 , n 2 ) , where planner-hired attackers are n, including n 1 strong attackers and n 2 weak attackers. a t = ( a t 1 , , a t n ) denotes each attacker’s type (strong or weak), where a t i { 1 , 2 } and a t i = 1 ( 2 ) represent that the i-th attacker is weak (strong). Each target holds a distinct importance to the defender and the planner, that is, they have different valuations of the targets. The defender’s valuation of the targets is v d = ( v 1 d , , v N d ) and the planner’s valuation is v a = ( v 1 a , , v N a ) . When a target j is attacked successfully, the defender incurs a penalty P j D and the planner receives a reward R j A . On the contrary, if an attack on a target j is unsuccessful, the defender will gain a reward R j D and the planner will incur a penalty P j A , and they satisfy R j D > P j D , R j A > P j A . This paper assumes that the utility of both sides is only related to the targets that were successfully attacked; in this case, R j D = P j A = 0 , P j D = v j d and R j A = v j a .

3.1.2. Strategy Expression

The defender’s pure strategy is to decide what level of protection to apply to each target within budget constraints which is denoted by s d = ( s d 1 , , s d N ) , where s d j { 0 , 1 , 2 } and j = 1 N c s d j + 1 B , s d j = 0 indicate that the target j is not protected; s d j = 1 means low-level defense against the target j and s d j = 2 denotes high-level defense against the target j. S D = { s d | s d j { 0 , 1 , 2 } , j = 1 N c s d j + 1 B } represents the set of all the defender’s pure strategies. The planner’s pure strategy involves determining the allocation of the hired attackers to execute an attack on a designated set of targets, represented by s a = ( s a i j ) i [ n ] , j [ N ] , where s a i j { 0 , 1 } and s a i j = 1 ( 0 ) mean that the i-th attacker is (not) arranged to attack the target j. Following a successful attack on a target, the attacker responsible is granted a commission proportional to the target’s value to the planner. Although it may appear that one attacker is assigned per target and vice versa, the diverse values of the targets result in a scenario of contention among multiple attackers, each seeking to secure a higher commission. Acknowledging this possibility, we assume that different attackers are permitted to attack the same target. However, the commission received from a successful attack will be evenly divided among the participating attackers, and the probability of a successful attack on a target remains unaffected by an increased number of attackers. This assumption facilitates adherence to the planner’s strategy among attackers. S A = { s a n × N | s a i j { 0 , 1 } , j = 1 N s a i j = 1 , i = 1 , , n } denotes the planner’s strategy space, where j = 1 N s a i j = 1 means that each attacker can only attack one target.

3.1.3. Utility Function

Given the attack strategy s a , we let s a 1 = ( s a i j 1 ) i [ n ] , j [ N ] , where s a i j 1 = 0 if s a i j = 0 and s a i j 1 = a t i if s a i j = 1 ; then, s a j 2 = m a x { s a i j 1 , i = 1 , , n } signifies the magnitude of the attack that target j is projected to encounter, denoting s a 2 = ( s a 1 2 , , s a N 2 ) . Specifically, should the planner allocate only weak attackers to a target, then the target faces only weak attacks; conversely, the target endures strong attacks. Then, for a strategy profile ( s d , s a ) , the probability that the target j is successfully attacked is
q j ( s d , s a ) = 1 , s a j 2 > s d j 0 , o t h e r w i s e ,
this is because the high-level defense is capable of repelling all attackers, while the low-level defense is effective only against weak attackers. In addition, we define s a j 3 = i = 1 n s a i j as the number of attackers attacking the target j, and let s a 3 = ( s a 1 3 , , s a N 3 ) .
The total loss of the defender, which is denoted as L D , is the sum of the value of the targets successfully attacked and the cost of the defensive strategy; then, for a strategy profile ( s d , s a ) , we have
L D ( s d , s a ) = j = 1 N ( q j ( s d , s a ) v j d + c s d j + 1 ) .
The utility of the planner, U A P , is the aggregate value derived from successfully attacked targets, minus the cost of hiring all attackers, then
U A P ( s d , s a ) = j = 1 N q j ( s d , s a ) v j a i = 1 n U i ( s d , s a ) ,
where U i is the benefit of the i-th employed attacker, i.e., the cost of hiring the i-th attacker. For each hired attacker, irrespective of the attack’s success, they receive a base salary, incur a certain attack cost, and upon a successful attack, they receive an additional commission proportional to the target’s value. Thus
U i ( s d , s a ) = u i + j = 1 N s a i j q j ( s d , s a ) · α v j a s a j 3 j = 1 N s a i j a c i j ,
where u i = b s a t i is the i-th attacker’s base salary, α ( 0 , 1 ) denotes the proportionality constant of the commission awarded to the attacker post-successful attack and a c i j denotes the i-th attacker’s cost when they attacks target j. Attackers, despite having identical capabilities, may demonstrate diverse target preferences, influenced by a multitude of factors, such as the distance between the attacker and the target. This study incorporates these individual differences among attackers by representing these differences through variable attack costs a c .

3.2. Defender–Attack Planner Game Model with Deception

Building upon the previously outlined model, we develop an advanced Defender–Attack Planner game model with Deception (D-APD). This model takes into account the planner’s strategic deception in two critical dimensions: the payoff type and the attacker type. The deception in payoff type could lead the defender to misestimate the importance of the targets to the planner. Similarly, deception in attacker type could involve obscuring the true attack capabilities, which can confuse the defender’s assessment of the attack’s treat level. By integrating these elements, the D-APD model encapsulates the intricate interplay of strategy and information, and thus provides a framework for analyzing sophisticated security threats.

3.2.1. The Attack Planner with Multiple Types

In reality, defenders are typically aware only of the range of possible scenarios and the likelihood of each occurring, which is encapsulated in the prior probability. First, denoting the set of all possible payment types as Θ = { θ 1 , , θ m } , the corresponding prior probability distribution is p p t : Θ [ 0 , 1 ] . A planner with payment type θ values each target as v θ = ( v 1 θ , , v N θ ) . Then, we denote the set of all possible hired attacker situations in which the planner with payment type θ as Φ θ = { φ 1 θ , , φ m θ θ } ; that is, the planner with payment type θ has m θ possible attacker types, where φ k θ = ( φ k 1 θ , φ k 2 θ ) means that the planner with payment type θ hired φ k 1 θ strong attackers and φ k 2 θ weak attackers. The corresponding prior probability distribution is p θ : Φ Θ [ 0 , 1 ] . Φ = θ Θ Φ θ represents the set of all attacker types. In practice, although the attacker type for each payment type planner is different, they can be viewed as the same by taking a union. For example, if the set of attacker types for a planner with payment type θ 1 is Φ θ 1 = { φ 1 , φ 2 } , the set of attacker types for a planner with payment type θ 2 is Φ θ 2 = { φ 2 , φ 3 } , then we can set Φ θ 1 = Φ θ 2 = { φ 1 , φ 2 , φ 3 } and p θ 1 ( φ 3 ) = 0 , p θ 2 ( φ 1 ) = 0 . Therefore, we can directly let Φ = { φ 1 , , φ M } , which represents the set of all possible attacker types (M in total). φ i = ( φ 1 i , φ 2 i ) means hiring a φ 1 i strong attacker and a φ 2 i weak attacker, and we denote the prior probability of the planner of payment type θ over all attacker types as p θ : Φ [ 0 , 1 ] . Obviously, each attacker type φ corresponds to a group of attackers, and the base salary of this group of attackers is recorded as u φ = ( u 1 φ , , u | | φ | | 1 φ ) ; the attack cost matrix is recorded as a c φ = ( a c i j φ ) i [ | | φ | | 1 ] , j [ N ] . Then, under the strategy profile ( s d , s a ) , the utility function of the planner with payment type θ and attacker type φ is
U θ , φ ( s d , s a ) = j = 1 N q j ( s d , s a ) v j θ i = 1 | | φ | | 1 U θ , φ , i ( s d , s a ) ,
where U θ , φ , i is each employed attacker’s benefit, that is
U θ , φ , i ( s d , s a ) = u i φ + j = 1 N s a i j q j ( s d , s a ) · α v j θ s a j 3 j = 1 N s a i j a c i j φ .

3.2.2. Attack Planner’s Deception

For private information, the planner is usually more inclined to mislead the defenders by deceiving rather than reporting the truth. The practice of sending preemptive signals has become the prevalent method of deception. These signals are crafted to influence the defender’s perception of the threat and subsequent resource allocation. However, it is important to note that there is a cost associated with this deceptive behavior, termed the deception cost, denoted as d c = ( d c 1 , d c 2 ) , d c 1 and d c 2 is the planner’s deception cost about payment and attacker type, respectively. First, the planner will achieve deception by simulating the behavior of a planner of other payment types, so the set of signals sent about the payment type is S 1 = Θ and the corresponding set of mixed-signal sending strategies is Δ o 1 = { o 1 = ( o 1 s 1 ) s 1 S 1 | o 1 s 1 [ 0 , 1 ] , | | o 1 | | 1 = 1 } , where o 1 s 1 is the probability of sending the signal s 1 . In addition, for the attacker type, the planner will deceive the defender by arranging the attacker to disguise or hide (such as plainclothes police), so that the defender will have a wrong understanding of the real number of hired attackers. Therefore, the number of attackers reported by the planner cannot be more than the number of real hired attackers. Then the set of signals sent by the planner with the payment type θ and the attacker type φ about the attacker type is S 2 , θ , φ = { s 2 N + | s 2 φ 1 + φ 2 } (the planner only reports the total number of attackers, not the specific number of strong and weak attackers). For example, suppose a planner’s set of attacker types is { ( 1 , 2 ) , ( 2 , 4 ) , ( 3 , 1 ) } , then the set of signals they send is { 1 , 2 , 3 , 4 , 5 , 6 } . But in fact, we can find that when the planner sends signal 1 or 2, the effect on the posterior probability is the same as when they send signal 3; this is because when sending signals 1, 2 and 3, the possible types of planner are ( 1 , 2 ) , ( 2 , 4 ) , ( 3 , 1 ) . Therefore, we reasonably let S 2 = { | | φ | | 1 | φ Φ } represent the set of signals about the planner’s attacker type; then, S 2 , φ = { s 2 S 2 | s 2 φ 1 + φ 2 } is the set of signals about the attacker type for the attacker type φ planner, and their collection of mixed signaling strategies is Δ o 2 φ = { o 2 = ( o 2 s 2 ) s 2 S 2 , φ | o 2 s 2 [ 0 , 1 ] , | | o 2 | | 1 = 1 } , where o 2 s 2 is the probability that the planner sends a signal s 2 about their attacker type. Then Δ o 2 = φ Φ Δ o 2 φ is the collection of all the mixed strategies about the attacker type.

3.2.3. Game Process

The game begins with the planner committing to a signaling scheme π o = ( π o 1 , π o 2 ) , where π o 1 : Θ × Φ Δ o 1 is the planner’s signaling scheme about their payment type, that is, π o 1 ( θ , φ ) represents the signaling scheme of the planner with payment type θ and the attacker type φ ; π o 1 ( s 1 | θ , φ ) is the probability that they send signal s 1 , π o 1 ( s 1 | θ ) = φ Φ π o 1 ( s 1 | θ , φ ) is the probability that the planner with payment type θ sends signal s 1 . Similarly, π o 2 : Θ × Φ Δ o 2 is the planner’s signaling scheme regarding the attacker type.
Then, after receiving the signal ( s 1 , s 2 ) , the defender updates their belief on the planner’s type according to Bayes’ rule, which is the posterior probability. The posterior probability about the payment type is
μ p t ( θ | s 1 ) = p p t ( θ ) π o 1 ( s 1 | θ ) θ Θ p p t ( θ ) π o 1 ( s 1 | θ ) = p p t ( θ ) φ Φ π o 1 ( s 1 | θ , φ ) θ Θ p p t ( θ ) φ Φ π o 1 ( s 1 | θ , φ ) ,
and the posterior probability about the attacker type of planner with payment type θ is
μ a n ( φ | θ , s 2 ) = p p t ( θ ) p θ ( φ ) π o 2 ( s 2 | θ , φ ) φ Φ p p t ( θ ) p θ ( φ ) π o 2 ( s 2 | θ , φ ) = p θ ( φ ) π o 2 ( s 2 | θ , φ ) φ Φ p θ ( φ ) π o 2 ( s 2 | θ , φ ) .
Then the posterior probability that the planner’s payment type is θ and attacker’s type is φ is μ ( θ , φ | s 1 , s 2 ) = μ p t ( θ | s 1 ) · μ a n ( φ | θ , s 2 ) . Subsequently, the defender chooses a defense strategy based on their posterior belief to protect targets; we denote their strategic scheme as π d : S 1 × S 2 S D , that is, π d ( s 1 , s 2 ) = ( π d ( j | s 1 , s 2 ) ) j [ N ] represents the strategy that the defender takes after receiving the signal s 1 , s 2 and π d ( j | s 1 , s 2 ) { 0 , 1 , 2 } is the defense level of the defender at target j at this time.
Finally, the planner arranges for the attackers to carry out the attack and the strategic scheme is π a : Θ × Φ × S 1 × S 2 S A ; π a ( θ , φ , s 1 , s 2 ) = ( π a ( i j | θ , φ , s 1 , s 2 ) ) i [ | | φ | | 1 ] , j [ N ] is the attacker arrangement strategy adopted by the planner with payment type θ and attacker type φ when sending signals s 1 , s 2 and π a ( i j | θ , φ , s 1 , s 2 ) { 0 , 1 } represents the case where the i-th attacker is scheduled to attack target j at this time. The game process is shown in Figure 1.

3.2.4. Utility of Both Sides

After receiving the signal s 1 , s 2 , the defender’s loss function is
L s 1 , s 2 ( π o , π d , π a ) = θ Θ φ Φ : | | φ | | 1 s 2 μ ( θ , φ | s 1 , s 2 ) · L D ( π d ( s 1 , s 2 ) , π a ( θ , φ , s 1 , s 2 ) ) .
The utility function of the planner with payment type θ and the attacker type φ when sending signal s 1 , s 2 is
U s 1 , s 2 θ , φ ( π d , π a ) = U θ , φ ( π d ( s 1 , s 2 ) , π a ( θ , φ , s 1 , s 2 ) ) i n d ( s 1 , θ ) · d c 1 i n d ( s 2 , | | φ | | 1 ) · d c 2 ,
where i n d ( a , b ) = 0 , a = b 1 , a b is indicator function. Thus, the total expected utility of the planner of type θ and attacker type φ is
U A P θ , φ ( π o , π d , π a ) = s 1 S 1 s 2 S 2 , φ π o 1 ( s 1 | θ , φ ) · π o 2 ( s 2 | θ , φ ) · U s 1 , s 2 θ , φ ( π d ( s 1 , s 2 ) , π a ( θ , φ , s 1 , s 2 ) ) ,
which corresponds to the gains of the i-th attacker, which are shown as
U i θ , φ ( π o , π d , π a ) = s 1 S 1 s 2 S 2 , φ π o 1 ( s 1 | θ , φ ) · π o 2 ( s 2 | θ , φ ) · U θ , φ , i ( π d ( s 1 , s 2 ) , π a ( θ , φ , s 1 , s 2 ) ) .

3.2.5. The Equilibrium

For each planner with the payment type θ and the attacker type φ , given the signaling strategy π o and defender strategy π d , let f k ( π a ) = U k θ , φ ( π o , π d , π a ) , k = 1 , , | | φ | | 1 ; then, by solving the Pareto optimal solutions of the multi-objective problem
m a x π a f = ( f 1 ( π a ) , , f k ( π a ) , , f | | φ | | 1 ( π a ) ) ,
we can obtain a Pareto optimal set, denoted as P a r S A θ , φ . In Definition 1, we give the specific definition of the model equilibrium solution.
Definition 1. 
The strategy profile ( π o * , π d * , π a * ) is called an equilibrium solution of model if and only if
  • π o * = < π o 1 * ( θ , φ ) , π o 2 * ( θ , φ ) > = a r g m a x π o 1 , π o 2 U A P θ , φ ( π o , π d * , π a * ) , θ , φ
  • π d * ( s 1 , s 2 ) = a r g m i n π d L s 1 , s 2 ( π o * , π d , π a * ) , s 1 , s 2
  • π a * ( θ , φ ) = a r g m a x π a P a r S A θ , φ U A P θ , φ ( π o * , π d * , π a ) , θ , φ .

3.3. Motivating Example

To explain the model more intuitively, we give an example. Assume there are three targets and the set is T = { t 1 , t 2 , t 3 } , and the planner has two payment types Θ = { θ 1 , θ 2 } and two attacker types Φ = { φ 1 , φ 2 } . The specific values are shown in Table 2.
When the planner chooses to reveal the true type, we can compute the SSE of this game. For the planner with payment type θ 1 and attacker type φ 1 , the defender strategy is ( 0 , 2 , 0 ) , i.e., they choose to implement a high-level defense against target t 2 , which makes sense because target t 2 is of the highest value to the planner with payment type θ 1 ; the corresponding attack strategy is ‘ t 3 ’. For a planner with payment type θ 1 and attacker type φ 2 where the defender chooses not to defend, the corresponding attack strategy is ( t 3 , t 2 ) , i.e., arrange for a strong attacker to attack target t 3 and a weak attacker to attack target ‘ t 2 ’. For a planner with payment type θ 2 and attacker type φ 1 where the defender chooses not to defend, the corresponding attack strategy is t 2 . For a planner with payment type θ 2 and attacker type φ 2 where the defender implements a high level of defense against the target t 1 , the corresponding attack strategy is ( t 3 , t 2 ) . Under the strategy profile, the loss suffered by the defender was 3750, and the gain gained by the planner was 5633. The game tree is shown in Figure 2.
When the planner first commits to a signaling strategy, the planner with payment type θ 1 and the attacker with type φ 1 will mislead the defender into thinking that their payment type is θ 2 by sending signals, causing the defender to mistake target t 1 as the most valuable to them, and will thus choose to implement a high-level defense against target t 1 . In this case, the planner will choose to attack the target t 2 so that it benefits and the defender loses. Similarly, the planner with payment type θ 2 and an attacker of type φ 2 misleads the defender into thinking that their attacker type is φ 1 by sending signals, making them gain and causing the defender to suffer losses. With the signaling strategy, the defender ultimately suffered a loss of 4022.5, while the planner made a gain of 6623.5. The game tree is given in Figure 3. It is observed that the outcomes for both parties are more pronounced when a signaling strategy is employed. This instance illustrates that the planner’s deception can lead to increased benefits for themselves, simultaneously resulting in more significant detriments for the defender.

4. Equilibrium Computation

In the introduced model, it is acknowledged that the planner’s strategy involves the transmission of two distinct signals, each necessitating individual calculation. It substantially increases the complexity of the resolution procedure. To address this issue, we propose a model simplification. Then, a tri-level programming framework is formulated based on the game process, and a decomposition algorithm is subsequently applied to solve it.

4.1. Modified Model

We propose a model simplification by considering all possible combinations of the two signals. This approach not only reduces the computational burden but also maintains the strategic essence of the signaling process, ensuring that the model remains a faithful representation. Specifically, the planner has m · M possible combinations of types, so the set of types can be reformulated into T Y = { t y θ , φ = ( θ , φ ) | θ Θ , φ Φ } . Then, the set of all types with payment type θ is represented as T Y θ = { t y θ , φ = ( θ , φ ) | φ Φ } and the set of all types with attacker type φ is represented as T Y φ = { t y θ , φ = ( θ , φ ) | θ Θ } . The corresponding prior probability is p t y : T Y [ 0 , 1 ] , where p t y ( t y θ , φ ) = p p t ( θ ) · p θ ( φ ) . Rewrite the set of signals sent by the planner as S = { s = ( s 1 , s 2 ) | s 1 S 1 , s 2 S 2 } . Denote the set of signals for each type t y θ , φ as S t y θ , φ = { s = ( s 1 , s 2 ) | s 1 S 1 , s 2 S 2 , φ } , and the corresponding mixed signal strategy set is Δ o t y θ , φ = { o = ( o s ) s S t y θ , φ | o s [ 0 , 1 ] , | | o s | | 1 = 1 } = { o = ( o s ) s S | o s [ 0 , 1 ] , | | o s | | 1 = 1 , o s = 0 s 2 > | | φ | | 1 } ; set Δ o = t y T Y Δ o t y θ , φ . With a slight abuse of the notation, denote the modified signaling strategy as π o : T Y Δ o , and π o ( s | t y θ , φ ) is the probability that the planner with type t y θ , φ sends the signal s. Then, accordingly, the defender’s strategy scheme is π d : S S D , where π d ( s ) = ( π d ( 1 | s ) , , π d ( N | s ) ) is the defender’s strategy after receiving the signal s, and π d ( j | s ) { 0 , 1 , 2 } represents the defense level in the j-th target. The planner’s attack scheme is π a : T Y × S S A , where π a ( t y θ , φ , s ) = ( π a ( i j | t y θ , φ , s ) ) i [ | | φ | | 1 ] , j [ N ] is the attack strategy after the planner with type t y θ , φ sends the signal s and π a ( i j | t y θ , φ , s ) { 0 , 1 } . After receiving the signal s, the defender’s belief that the planner type is t y θ , φ is redenoted as μ ( t y θ , φ | s ) = p t y ( t y θ , φ ) · π o ( s | t y θ , φ ) t y T Y p t y ( t y ) · π o ( s | t y ) . Then, combined with Equations (2), (4) and (10), the total utility function of the planner is   
U t o ( π o , π d , π a ) = t y θ , φ T Y p t y ( t y θ , φ ) s S t y θ , φ π o ( s | t y θ , φ ) · ( j = 1 N q j ( π d ( s ) , π a ( t y θ , φ , s ) ) v j θ i = 1 | | φ | | 1 U i t y θ , φ i n d ( s 1 , θ ) · d c 1 i n d ( s 2 , φ ) · d c 2 ) ,
where U i t y θ , φ is the income of the i-th attacker hired by the planner with type t y θ , φ ; that is
U i t y θ , φ = u i φ + j = 1 N π a ( i j | t y θ , φ , s ) q j ( π d ( s ) , π a ( t y θ , φ , s ) ) · α v j θ π a ( t y θ , φ , s ) j 3 j = 1 N π a ( i j | t y θ , φ , s ) a c i j φ .
Combined with Equations (1) and (8), the total loss function of the defender is
L t o ( π o , π d , π a ) = s S t y θ , φ T Y π o ( s | t y θ , φ ) · t y θ , φ T Y : | | φ | | 1 s 2 μ ( t y θ , φ | s ) j = 1 N ( q j ( π d ( s ) , π a ( t y θ , φ , s ) ) v j d + c π d ( j | s ) + 1 ) .

4.2. Tri-Level Programming

In this section, based on the game process, we delineate the D-APD model as a tri-level programming framework that encapsulates the strategic interactions between an attack planner and a defender, as depicted in Figure 4. The following discussion provides an in-depth exposition of this tri-level programming framework.

4.2.1. Signaling-Level Problem

At the beginning of the game, the planner chooses a signaling strategy to confuse the defender and thus make their own gains greater. Therefore, the first level is the signaling-level (SL), where the planner chooses a signal strategy designed to mislead the defender’s decision-making process to achieve greater benefits. Formally, the SL problem optimizes the planner’s utility and finds the corresponding optimal signaling strategy and is denoted by P 1 :   
P 1 : m a x π o U t o ( π o , π d * , π a * ) ( 1 a ) s.t. π o ( s | t y θ , φ ) [ 0 , 1 ] , s S , t y θ , φ T Y ( 1 b ) s s π o ( s | t y θ , φ ) = 1 , t y θ , φ T Y ( 1 c ) π o ( s | t y θ , φ ) = 0 f o r s 2 > | | φ | | 1 , t y θ , φ T Y ( 1 d )
In P 1 , the objective function (1a) represents the planner’s objective to maximize their utility. The decision is to determine the probability of sending each signal. Constraint (1b) and (1c) ensure the validity of the decision variable. Constraint (1d) illustrates the feasibility of the decision variable, that is, the planner can only send the signal with a smaller number of attackers than the actual number of hired attackers.

4.2.2. Defense-Level Problem

Upon receiving the signal from the planner, the defender updates their beliefs in light of the received signals, and accordingly, chooses a defense strategy, aiming to minimize their loss. Thus, the second level, designated as the defense-level (DL), is designed to identify a defense strategy for the defender that minimizes the potential loss. This level is mathematically formalized as P 2 :
P 2 : m i n π d L t o ( π o * , π d , π a * ) ( 2 a ) s.t. j = 1 N c π d ( j | s ) + 1 B ( 2 b ) π d ( j | s ) { 0 , 1 , 2 } , j [ N ] , s S ( 2 c )
In P 2 , the objective function (2a) is to minimize the defender’s loss. Constraint (2b) indicates that the defense cost cannot exceed its budget. Constraint (2c) sets a ternary restriction on defense decisions.

4.2.3. Attack-Level Problem

After observing the defense strategy, the planner arranges for the attackers to attack a subset of targets and the game ends. Consequently, the third and final level, known as the attack-level (AL), involves the planner deploying the attackers to execute an attack strategy that maximizes the utility. Specifically, in the AL problem, the decision is to decide the specific target for each attacker. It is essential to consider the individual benefits of each attacker simultaneously in formulating the attack strategy, ensuring that the strategy aligns with the Pareto optimal solutions within the multi-objective framework defined by the attackers’ benefits. This problem is formalized as P 3 :
P 3 : m a x π o U t o ( π o * , π d * , π a ) ( 3 a ) s.t. π a ( t y θ , φ , s ) P a r S A t y θ , φ , s S ( 3 b ) j = 1 N π a ( i j | t y θ , φ , s ) = 1 , i [ | | φ | | 1 ] , s S , t y θ , φ T Y ( 3 c ) π a ( i j | t y θ , φ , s ) { 0 , 1 } , i [ | | φ | | 1 ] , j [ N ] , s S , t y θ , φ T Y ( 3 d )
In P 3 , the objective function (3a) is to maximize the planner’s payoffs. Constraint (3c) states that each attacker can only attack one target. Constraint (3d) sets a binary restriction on attack decisions. Constraint (3b) guarantees a profit for each attacker. Formally, for the planner with type t y θ , φ , let g k ( π a ) = U k t y θ , φ ( π o , π d , π a ) , k = 1 , , | | φ | | 1 ; then, P a r S A t y θ , φ is the Pareto optimal set for the following multi-objective problem P 4 .
P 4 : m a x π a g = ( g 1 ( π a ) , , g k ( π a ) , , g | | φ | | 1 ( π a ) ) ( 4 a ) s.t. j = 1 N π a ( i j | t y θ , φ , s ) = 1 , i [ | | φ | | 1 ] , s S ( 4 b ) π a ( i j | t y θ , φ , s ) { 0 , 1 } , i [ | | φ | | 1 ] , j [ N ] , s S ( 4 c )

4.3. Algorithm Description

Solving tri-level programming problems is inherently complex, with even the most straightforward scenarios being NP-hard [60]. The extant algorithms for addressing these problems are primarily divided into exact and heuristic methods. Exact algorithms encompass implicit enumeration [61,62], reformation techniques [63] and decomposition algorithms [37,64]. However, implicit enumeration is limited to very small-scale problems. The reformation technique is tailored to problems with continuous lower-level decision variables, which does not align with the discrete variables present in our model. Consequently, the decomposition algorithm has garnered widespread application in research due to its effectiveness. The crux of employing this algorithm lies in its successful decomposition of the problem. Heuristic algorithms are also favored for their swift computation times [65]. But it may yield suboptimal solutions. So, the column and constraint generation algorithm (C&CG) has been introduced as an extension of decomposition algorithms [66,67]. While previous C&CG algorithms are tailored to solve min–max–min or max–min–max problems, they are not directly applicable to the model presented in this paper. In response, we extend the decomposition algorithm and devise a customized decomposition algorithm based on a genetic algorithm (DA-GA) for the tri-level programming problem outlined herein. Our proposed algorithm operates within a master-subproblem framework. The master problem (MP) is tasked with determining the signal strategy, while the subproblem (SP) is designed to extract a defense strategy for a given signal strategy. We will now delve into the construction and computation of both the master problem and subproblem, as well as the intricate process of our customized DA-GA. Given that the resolution of the master problem and subproblem is contingent upon the solution of the lowest attack level problem, we initially present a detailed methodology for addressing the attack level problem in Section 4.3.1, prior to discussing the master problem and subproblem in depth.

4.3.1. Attack-Level Algorithm

The implementation steps of the AL problem’s algorithm are demonstrated in Algorithm 1. In fact, the essence of resolving the attack-level problem P 3 lies in addressing the multi-objective problem P 4 . Given the vast number of strategy combinations available to planners of various types, a direct computational approach would incur significant computational expense. However, since the problem-solving process for planners of different types is analogous and their decisions are mutually independent, we can substantially mitigate computational costs by addressing each type of planner separately. Furthermore, the dimensionality of the attack scheduling strategy space escalates exponentially as the number of hired attackers increases, making it impractical to identify all Pareto optimal solutions by exhaustively examining each potential attack strategy. Leveraging the effectiveness of the NSGA-II algorithm in tackling multi-objective problems, this paper introduces a genetic algorithm (GA) tailored to problem P 4 . Taking the planner with type t y θ , φ as an example, when they send signal s, the specific steps are as follows:
0.
Code: The planner’s attack scheduling strategy is a 0 1 matrix with size | | φ | | 1 × N , so we first need to encode it. Specifically, we encode each attack scheduling strategy as an | | φ | | 1 -dimensional row vector with elements in set [ N ] . For example, a strategy 1 0 0 0 0 0 0 0 0 1 0 0 1 0 0 can be encoded as ( 1 , 5 , 3 ) , which represents that the 1-th attacks the 1-th target, the 2-th attacks the 5-th target and the 3-th attacks the 3-th target. In fact, the number of columns in each row of the policy matrix with the value of 1 is taken as the encoding of the policy.
1.
Determine an initial population P consisting of n individuals that meet the constraints;
2.
Calculate the fitness f = ( f 1 , f 2 ) of each individual in P;
(1)
For f 1 :
  • Calculate the function value for each individual on each objective;
  • For each individual p, calculate the number of individuals n p that dominate it and the set of individuals S p that it dominates;
  • Set the f 1 of the individuals with n p = 0 to 1, and the set of these individuals to F 1 , then subtract the n p values of the individuals in S p by 1; then, set the f 1 of the individuals with n p = 0 at this time to 2, the set of these individuals to F 2 , and so on until the f 1 of each individual is calculated.
(2)
For f 2 , solve it for each individual in F l in ( 1 ) separately:
  • For each individual i in F l , let i’s crowding distance L [ i ] d = 0 ;
  • Calculate the m-th function value for each individual and sort them in ascending order according to these values;
  • In order to give individuals on the edge a selection advantage, let L [ 1 ] d = L [ e n d ] d = i n f ; that is, after sorting, let the crowding distance of the first and last individuals be infinity. Then for the other individuals i = 2 , , | F l | 1
    L [ i ] d = L [ i ] d + ( L [ i + 1 ] m L [ i 1 ] m ) / ( f m m a x f m m i n )
    where L [ i + 1 ] m and L [ i 1 ] m represent the m-th objective function value for the ( i + 1 ) -th and ( i 1 ) -th individuals, and f m m a x and f m m i n represent the maximum and minimum values of the m-th objective function of the individuals in F l , respectively;
  • The above three steps are repeatedly calculated for each objective function, and the crowding distance obtained when each individual corresponds to different objective functions is added up to obtain the final crowding distance, which is denoted as f 2 .
3.
By selecting, crossing and mutating for P, a new population Q is generated;
(1)
Selection operator: The selection is based on fitness using the binary tournament method. Randomly select two individuals for comparison, select the better individual to enter the next generation and repeat this operation until we select enough n individuals where an individual p is better than an individual q, if, and only if, the f 1 of p is less than q or the f 1 of p is equal to q and the f 2 of p is greater than q. That is, compare the rank first; the individual with a lower rank is better, individuals with an equal rank have a comparative crowding distance, and individuals with larger crowding distances are better;
(2)
Crossover operator: For two individuals in the population, a location is randomly selected and the genes at that location are exchanged with a certain probability;
(3)
Mutation operator: In order to ensure the diversity of the population and avoid falling into the local optimal situation, the individuals in the population are randomly selected in a position and the genes at that position are changed.
4.
According to step 2, calculate the fitness of each individual in P Q ;
5.
Selection* operator: Select the n individuals in P Q as the offspring; if there is an l, that makes i = 1 l | F i | = n true, so P = i = 1 l F i is the offspring population. Otherwise, when l satisfies the conditions i = 1 l 1 | F i | < n and i = 1 l | F i | > n , the individuals in F l are sorted in descending order of crowding distance ( f 2 ) , and then the first ( n i = 1 l 1 | F i | ) individuals are selected to form offspring P together with i = 1 l 1 F i ;
6.
Return to step 3 and loop until the termination condition is satisfied, and finally obtain a Pareto optimal solution set.
Algorithm 1: GA for AL problem
Mathematics 12 02532 i001

4.3.2. Master Problem

First, we construct and solve a master problem by considering a subset of the defense strategy schemes to obtain a signaling strategy. Specifically, we give a set of defense strategies Π ^ d = { π ^ d 1 , , π ^ d h , , π ^ d H } , where h represents the iteration index. For each π ^ d h , we define a attack scheme π a h . Then the master problem can be formulated as follows:
P 5 : m a x π o , π a h U t o ( π o , π ^ d h , π a h ) ( 5 a ) s.t. π o ( s | t y θ , φ [ 0 , 1 ] , s S , t y θ , φ T Y ) ( 5 b ) s S π o ( s | t y θ , φ ) = 1 , t y θ , φ T Y ( 5 c ) π o ( s | t y θ , φ ) = 0 f o r s 2 > | | φ | | 1 , t y θ , φ T Y ( 5 d ) π a h ( t y θ , φ , s ) P a r S A t y θ , φ , s S , h [ H ] ( 5 e ) j = 1 N π a h ( i j | t y θ , φ , s ) = 1 , i | | φ | | 1 , s S , t y θ , φ T Y , h [ H ] ( 5 f ) π a h ( i j | t y θ , φ , s ) { 0 , 1 } , i | | φ | | 1 , j [ N ] , s S , t y θ , φ T Y , h [ H ] ( 5 g )
The planner’s signaling strategy is crafted specifically to mislead the defender, and as such, it exerts a direct influence on the defense strategy within the middle-level problem. Then problem P 5 effectively translates into a linear programming problem, in which, the dimensions of columns and constraints are finite and the constraints (5e) can be effectively addressed using Algorithm 1. So the master problem can be solved directly in polynomial time by using a professional linear programming solver.

4.3.3. Subproblem

The subproblem is centered on determining the defender’s strategy in response to a predefined signaling strategy π o * . It is important to recognize that the subproblem is concerned solely with deriving the defender’s strategy. The defender’s eventual loss, the corresponding attack strategy and the planner’s utility are all contingent upon the defender’s potentially incorrect perception of the planner’s intentions. These outcomes, which are based on this misaligned cognition, require further analysis and are addressed in the attack-level problem. Then, the subproblem is constructed as the following bilevel problem:
P 6 : m i n π d L t o ( π o * , π d , π a ) ( 6 a ) m a x π a U t o ( π o * , π d , π a ) ( 6 b ) s.t. j = 1 N c π d ( j | s ) + 1 B ( 6 c ) π a ( t y θ , φ , s ) P a r S A t y θ , φ , s S ( 6 d ) j = 1 N π a ( i j | t y θ , φ , s ) = 1 , i [ | | φ | | 1 ] , s S , t y θ , φ T Y ( 6 e ) π d ( j | s ) { 0 , 1 , 2 } , j [ N ] , s S ( 6 f ) π a ( i j | t y θ , φ , s ) { 0 , 1 } , i [ | | φ | | 1 ] , j [ N ] , s S ( 6 g )
This problem’s complexity arises from the vast number of possible defense strategy combinations for each signal, and the dimension of the defense strategy space grows exponentially with an increasing number of targets. To address this, we employ a genetic algorithm to find solutions for each signal that the defender might receive. Each defense strategy is represented as an N-dimensional row vector with elements from the set { 0 , 1 , 2 } , and these vectors are used directly as chromosomes without the need for additional encoding. The defender’s objective is to minimize their loss; that is, an individual with a lower loss function value would conventionally be considered more advantageous. Thus, we instead define the fitness of each individual as f = j = 1 N v j d L . Here, L represents the objective function value that corresponds to the defender’s strategy under misperception. It is essential to note that during the algorithm’s iterative process, we must evaluate whether the individuals in each new generation adhere to the budget constraints. This is a critical step to ensure that the solutions remain feasible within the given parameters. With the signaling scheme π o * as a given, the detailed steps of the algorithm are provided in Algorithm 2.
Algorithm 2: GA for SP
Mathematics 12 02532 i002

4.3.4. Customized Decomposition Algorithm Based on Genetic Algorithm

To solve the tri-level programming problem in this paper, we propose a customized decomposition algorithm based on a genetic algorithm. Since the goal of the algorithm is to maximize the utility of the planner, first make P ̲ = starting with Π ^ d = , representing that the defender does not defend against any target, and then make the utility greater by comparing the utility with P ̲ after each iteration. At each iteration, solve the MP for the defense scheme Π ^ d and obtain a signal scheme π ^ o . Then for this π ^ o , solve SP using Algorithm 2; obtain the defense scheme π ^ d h and let Π ^ d = Π ^ d π ^ d h . Subsequently, solve the AL problem using Algorithm 1 and obtain the objective function value o b j A L and corresponding attack scheme π ^ a . Compare o b j A L and P ̲ ; if o b j A L is larger, update π o * , π d * , π a * , P ̲ to π ^ o , π ^ d h , π ^ a , o b j A L , respectively, otherwise do not update. Then, add a variable and a corresponding set of constraints to MP, and continue solving until the iteration stops. It is important to note that the π o * , π d * , π a * is the final strategy scheme, and U is given by P ̲ , but L needs to solve L t o ( π o * , π d * , π a * ) for the strategy profile ( π o * , π d * , π a * ) , and cannot be directly given by the objective function value of SP. This is because the SP is carried out under the wrong cognition formed by the defender after receiving the deception signal of the planner; the attack scheme is not the real strategy scheme taken by the planner. The specific steps are shown in Algorithm 3.
Algorithm 3: Customized DA-GA
Mathematics 12 02532 i003

5. Experiment and Analysis

In this section, we analyze the model through some numerical experiments. Suppose that the value of each target to both sides is divided into three levels according to their importance to both sides, and the specific values are 10,000, 15,000 and 25,000. v d and v Θ are randomly selected among all combinations of the three levels. Let b s = ( 350 , 200 ) and α = 0.01 . Each element in the attacker’s cost matrix a c takes a random value from [ 0 , 50 ] .

5.1. Algorithm Performance

In this section, we evaluate the computational performance of the customized DA-GA. Figure 5 illustrates that the algorithm is finitely convergent and influenced by the budget constraints and the number of targets. Specifically, the number of iterations increases as the budget increases, as evidenced by comparing (a) with (c) and (b) with (d) in Figure 5. This trend is attributed to the expansion of the strategy space with an increased budget, complicating the search for the optimal strategy and consequently increasing the iterations needed for convergence. Similarly, an increase in the number of targets also leads to a higher number of iterations, as demonstrated by the comparison between (a) and (b), and (c) and (d) in Figure 5.
Then, we have compared the runtime of the DA-GA algorithm introduced in this paper against the Enumeration Algorithm (EA). The comparative results are illustrated in Figure 6. It is observed that while the EA holds a marginal advantage in scenarios with a small number of targets, the DA-GA algorithm consistently outperforms EA when the number of targets exceeds four. The performance gap becomes particularly pronounced as the number of targets increases. For instance, when the target count reaches six, the EA’s execution time exceeds 500 min, demonstrating the computational inefficiency of EA at higher target volumes. In contrast, the DA-GA algorithm exhibits a significant performance advantage under these conditions. This comparative analysis underscores the efficiency of the DA-GA in solving complex problems involving multiple targets.
Furthermore, to validate the accuracy of the algorithm, we have compared its computational results with the exact algorithm, as depicted in the Table 3. The findings indicate that the error margin of the DA-GA algorithm is confined within a narrow range.

5.2. The Impact of Defense Budgets

The impact of defense budgets on the planner’s utility is analyzed and presented in Figure 7. With an increase in the defense budget, the planner’s utility is observed to decrease in a stepwise pattern. This inverse relationship is logical, given that a larger defense budget enables broader target protection, which in turn lowers the probability of a successful attack and thus reduces the planner’s utility. Moreover, within a certain range of the defense budget, minor budget variations lead to negligible changes in the planner’s utility. This phenomenon occurs as minor budget adjustments do not substantially change the defender’s strategy. For example, a defense budget of 150 or 200 units allows the defender to provide only low-level protection for a single target. This analysis highlights the critical role of budget allocation in defense strategy.

5.3. The Impact of Signaling Strategy

Following the results from Section 5.2, in this section, we calculate the utility of the attack planner when the defense budget is 300, 450, 600, 750 and 900, respectively, with and without How to solve without the signaling strategy is in the Appendix A. the signaling strategy, and subsequently analyze the impact of the signaling strategy on the utility. For each subfigure (aa) in Figure 8a–e, we set m = 2 , M = 3 and for each subfigure (bb) in Figure 8a–e, we set N = 5 . The findings indicate that the utility of the planner employing a signaling strategy exceeds that of the planner not employing such a strategy. This suggests a propensity for the planner to deceive the defender in order to reap greater benefits. Additionally, each (aa) in Figure 8 shows that an increase in the number of targets causes the planner’s utility to increase. This increase is logical, given that a constant defense budget implies that the defender cannot effectively protect all targets, facilitating the planner’s selection of high-value targets for attack to maximize their utility. Furthermore, each (bb) in Figure 8 exhibits a more pronounced difference in the planner’s utility between scenarios with and without signaling strategies compared to corresponding (aa). This suggests that signaling strategies are more responsive to the diversity of payment types than to the number of targets. With an increased number of payment types, the defender’s uncertainty about the planner’s intentions escalates, enabling the planner to exploit this uncertainty through signaling and to enhance the signal strategy’s effectiveness.

5.4. The Impact of Prior Probability

Furthering our investigation, we examine the impact of prior probability on the planner’s utility. To facilitate the analysis, we focus on a scenario involving solely two types of planners. We propose the following theorem:
Theorem 1. 
When there are two types of attack planner, there is a linear relationship between their utility and the prior probability of each type (the proof of this theorem is in the Appendix B).
We proceed to validate Theorem 1 through experimentation. The number of targets was fixed at five. Utilizing the outcomes from Section 5.2, we analyzed how the planner’s utility varied with changes in the prior probability across defense budgets of 300, 450, 600, 750 and 900, respectively. The findings are presented in Figure 9. The results demonstrate a linear decrease in the planner’s utility with an increase in prior probability, consistent with the predictions of Theorem 1. This trend is attributed to the reduction in the defender’s uncertainty about the planner’s type due to an elevated prior probability. Consequently, the defender protects the targets with increased precision, resulting in a corresponding decrease in the planner’s utility.

5.5. The Impact of Deception Cost

Our analysis further explores the influence of deception costs on signaling strategies, with a summary of the findings presented in Table 4. The data indicate a positive correlation between the cost of deception and the planner’s tendency to report their types truthfully. As the deception cost increases, the probability of the planner reporting their types truthfully also rises. Furthermore, the analysis indicates that the relative magnitude of deception costs associated with payment and attacker types significantly influences the planner’s signaling strategy. When the deception cost is lower for the payment type than for the attacker type, the planner is more inclined to deceive the defender regarding the payment type. Conversely, the planner is more likely to mislead the defender about the attacker type.

6. Conclusions

This paper models the interaction between an attack planner and a defender, in which the attack planner hires some attackers and arranges them to attack the target. To ensure the attackers’ compliance with the arrangement, we safeguard their interests by incorporating a multi-objective problem into our model. In addition, considering that defenders may not fully grasp various details about the attack planner, which could be exploited using the attack planner to induce misjudgment and ineffective defense strategies, we develop a D-APD signaling game model to make up for the deficiencies in prior research on this type of attack. To address the model, we formulate it as a tri-level programming framework according to the game process, and propose a customized DA-GA for its computation. Then, we validate the algorithm’s effectiveness through experimental validation. Moreover, by assessing the benefits of an honest planner, we determine that the planner is more likely to opt for deception to secure greater benefits. Hence, there is an imperative to investigate effective defenses against these attacks. Finally, through controlled variable experiments, we establish that the defense budget can significantly curtail the attack planner’s incomes, which further verify the significance of budgetary considerations in defense strategy.
We proceed under the assumption that attackers will adhere to the directives of the planner. However, in reality, hired attackers may defy the planner’s attack arrangements due to inadequate compensation or financial incentives from the defenders, among other factors, introducing a layer of complexity. Consequently, our future research will delve into this issue. Furthermore, despite our detailed examination of a specific attack behavior, the critical aspect lies in our response to it. Thus, the implementation of effective defenses by defenders in response to such complex attacks remains a central focus of our future research, and is the paramount objective of our study on attack behaviors.

Author Contributions

Conceptualization, S.H. and M.L.; methodology, S.H.; validation, S.H. and R.Z.; formal analysis, S.H.; investigation, S.H., R.Z. and M.L.; writing—original draft preparation, S.H.; writing—review and editing, S.H.; supervision, M.L. and R.Z.; funding acquisition, M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Nature Science Foundation of China under grant number: T2350710232.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

When there is no signal strategy, the model is a Stackelberg game; therefore, the strategies and utility of both the attack planner and the defender are solved using the following problem P7:
P 7 : m i n s d j = 1 N ( q j ( s d , s a * ) v j d + c s d j + 1 ) ( A 1 ) s.t. j = 1 N q j ( s d , s a * ) v j θ i = 1 | | φ | | 1 ( u i + j = 1 N s a i j * q j ( s d , s a * ) · α v j θ s a j * 3 j = 1 N s a i j * a c i j ) j = 1 N q j ( s d , s a ) v j θ i = 1 | | φ | | 1 ( u i + j = 1 N s a i j q j ( s d , s a ) · α v j θ s a j 3 j = 1 N s a i j a c i j ) , s a P a r S A θ , φ ( A 2 ) j = 1 N c s d j + 1 B ( A 3 ) s d j { 0 , 1 , 2 } , j [ N ] ( A 4 )
This is a discrete planning problem, and in fact, it is equivalent to the SP in which the signal strategy is an identity matrix, i.e., the attack planner sends a truthful signal about its own type. So, we can solve it through solve SP(E) (E represents an identity matrix), using Algorithm 2.

Appendix B

Proof of Theorem 1. 
Suppose that there are two types of attack planner, that is t y 1 , t y 2 ; the prior probability of t y 1 is p 1 , so the probability of t y 2 is 1 p 1 . Then, the attack planner’s total utility is
U = p 1 · s S t y 1 π o ( s | t y 1 ) · ( j = 1 N q j ( π d ( s ) , π a ( t y 1 , s ) ) v j a 1 i = 1 | | n 1 | | 1 U i t y 1 i n d ( s 1 , t y 1 1 ) · d c 1 i n d ( s 2 , t y 1 2 ) · d c 2 ) + ( 1 p 1 ) · s S t y 2 π o ( s | t y 2 ) · ( j = 1 N q j ( π d ( s ) , π a ( t y 2 , s ) ) v j a 2 i = 1 | | n 2 | | 1 U i t y 2 i n d ( s 1 , t y 2 1 ) · d c 1 i n d ( s 2 , t y 2 2 ) · d c 2 ) = p 1 · ( s S t y 1 π o ( s | t y 1 ) · ( j = 1 N q j ( π d ( s ) , π a ( t y 1 , s ) ) v j a 1 i = 1 | | n 1 | | 1 U i t y 1 i n d ( s 1 , t y 1 1 ) · d c 1 i n d ( s 2 , t y 1 2 ) · d c 2 ) s S t y 2 π o ( s | t y 2 ) · ( j = 1 N q j ( π d ( s ) , π a ( t y 2 , s ) ) v j a 2 i = 1 | | n 2 | | 1 U i t y 2 i n d ( s 1 , t y 2 1 ) · d c 1 i n d ( s 2 , t y 2 2 ) · d c 2 ) ) + s S t y 2 π o ( s | t y 2 ) · ( j = 1 N q j ( π d ( s ) , π a ( t y 2 , s ) ) v j a 2 i = 1 | | n 2 | | 1 U i t y 2 i n d ( s 1 , t y 2 1 ) · d c 1 i n d ( s 2 , t y 2 2 ) · d c 2 ) .
From this equation, we can obtain that U is linearly related to p 1 . □

References

  1. van Stackelberg, H. Marktform und Gleichgewicht; Springer: Berlin, Germany, 1934. [Google Scholar]
  2. Page, F.H. Optimal contract mechanisms for principal-agent problems with moral hazard and adverse selection. Econ. Theory 1991, 1, 323–338. [Google Scholar] [CrossRef]
  3. Gan, J.; An, B.; Wang, H.; Sun, X.; Shi, Z. Optimal pricing for improving efficiency of taxi systems. In Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence, Beijing, China, 3–9 August 2013. [Google Scholar]
  4. Carnes, T.; Nagarajan, C.; Wild, S.M.; van Zuylen, A. Maximizing influence in a competitive social network: A follower’s perspective. In Proceedings of the Ninth International Conference on Electronic Commerce, Minneapolis, MN, USA, 19–22 August 2007; pp. 351–360. [Google Scholar]
  5. Li, Y.; Conitzer, V. Game-theoretic question selection for tests. J. Artif. Intell. Res. 2017, 59, 437–462. [Google Scholar] [CrossRef]
  6. Tambe, M. Security and Game Theory: Algorithms, Deployed Systems, Lessons Learned; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  7. Brown, M.; Sinha, A.; Schlenker, A.; Tambe, M. One size does not fit all: A game-theoretic approach for dynamically and effectively screening for threats. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. [Google Scholar]
  8. Fang, F.; Nguyen, T.H. Green security games: Apply game theory to addressing green security challenges. ACM SIGecom Exch. 2016, 15, 78–83. [Google Scholar] [CrossRef]
  9. Fang, F.; Stone, P.; Tambe, M. When Security Games Go Green: Designing Defender Strategies to Prevent Poaching and Illegal Fishing. IJCAI 2015, 15, 2589–2595. [Google Scholar]
  10. Fang, F.; Nguyen, T.; Pickles, R.; Lam, W.; Clements, G.; An, B.; Singh, A.; Tambe, M.; Lemieux, A. Deploying paws: Field optimization of the protection assistant for wildlife security. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30, pp. 3966–3973. [Google Scholar]
  11. Zhang, C.; Gholami, S.; Kar, D.; Sinha, A.; Jain, M.; Goyal, R.; Tambe, M. Keeping pace with criminals: An extended study of designing patrol allocation against adaptive opportunistic criminals. Games 2016, 7, 15. [Google Scholar] [CrossRef]
  12. Kar, D.; Nguyen, T.H.; Fang, F.; Brown, M.; Sinha, A.; Tambe, M.; Jiang, A.X. Trends and applications in Stackelberg security games. In Handbook of Dynamic Game Theory; Springer: Cham, Switzerland, 2017; pp. 1–47. [Google Scholar]
  13. Delle Fave, F.M.; Jiang, A.X.; Yin, Z.; Zhang, C.; Tambe, M.; Kraus, S.; Sullivan, J.P. Game-theoretic patrolling with dynamic execution uncertainty and a case study on a real transit system. J. Artif. Intell. Res. 2014, 50, 321–367. [Google Scholar] [CrossRef]
  14. An, B.; Ordóñez, F.; Tambe, M.; Shieh, E.; Yang, R.; Baldwin, C.; DiRenzo, J., III; Moretti, K.; Maule, B.; Meyer, G. A deployed quantal response-based patrol planning system for the US Coast Guard. Interfaces 2013, 43, 400–420. [Google Scholar] [CrossRef]
  15. Leitmann, G. On generalized Stackelberg strategies. J. Optim. Theory Appl. 1978, 26, 637–643. [Google Scholar] [CrossRef]
  16. Von Stengel, B.; Zamir, S. Leadership with Commitment to Mixed Strategies; Technical Report; London School of Economics: London, UK, 2004. [Google Scholar]
  17. Letchford, J.; Conitzer, V.; Munagala, K. Learning and approximating the optimal strategy to commit to. In Algorithmic Game Theory: Proceedings of the Second International Symposium, SAGT 2009, Paphos, Cyprus, 18–20 October 2009; Proceedings 2; Springer: Berlin/Heidelberg, Germany, 2009; pp. 250–262. [Google Scholar]
  18. Kiekintveld, C.; Islam, T.; Kreinovich, V. Security games with interval uncertainty. In Proceedings of the Twelfth International Conference on Autonomous Agents and Multiagent Systems AAMAS’2013, Saint Paul, MN, USA, 6–10 May 2013. [Google Scholar]
  19. Nguyen, T.; Yadav, A.; An, B.; Tambe, M.; Boutilier, C. Regret-based optimization and preference elicitation for Stackelberg security games with uncertainty. In Proceedings of the AAAI Conference on Artificial Intelligence, Quebec City, QC, Canada, 27–31 July 2014; Volume 28. [Google Scholar]
  20. Conitzer, V.; Sandholm, T. Computing the optimal strategy to commit to. In Proceedings of the 7th ACM conference on Electronic Commerce, Ann Arbor, MI, USA, 11–15 June 2006; pp. 82–90. [Google Scholar]
  21. Paruchuri, P.; Pearce, J.P.; Marecki, J.; Tambe, M.; Ordonez, F.; Kraus, S. Playing games for security: An efficient exact algorithm for solving Bayesian Stackelberg games. In Proceedings of the 7th International Joint Conference on Autonomous Agents and Multiagent Systems, Estoril, Portugal, 12–16 May 2008; Volume 2, pp. 895–902. [Google Scholar]
  22. Jain, M.; Pita, J.; Tambe, M.; Ordónez, F.; Paruchuri, P.; Kraus, S. Bayesian Stackelberg games and their application for security at Los Angeles International Airport. ACM SIGecom Exch. 2008, 7, 1–3. [Google Scholar] [CrossRef]
  23. Blum, A.; Haghtalab, N.; Procaccia, A.D. Learning optimal commitment to overcome insecurity. Adv. Neural Inf. Process. Syst. 2014, 27, 1826–1834. [Google Scholar]
  24. Haghtalab, N.; Fang, F.; Nguyen, T.H.; Sinha, A.; Procaccia, A.D.; Tambe, M. Three strategies to success: Learning adversary models in security games. In Proceedings of the 25th International Joint Conference on Artificial Intelligence (IJCAI), New York, NY, USA, 9–15 July 2016. [Google Scholar]
  25. Roth, A.; Ullman, J.; Wu, Z.S. Watch and learn: Optimizing from revealed preferences feedback. In Proceedings of the Forty-Eighth Annual ACM Symposium on Theory of Computing, Cambridge, MA, USA, 19–21 June 2016; pp. 949–962. [Google Scholar]
  26. Peng, B.; Shen, W.; Tang, P.; Zuo, S. Learning optimal strategies to commit to. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 2149–2156. [Google Scholar]
  27. Gan, J.; Xu, H.; Guo, Q.; Tran-Thanh, L.; Rabinovich, Z.; Wooldridge, M. Imitative follower deception in stackelberg games. In Proceedings of the 2019 ACM Conference on Economics and Computation, Phoenix, AZ, USA, 24–28 June 2019; pp. 639–657. [Google Scholar]
  28. Gan, J.; Guo, Q.; Tran-Thanh, L.; An, B.; Wooldridge, M. Manipulating a learning defender and ways to counteract. Adv. Neural Inf. Process. Syst. 2019, 32, 8274–8283. [Google Scholar]
  29. Nguyen, T.H.; Wang, Y.; Sinha, A.; Wellman, M.P. Deception in finitely repeated security games. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 2133–2140. [Google Scholar]
  30. Nguyen, T.H.; Sinha, A.; He, H. Partial adversarial behavior deception in security games. In Proceedings of the 29th International Joint Conference on Artificial Intelligence, Virtual, 7–15 January 2021. [Google Scholar]
  31. Nguyen, T.H.; Sinha, A. Sequential Manipulative Attacks in Security Games. In Proceedings of the 20th International Conference on Autonomous Agents and Multiagent Systems, London, UK, 3–7 May 2021. [Google Scholar]
  32. Nguyen, T.H.; Sinha, A. The Art of Manipulation: Threat of Multi-Step Manipulative Attacks in Security Games. arXiv 2022, arXiv:2202.13424. [Google Scholar]
  33. Guo, Q.; An, B.; Bosanský, B.; Kiekintveld, C. Comparing Strategic Secrecy and Stackelberg Commitment in Security Games. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence IJCAI, Melbourne, Australia, 19–25 August 2017; pp. 3691–3699. [Google Scholar]
  34. Cheng, Z.; Chen, G.; Hong, Y. Single-leader-multiple-followers Stackelberg security game with hypergame framework. IEEE Trans. Inf. Forensics Secur. 2022, 17, 954–969. [Google Scholar] [CrossRef]
  35. Korzhyk, D.; Conitzer, V.; Parr, R. Security games with multiple attacker resources. In Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence, Barcelona, Spain, 16–22 July 2011. [Google Scholar]
  36. Li, Q.; Li, M.; Tian, Y.; Gan, J. A risk-averse tri-level stochastic model for locating and recovering facilities against attacks in an uncertain environment. Reliab. Eng. Syst. Saf. 2023, 229, 108855. [Google Scholar] [CrossRef]
  37. Ghorbani-Renani, N.; González, A.D.; Barker, K.; Morshedlou, N. Protection-interdiction-restoration: Tri-level optimization for enhancing interdependent network resilience. Reliab. Eng. Syst. Saf. 2020, 199, 106907. [Google Scholar] [CrossRef]
  38. Zhuang, J.; Bier, V.M. Reasons for secrecy and deception in homeland-security resource allocation. Risk Anal. Int. J. 2010, 30, 1737–1743. [Google Scholar] [CrossRef] [PubMed]
  39. Zhuang, J.; Bier, V.M. Secrecy and deception at equilibrium, with applications to anti-terrorism resource allocation. Def. Peace Econ. 2011, 22, 43–61. [Google Scholar] [CrossRef]
  40. Zhuang, J.; Bier, V.M.; Alagoz, O. Modeling secrecy and deception in a multiple-period attacker–defender signaling game. Eur. J. Oper. Res. 2010, 203, 409–418. [Google Scholar] [CrossRef]
  41. Zhang, C.; Ramirez-Marquez, J.E.; Wang, J. Critical infrastructure protection using secrecy—A discrete simultaneous game. Eur. J. Oper. Res. 2015, 242, 212–221. [Google Scholar] [CrossRef]
  42. Chen, L.; Li, M.; Liu, J. Modeling bluffing behavior in signaling security games. Int. Trans. Oper. Res. 2022, 29, 1825–1841. [Google Scholar] [CrossRef]
  43. Xu, H.; Freeman, R.; Conitzer, V.; Dughmi, S.; Tambe, M. Signaling in Bayesian Stackelberg Games. In Proceedings of the 15th International Conference on Au- tonomous Agents and Multiagent Systems AAMAS, Singapore, 9–13 May 2016; pp. 150–158. [Google Scholar]
  44. Xu, H.; Rabinovich, Z.; Dughmi, S.; Tambe, M. Exploring information asymmetry in two-stage security games. Proc. AAAI Conf. Artif. Intell. 2015, 29, 1057–1063. [Google Scholar] [CrossRef]
  45. Li, X.; Yi, S.; Sycara, K. Multi-agent deception in attack-defense stochastic game. In Distributed Autonomous Robotic Systems: Proceedings of the 15th International Symposium, Online, 1–4 June 2022; Springer: Cham, Switzerland, 2022; pp. 242–255. [Google Scholar]
  46. Nguyen, T.; Xu, H. When can the defender effectively deceive attackers in security games? In Proceedings of the AAAI Conference on Artificial Intelligence, Philadephia, PA, USA, 27 February–2 March 2022; Volume 36, pp. 9405–9412. [Google Scholar]
  47. Esmaeeli, S.; Hassanpour, H.; Bigdeli, H. Deception in multi-attacker security game with nonfuzzy and fuzzy payoffs. Iran. J. Numer. Anal. Optim. 2022, 12, 542–566. [Google Scholar]
  48. Wan, Z.; Cho, J.H.; Zhu, M.; Anwar, A.H.; Kamhoua, C.; Singh, M.P. Resisting multiple advanced persistent threats via hypergame-theoretic defensive deception. IEEE Trans. Netw. Serv. Manag. 2023, 20, 3816–3830. [Google Scholar] [CrossRef]
  49. Nguyen, T.H.; Butler, A.; Xu, H. Tackling Imitative Attacker Deception in Repeated Bayesian Stackelberg Security Games. In Proceedings of the ECAI, Santiago de Compostela, Spain, 29 August–8 September 2020; pp. 187–194. [Google Scholar]
  50. Long, N.V.; Sorger, G. A dynamic principal-agent problem as a feedback Stackelberg differential game. Cent. Eur. J. Oper. Res. 2010, 18, 491–509. [Google Scholar] [CrossRef]
  51. Caldentey, R.; Haugh, M. A Cournot-Stackelberg model of supply contracts with financial hedging and identical retailers. Found. Trends Technol. Inf. Oper. Manag. 2017, 11, 124–143. [Google Scholar] [CrossRef]
  52. Lin, Y.; Ren, Z.; Touzi, N.; Yang, J. Random horizon principal-agent problems. SIAM J. Control. Optim. 2022, 60, 355–384. [Google Scholar] [CrossRef]
  53. Behnk, S. On the Reduction of Deception in Principal-Agent Relationships. Ph.D. Thesis, Universitat Jaume I, Castello de la Plana, Spain, 2015. [Google Scholar]
  54. Duffy, R.; St John, F.A.; Büscher, B.; Brockington, D. The militarization of anti-poaching: Undermining long term goals? Environ. Conserv. 2015, 42, 345–348. [Google Scholar] [CrossRef]
  55. Gill, C.; Weisburd, D.; Telep, C.W.; Vitter, Z.; Bennett, T. Community-oriented policing to reduce crime, disorder and fear and increase satisfaction and legitimacy among citizens: A systematic review. J. Exp. Criminol. 2014, 10, 399–428. [Google Scholar] [CrossRef]
  56. Linkie, M.; Martyr, D.J.; Harihar, A.; Risdianto, D.; Nugraha, R.T.; Maryati; Leader-Williams, N.; Wong, W.M. Editor’s Choice: Safeguarding Sumatran tigers: Evaluating effectiveness of law enforcement patrols and local informant networks. J. Appl. Ecol. 2015, 52, 851–860. [Google Scholar] [CrossRef]
  57. Moreto, W.D. Introducing intelligence-led conservation: Bridging crime and conservation science. Crime Sci. 2015, 4, 15. [Google Scholar] [CrossRef]
  58. Smith, M.; Humphreys, J. The Poaching Paradox: Why South Africa’s ‘Rhino Wars’ Shine a Harsh Spotlight on Security and Conservation. In Environmental Crime and Social Conflict; Routledge: London, UK, 2016; pp. 197–220. [Google Scholar]
  59. Huang, T.; Shen, W.; Zeng, D.; Gu, T.; Singh, R.; Fang, F. Green security game with community engagement. arXiv 2020, arXiv:2002.09126. [Google Scholar]
  60. Bard, J.F. Some properties of the bilevel programming problem. J. Optim. Theory Appl. 1991, 68, 371–378. [Google Scholar] [CrossRef]
  61. Parajuli, A.; Kuzgunkaya, O.; Vidyarthi, N. Responsive contingency planning of capacitated supply networks under disruption risks. Transp. Res. Part Logist. Transp. Rev. 2017, 102, 13–37. [Google Scholar] [CrossRef]
  62. Parajuli, A.; Kuzgunkaya, O.; Vidyarthi, N. The impact of congestion on protection decisions in supply networks under disruptions. Transp. Res. Part E Logist. Transp. Rev. 2021, 145, 102166. [Google Scholar] [CrossRef]
  63. Lei, X.; Shen, S.; Song, Y. Stochastic maximum flow interdiction problems under heterogeneous risk preferences. Comput. Oper. Res. 2018, 90, 97–109. [Google Scholar] [CrossRef]
  64. Ghorbani-Renani, N.; González, A.D.; Barker, K. A decomposition approach for solving tri-level defender-attacker-defender problems. Comput. Ind. Eng. 2021, 153, 107085. [Google Scholar] [CrossRef]
  65. Fakhry, R.; Hassini, E.; Ezzeldin, M.; El-Dakhakhni, W. Tri-level mixed-binary linear programming: Solution approaches and application in defending critical infrastructure. Eur. J. Oper. Res. 2022, 298, 1114–1131. [Google Scholar] [CrossRef]
  66. Ding, T.; Yao, L.; Li, F. A multi-uncertainty-set based two-stage robust optimization to defender–attacker–defender model for power system protection. Reliab. Eng. Syst. Saf. 2018, 169, 179–186. [Google Scholar] [CrossRef]
  67. Fang, Y.P.; Zio, E. An adaptive robust framework for the optimization of the resilience of interdependent infrastructures under natural hazards. Eur. J. Oper. Res. 2019, 276, 1119–1136. [Google Scholar] [CrossRef]
Figure 1. Game process.
Figure 1. Game process.
Mathematics 12 02532 g001
Figure 2. Game tree without deception.
Figure 2. Game tree without deception.
Mathematics 12 02532 g002
Figure 3. Game tree with deception.
Figure 3. Game tree with deception.
Mathematics 12 02532 g003
Figure 4. The tri-level programming framework.
Figure 4. The tri-level programming framework.
Mathematics 12 02532 g004
Figure 5. Convergence of DA-GA.
Figure 5. Convergence of DA-GA.
Mathematics 12 02532 g005
Figure 6. Algorithm runtime.
Figure 6. Algorithm runtime.
Mathematics 12 02532 g006
Figure 7. The impact of defense budgets.
Figure 7. The impact of defense budgets.
Mathematics 12 02532 g007
Figure 8. The impact of signaling strategy.
Figure 8. The impact of signaling strategy.
Mathematics 12 02532 g008
Figure 9. The impact of prior probability.
Figure 9. The impact of prior probability.
Mathematics 12 02532 g009
Table 1. Legend of common symbols.
Table 1. Legend of common symbols.
SymbolDescription
Sets:
S D The set of defender’s pure strategy
S A The set of attack planner’s pure strategy
Θ The set of attack planner’s payment types
Φ The set of attacker types
S 1 The set of signals about payment type
S 2 The set of signal about attacker type
T Y The set of attack planner’s types
SThe set of signals
Δ o 1 The set of mixed signaling strategy about planner’s payoff type
Δ o 2 The set of mixed signaling strategy about planner’s attacker type
Parameters:
NThe number of targets
b s Attacker’s base salary vector
cDefender’s cost vector
BDefender’s budget constraint
nThe number of strong and weak attacker
v d Defender’s valuation of targets
v a Attack planner’s valuation of targets
s d Defender’s pure strategy
s a Attack planner’s pure strategy
q j The probability that the target j is attacked successfully
u i i-th attacker’s base salary
a c i j The i-th attacker’s cost when they attacks target j
α The ratio of the commission when attack is successful
p p t The prior probability distribution about payment types
v θ The target valuation of Attack planner with payment type θ
p θ The prior probability distribution about attacker type
μ p t The posterior probability about the payment types
μ a n The posterior probability about the attacker type
d c Attack planner’s deception cost
p t y The prior probability distribution of attack planner’s type
μ ( t y | s ) The probability that the attack planner’s type is t y after receiving the signal s
L D , U A P , U i The utility of defender, planner and the i-th attacker without deception
L t o , U t o , U i t y The utility of defender, planner and the i-th attacker with deception
Decision variables:
π o The signaling strategy
π d Defender’s strategy scheme
π a Attack planner’s attack scheme
Table 2. Motivation example.
Table 2. Motivation example.
VariableValue
N3
m2
p 1 [0.5 0.5]
M2
p 2 [0.5 0.5]
α 0.01
v d [5000 3000 2000]
Θ [2000 5000 2000;5000 5000 3000]
b s [50 30]
Φ [1 0;1 1]
u 1 50
u 2 [30 50]
a c 1 [2 1 3]
a c 2 [1.5 2 1;1 3 2]
d c [3 3]
c[0 20 30]
B40
Table 3. Effectiveness of DA-GA.
Table 3. Effectiveness of DA-GA.
B = 400B = 600
NEADA-GAEADA-GA
3 4.1894 × 10 4 4.1894 × 10 4 2.4163 × 10 4 2.4163 × 10 4
4 5.5653 × 10 4 5.5653 × 10 4 4.6396 × 10 4 4.6399 × 10 4
5 5.8032 × 10 4 5.8039 × 10 4 6.4190 × 10 4 6.4192 × 10 4
Table 4. The impact of deception costs on signaling strategy.
Table 4. The impact of deception costs on signaling strategy.
Type12345
d c = [0,50]1(1,1)(2,1)(1,1)(2,1)(2,1)
2(1,2)(4,2)(5,2)(5,2)(1,2)
3(1,3)(5,3)(3,3)(3,3)(1,3)
d c = [10,50]1(3,1)(5,1)(2,1)(5,1)(5,1)
2(1,2)(1,2)(3,2)(4,2)(5,2)
3(1,3)(2,3)(4,3)(4,3)(5,3)
d c = [20,50]1(1,1)(5,1)(5,1)(4,1)(5,1)
2(1,2)(2,2)(3,2)(4,2)(5,2)
3(1,3)(3,3)(3,3)(4,3)(5,3)
d c = [30,50]1(1,1)(1,1)(3,1)(4,1)(5,1)
2(1,2)(2,2)(3,2)(4,2)(5,2)
3(1,3)(2,3)(3,3)(4,3)(5,3)
d c = [40,50]1(1,1)(5,1)(3,1)(4,1)(5,1)
2(1,2)(2,2)(3,2)(4,2)(5,2)
3(1,3)(2,3)(3,3)(4,3)(5,3)
d c = [50,50]1(1,1)(5,1)(3,1)(4,1)(5,1)
2(1,2)(2,2)(3,2)(4,2)(5,2)
3(1,3)(2,3)(3,3)(4,3)(5,3)
d c = [60,50]1(1,1)(2,3)(3,1)(4,1)(5,1)
2(1,2)(2,2)(3,2)(4,2)(5,2)
3(1,3)(2,3)(3,3)(4,3)(5,3)
d c = [70,50]1(1,1)(2,3)(3,2)(4,1)(5,1)
2(1,2)(2,2)(3,2)(4,2)(5,2)
3(1,3)(2,3)(3,3)(4,3)(5,3)
d c = [80,50]1(1,1)(2,2)(3,1)(4,2)(5,3)
2(1,2)(2,2)(3,2)(4,2)(5,2)
3(1,3)(2,3)(3,3)(4,3)(5,3)
d c = [90,50]1(1,1)(2,3)(3,1)(4,1)(5,1)
2(1,2)(2,2)(3,2)(4,2)(5,2)
3(1,3)(2,3)(3,3)(4,3)(5,3)
d c = [100,50]1(1,1)(2,3)(3,1)(4,1)(5,1)
2(1,2)(2,2)(3,2)(4,2)(5,2)
3(1,3)(2,3)(3,3)(4,3)(5,3)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, S.; Li, M.; Zhang, R. Signaling Security Games with Attack Planner Deception. Mathematics 2024, 12, 2532. https://doi.org/10.3390/math12162532

AMA Style

He S, Li M, Zhang R. Signaling Security Games with Attack Planner Deception. Mathematics. 2024; 12(16):2532. https://doi.org/10.3390/math12162532

Chicago/Turabian Style

He, Santing, Mingchu Li, and Runfa Zhang. 2024. "Signaling Security Games with Attack Planner Deception" Mathematics 12, no. 16: 2532. https://doi.org/10.3390/math12162532

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop