1. Introduction
The wireless communication connection between physical components and network computing layers in cyber-physical systems (CPSs) introduces vulnerabilities in network attacks [
1,
2]. Consequently, there is a crucial need to thoroughly study and address the security issues associated with CPSs to enhance the security of CPSs and safeguard CPSs against cyber-attacks.
Thus, many researchers focus on improving the means of keeping CPSs secure. The secure state estimation (SSE) is the problem of estimating the system state from the compromised data packets and the noisy [
3]. To ensure an SSE under cyber-attacks, it is crucial to identify the information transmission channel affected when a portion of the collected information is tampered with. In the context of an SSE amidst cyber-attacks, the challenge lies in identifying the precise set of compromised channels among numerous possibilities, rendering the problem an instance of combinatorial optimization and falling within the realm of NP-hard problems [
4,
5,
6,
7,
8]. To solve such difficulties, the author devotes the
relaxation technique in reference [
4]. The author applies the ideas from reference [
4] to the event-triggered projected gradient descent algorithm [
5]. Note that the aforementioned algorithms [
4,
5] can only guarantee the correctness of the estimates under certain restrictive assumptions about the system structure. Reference [
6] investigates the impact of cyber-attacks, establishing that the stability of the observation error system is stable as long as the
-sparse observable condition of the original system is satisfied. The existing work provides fundamental limits for estimation against integrity attacks, specifically proving the necessity of
-sparse observable for SSE against
-sparse attacks. For example, reference [
7] analyzes the convergence of observation errors under
-sparse attacks using the separation principle and Lyapunov stability theory via a system based on
-sparse observable. However, it is worth noting that control systems are often nonlinear. In such complex systems, accurate SSE may require the use of complex models and algorithms, increasing computational complexity and cost.
Existing studies suggest that increasing secure/resilient control or prior knowledge can alleviate the damage caused by cyber-attacks. Through further research on the assumption that attacks can be parameterized, which is state-dependent, reference [
9,
10] relaxed the requirement for attacks to be bounded, allowing attacks to be unbounded. Reference [
11] introduces the coordinate transformations in controller design, which can overcome the negative impact caused by cyber-attacks. References [
12,
13] propose an attack compensator that effectively suppresses the impact of attacks by removing extreme values, thereby ensuring that normal nodes correctly estimate the system state under malicious node interference and avoid traversal search methods. Meanwhile, reference [
14] proposes a novel attack compensator to suppress the effect of the attack signal on state information by compressing the amplitude of abnormal measurement signals. Due to the challenge of unknown time-varying gains in the design of control systems after the parameterization of attacks, to solve it, the author proposes a Nussbaum gain technique [
15]. In addition, the authors propose an observer with attack compensation in [
15], which successfully solved the problem of unknown control gain caused by unknown time-varying coefficients. However, the assumption that attacks can be parameterized is too strict, and it is difficult to find conditions that satisfy this assumption in real physical systems.
As deep learning continues to be studied, the performance of NN models is constantly improving. Researchers have proposed many prediction methods [
16], making the application of NN more reliable and understandable. Therefore, in security maintenance and system performance evaluations, neural networks are usually used to estimate attack functions to maintain the security of the system [
17,
18,
19]. Therefore, the authors propose the NN estimation algorithm for the systems in [
17,
18], excluding the strict assumption of parameterization. At the same time, in reference [
19], the authors use nonlinear mapping to the cyber-attack function, applying it to a compact set, achieving the estimation requirements for the attack. In addition, an anti-attack estimation algorithm is also designed to suppress the impact of attack signals [
19]. Hence, developing new control algorithms against sensor and actuator attacks is of significant importance in both theory and practice. Therefore, the main challenge of adversarial attack recovery algorithms lies in how to achieve complete compensation or suppression of attack signals and ensure convergence to the equilibrium point in any situation (with or without attacks).
In this paper, we propose an NN adaptive secure control scheme under the NN estimation algorithm framework. Specifically, we consider the following situations:
In the above research on secure/resilient control [
9,
10,
11,
12,
13,
14,
15,
16], secure/resilient control strategies are designed under the assumption that attacks can be parameterized. If an attack cannot be parameterized, how can we design an attack reconstruction strategy to eliminate the adverse effects of the cyber-attack? Furthermore, compared to [
17,
18,
19], how can we utilize this insecure information to restore system security when attacking all channels?
The goal of this paper is to obtain the NN attack reconstruction strategies in an insecure network with all information unavailable to cope with the aforementioned situations. The innovations of this paper are as follows:
- (1)
We formulate the attack reconstruction algorithm in an insecure network with all information unavailable, where the attack reconstruction strategy involves the final solution of the estimation algorithm. Specifically, by introducing mapping, time is first mapped into a compact set. Then, we use the NN estimation algorithm to solve approximate learning for unknown adversarial attacks, thereby improving the ability to cope with the adversarial attacks. Note that, although [
9,
10,
11,
12,
13,
14,
15,
16] also investigates the problems of adversarial attacks, the adversarial attacks model in [
9,
10,
11,
12,
13,
14,
15,
16] requires that the adversarial attacks must be parameterized; this article removes this restrictive assumption and is more applicable to general attack models.
- (2)
We study the NN adaptive secure control scheme within the framework of backstepping, where the attack reconstruction algorithm is utilized to solve approximate learning for unknown adversarial attacks. We have demonstrated, based on the Lyapunov stability analysis theory, that the proposed secure control can ensure system performance even in the presence of adversarial attacks. Different from the requirements in the literature [
20,
21], and considering that the sensor attacks should be
-sparse [
3,
4,
5,
6,
7,
8,
21,
22], this paper utilizes the attack reconstruction algorithm to solve approximate learning for unknown adversarial attacks, no longer requiring these restrictive assumptions.
To ensure clarity and consistency in the text, the abbreviations employed throughout the paper are summarized in Abbreviations prat.
The remainder of this paper is organized as follows.
Section 2 introduces the system architecture.
Section 3 introduces the NN attack reconstruction algorithm to estimate the cyber-attacks.
Section 4 discusses the NN adaptive secure control method. In
Section 5, a stability analysis was conducted on the designed the NN adaptive secure control method.
Section 6 discusses the feasibility of the algorithm through simulation.
2. Preliminaries and Problem Formulation
- A.
System Descriptions
Consider the following nonlinear CPSs as
where
,
is the state vector [
16].
denotes the actuator output.
and
are unknown smooth nonlinear dynamics functions of the CPSs (1).
denotes the output of the controlled system.
- B.
Adversarial Attack Model
In this paper, we consider the adversarial attack as follows [
9,
23]:
where the compromise state
is the available variable after being attacked.
is the unavailable control input caused by the adversarial attack.
is defined as a time-varying, state-dependent malicious adversarial attack function, meaning it changes over time and is influenced by the system’s current state.
Remark 1. In the presence of sensor attacks, this paper proposes an NN estimation algorithm to estimate the unknown sensor attacks, essentially letting the network train itself and determine which attack values are most important. The proposed a NN estimation algorithm method does not require that the sensor attacks must be parameterized. Note that, although [10,11,12,13,14,15] also investigate the problems of injection unknown sensor attacks, the proposed control methods in [10,11,12,13,14,15] require that the sensor attacks must be parameterized, that is, the restrictive assumption is removed in this paper.
In defining the reference signal , the control objectives are given as follows:
Control objective. The control objective of this paper is to design a state-feedback secure control scheme based on the NN attack reconstruction strategies in an insecure network with all the information available for nonlinear CPSs (1) against adversarial attack (2) and (3), such that:
- (1)
The obtained NN attack reconstruction strategies in an insecure network with all information unavailable can effectively estimate unknown attacks.
- (2)
The above secure control process reaches the expected trajectory, and all the closed-loop signals are bounded.
- (3)
The output error of the system converges to a small neighborhood of zero.
3. NN Attack Reconstruction Algorithm
Due to the unknown adversarial attack functions and being unusable, this section needs to construct the attack reconstruction strategies during the control design process. Thus, the NN estimation algorithm is proposed to solve approximate learning for unknown adversarial attacks.
According to the research in reference [
16,
19], the estimation performance of neural networks (NNs) is only guaranteed on the compact set
. However, adversarial attack functions
are time-dependent; therefore, according to reference [
19], we propose a nonlinear mapping that maps
to a compact set:
This paper takes the inverse mapping
of (4), which we can obtain as follows:
According to (4) and (5), the time-related adversarial attack functions
and
have been mapped to the compact set
. NNs are capable of learning complex nonlinear relationships, making them highly adaptable to various types of data and problems. Next, we use NNs to capture adversarial attack patterns and features. We define
Similarly to work [
19], we have
where
and
present an ideal weight vector, which is the optimal weight vector found through training to minimize error, while
is the NN rules number;
and
are activation function vectors, aiding in the Gaussian activation function, with parameters being tuned during training for optimal performance; and
and
are approximation errors of the NN and satisfy that
and
, respectively.
and
are positive constants.
Based on (8) and (9), we can use and to solve approximate learning for unknown adversarial attacks, respectively.
4. Controller Design
Based on the NN attack reconstruction strategies (8) and (9), and under the framework of a backstepping recursive control design algorithm, an NN-adaptive control scheme is developed which can solve approximate learning for unknown adversarial attacks.
To analyze the stability of the designed controller, we define the following error coordinate change as follows:
where
is the virtual controller, which is a virtual controller, which means that the value of
is not a virtual value (it represents the control parameters within the system);
is the error surface;
is a coordinate of the first-order filter output;
is a variable, which is obtained through a first-order filter on intermediate virtual control
with a constant
to obtain
, i.e.,
where
is positive constant.
Step 1: Similarly to Equations (8) and (9), we can express the approximation of the NN as follows:
where
is an ideal weight vector, which is the optimal weight vector found through training to minimize error;
is an activation function vector, which is the Gaussian activation function, with parameters tuned during training for optimal performance; given the positive constant
, the error
of NN and satisfies
. For unknown nonlinear dynamics, a NN learns it by adjusting weights during training based on input-output data
.
By taking the derivative of both sides of (10) simultaneously, and using (1), it can be derived that
To analyze the system stability, we have chosen the following quadratic Lyapunov function. This function is selected to effectively capture the system’s dynamics and provide a suitable framework for evaluating its stability; we have
where
.
To ensure stability, we design the virtual controller and the parameter adaptation laws as follows, and the stability will be proven in the next section.
where
,
,
,
and
are design parameters.
and the vector
provide estimates for the ideal learning weight of
.
Remark 2. Therefore, the main challenge of adversarial attack recovery algorithms lies in how to achieve complete compensation or suppression of attack signals. Due to the unavailability of the tracking error in (10), we introduce a novel error transformation to construct an equivalent instead of . The equivalence between and depends on the performance of (8) and (9), respectively. The estimation performance of NNs can be supported within tight sets.
Step
: By taking the derivative of both sides of (10) simultaneously, it can be obtained that
From (11), one has
where
is a known continuous function.
Similarly to the Lyapunov function in (14), we choose the following Lyapunov function in step
; we obtain
To ensure stability, we design the virtual controller and the parameter adaptation laws as follows, and the stability will be proven in the next section.
where
,
,
,
and
are design parameters. And
.
Remark 3. The author has designed an anti-attack observer that can compensate for attacks to restore the system state in [19], which only requires the sensor output to be attacked. Unlike reference [19], this article studies all states being attacked, and all variables are unavailable. The estimation performance of NNs can be supported within tight sets.
Step
: Similarly to Step
, from (1), (3) and (12) and (10), by taking the derivative of both sides of (10) simultaneously, it can be gathered that
Similarly to the Lyapunov functions (14) and (20), we choose the following Lyapunov function in step
, finding the following:
To ensure stability, we design the virtual controller and the parameter adaptation laws as follows, and the stability will be proven in the next section.
where
,
,
,
,
and
are design parameters. Additionally,
.
The configuration of the NN-adaptive state-feedback control scheme is shown in
Figure 1.
5. Stability Analysis
Theorem 1. For nonlinear CPSs (1), if we adopt the NN attack reconstruction strategies (8) and (9), the secure control scheme (26), adaptive laws (16) and (17), (22) and (23), (28) and (29) and the virtual control laws (15), (21), (27), and if, for all the initial conditions, is bounded, then a whole-state-feedback secure control scheme makes the following properties valid:
- (1)
The obtained NN attack reconstruction strategies in an insecure network with all information unavailable can effectively estimate unknown attacks.
- (2)
The above secure control process reaches the expected trajectory and all the closed-loop signals are bounded.
- (3)
The output error of the system converges to a small neighborhood of zero.
Proof. According to (19) and (20), we can obtain
According to (1) and (10), we have
From (14), by applying the inequality
, we can derive the following expression/relationship, which will help in simplifying the analysis and in providing a clearer understanding of the system’s behavior.
Substituting (30)–(32) into (12) yields
Substituting (14)–(16) into (35) yields
By employing the inequality
, we can obtain
According to (36) and (37), we have
where
.
According to (12) and (18), we can obtain
Let
. Since
is a compact set and
satisfies
on
with a constant
, from (11) and (20), and by using the inequality
, we can obtain
Substituting (40)–(43) into (20) yields
Substituting (21)–(23) into (44) yields
By employing the inequality
, we can obtain
According to (45) and (46), we have
where
.
According to (1), (3), (9) and (10), we can obtain
Similary to (40)–(46), we can gather that
where
.
By employing the inequality
, we can obtain
According to (49) and (50), we have
where
.
Then, (51) can be written as
where
. This means that the system exhibits stability and boundedness over time, ensuring that the state variables remain within a certain range, even in the presence of potential disturbances or uncertainties.
By integrating both sides of (52) simultaneously with respect to
, it can be found that
Denote the initial conditions are , , , and . According to (53), it can be seen that, with the definition of , we can obtain , and the control tracking error satisfy .
The proof of Theorem 1 is complete. □
Remark 4. The time complexity is generally , where is the number of iterations, is the number of layers in the network, and is the number of neurons per layer; the time complexity for such methods is generally , where is the number of parameters to be optimized and is the number of optimization iterations. Although parameterized attack algorithms typically have lower computational complexity, the parameterization assumption they rely on is quite restrictive. This assumption limits their applicability to parameterized attack models exclusively, making them less flexible for general adversarial attack scenarios. In contrast, our approach removes this restrictive assumption, offering a more general and adaptable framework for evaluating a broader variety of adversarial attack methods.
In this paper, parameter selection guidelines are as follows:
- (1)
Select design parameters and to determine the adversarial attack reconstruction adaptive laws and .
- (2)
Select design parameters such that , , , and to determine the virtual controller and the adaptive law .
6. Simulation Results
In this section, we provide two-joint robots to verify the effectiveness of the proposed NN attack reconstruction strategies and state-feedback secure control scheme against adversarial attacks. We expect to achieve the following control objective:
- (1)
The obtained NN attack reconstruction strategies in an insecure network with all information unavailable can effectively estimate unknown attacks.
- (2)
The above secure control process reaches the expected trajectory, and all the closed-loop signals are bounded, i.e., .
- (3)
The output error of the system converges to a small neighborhood of zero, i.e., . is an infinitesimal constant.
Example 1. The dynamic equation of a two-joint manipulator is defined as follows:where ,
and
are the angle, angular velocity and angular acceleration of the manipulator, respectively.
is an inertia;
is a centripetal force and Coriolis moment;
refers to an external disturbance, which does not affect the system’s convergence but allows for a more realistic representation of the system’s behavior under practical conditions;
is a control input torque vector;
is an output vector. In order to apply the backstepping method, we define and
,
which is written (54) aswith where is the mass of the first connecting rod and is the mass of the second connecting rod ;
is the length of the first connecting rod and is the length of the second connecting rod;
is the distance from the first joint to the center of gravity of the first connecting rod and is the distance from the second joint to the center of gravity of the first connecting rod;
is the angle of the first connecting rod and is the angle of the second connecting rod;
is the inertia matrix of the first connecting rod and is the inertia matrix of the second connecting rod. We describe the two-joint robotic in simulation by means of a 3-dimensional model, which is shown in Figure 2.
The FDI attacks model in (2) is defined as follows:
with
The actuator attacks model in (3) is defined as follows:
with
The parameter (the center and variance of radial basis function and the weight value from hidden layer to output layer) learning process follows the principle of automatically increasing the number of hidden neurons of the network until the number of neurons reaches the maximum. The spread of radial basis function is set to 1. In the simulation, the number of neurons for unknown function
and
are chosen as 11. The number of neurons for sensor attack
and
are chosen as 101. The activation functions are chosen by
. The layers of the NN can be presented as follows
Figure 3:
We design the secure controller, virtual controller, and the parameter adaptive laws as
The system initial conditions are chosen as
and
. The other initial conditions of the systems are chosen as zero. Additionally, the reference signal is
and the external disturbance is
. The design parameters in (56)–(58) and (64)–(71) are given in
Table 1.
In this simulation, we simulated a sustained confrontation between attackers and defenders. The NN attack reconstruction strategies have been plotted in
Figure 4,
Figure 5 and
Figure 6.
Figure 4,
Figure 5 and
Figure 6 show that, in regard to the NN estimation, Algorithm (8) and (9) solve approximate learning for unknown adversarial attacks and improve the ability to cope with the adversarial attacks at
and
. We use the developed NN attack reconstruction strategies (8) and (9) to solve the state-compromised problem caused by adversarial attacks, which further confirms that, in terms of NN estimation, Algorithm (8) and (9) can solve approximate learning for unknown adversarial attacks. Thus, the obtained NN attack reconstruction strategies in an insecure network with all information unavailable can effectively estimate unknown attacks.
Figure 7,
Figure 8 and
Figure 9 show that the performance trajectory diagram of the designed state-feedback secure control scheme in a closed-loop control system.
Figure 7 shows the attitude angles
and their observations of two-joint robots under the proposed NN attack reconstruction strategies, and the curves of the corresponding angular velocities
and their observations are plotted in
Figure 8. Thus, it can be seen that the output error of the system converges to a small neighborhood of zero, i.e.,
.
Figure 9 shows that the two-joint robots control the input signal. It indicates that the followers’ attitude can still be consistent with the reference signal trajectory attitude despite adversarial attacks. Thus, it can be seen that the above secure control process reaches the expected trajectory and that all the closed-loop signals are bounded, i.e.,
.
To emphasize the superiority of the designed scheme, both the algorithms in [
21] and our proposed algorithm are implemented under identical conditions.