Next Article in Journal
Evaluation of a Voltametric E-Tongue Combined with Data Preprocessing for Fast and Effective Machine Learning-Based Classification of Tomato Purées by Cultivar
Previous Article in Journal
ADM-SLAM: Accurate and Fast Dynamic Visual SLAM with Adaptive Feature Point Extraction, Deeplabv3pro, and Multi-View Geometry
Previous Article in Special Issue
A Survey on Satellite Communication System Security
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adversarial Attacks against Deep-Learning-Based Automatic Dependent Surveillance-Broadcast Unsupervised Anomaly Detection Models in the Context of Air Traffic Management

School of Information and Navigation, Air Force Engineering University, Xi’an 710051, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(11), 3584; https://doi.org/10.3390/s24113584
Submission received: 19 March 2024 / Revised: 15 May 2024 / Accepted: 28 May 2024 / Published: 2 June 2024
(This article belongs to the Special Issue Cybersecurity Attack and Defense in Wireless Sensors Networks)

Abstract

:
Deep learning has shown significant advantages in Automatic Dependent Surveillance-Broadcast (ADS-B) anomaly detection, but it is known for its susceptibility to adversarial examples which make anomaly detection models non-robust. In this study, we propose Time Neighborhood Accumulation Iteration Fast Gradient Sign Method (TNAI-FGSM) adversarial attacks which fully take into account the temporal correlation of an ADS-B time series, stabilize the update directions of adversarial samples, and escape from poor local optimum during the process of iterating. The experimental results show that TNAI-FGSM adversarial attacks can successfully attack ADS-B anomaly detection models and improve the transferability of ADS-B adversarial examples. Moreover, the TNAI-FGSM is superior to two well-known adversarial attacks called the Fast Gradient Sign Method (FGSM) and Basic Iterative Method (BIM). To the best of our understanding, we demonstrate, for the first time, the vulnerability of deep-learning-based ADS-B time series unsupervised anomaly detection models to adversarial examples, which is a crucial step in safety-critical and cost-critical Air Traffic Management (ATM).

1. Introduction

As a new generation of air traffic management surveillance technology, ADS-B can broadcast aircraft position, velocity, heading and other flight information in real-time which improves the visibility of air traffic, reduces flight conflicts and enhances the efficiency and safety of ATM [1]. However, since ADS-B data are broadcast in plaintext format and lack an authentication mechanism, the attackers can easily jam, eavesdrop, modify, inject and delete ADS-B messages [2,3,4]. Therefore, it is crucial to detect anomaly data caused by ADS-B attacks [5,6].
Compared to other methods such as encryption, physical layer information and multilateration, deep learning has demonstrated tremendous success in the field of detecting ADS-B anomaly data [7,8,9,10]. The deep learning method usually detects ADS-B anomaly data based on prediction errors and reconstruction errors. Taking into account the temporal correlation of ADS-B data, Long Short-Term Memory (LSTM) has been utilized for detecting ADS-B anomaly data by analyzing prediction errors [11]. In order to improve the accuracy of anomaly detection, an LSTM-Encoder-Decoder is used to reconstruct ADS-B data and determine whether ADS-B data are anomalous [12]. Considering the maneuvering characteristic of ADS-B flight data, a Variational Autoencoder (VAE) and Gate Recurrent Unit (GRU) model are used to detect ADS-B anomaly data on the basis of reconstruction errors [13]. The deep learning method makes full use of the temporal correlation of ADS-B data to establish anomaly detection models, which have the advantage of detecting ADS-B anomaly data accurately and quickly.
However, deep learning itself can be easily fooled by adversarial examples [14,15]. Therefore, deep-learning-based ADS-B anomaly detection models could also be attacked by adversarial examples. The applications of ADS-B anomaly detection models end with decision making by pilots or onboard automation systems [16]. Therefore, it may result in serious consequences such as flight deviation, flight delays and aircraft collisions once deep-learning-based ADS-B anomaly detection models are attacked by adversarial examples [17]. Not only the accuracy, but also the robustness of ADS-B anomaly detection models need to be investigated. At present, adversarial examples have been widely studied in computer vision fields such as image classification, object detection and face recognition [18,19,20]. From the perspective of image classification, adversarial examples refer to modifying the original image with small, imperceptible adversarial perturbations, causing the modified image to be classified incorrectly [21,22]. However, the application of adversarial samples in non-image fields, especially in time series prediction and regression problems, is vastly limited [23]. This also includes the lack of research on adversarial samples for ADS-B time series anomaly detection despite the increasing success of deep learning in the field of ADS-B time series anomaly detection. In our previous work, we managed to craft adversarial examples to fool an ADS-B spoofing detection classifier based on a Manchester Encoding Attack (MEA) [24]. The biggest flaw of this work is that it assumes that the victim model is supervised and considers ADS-B spoofing detection as a classification problem. However, there is scarce labeled ADS-B anomaly data in the real-world environment [25]. Therefore, the victim ADS-B anomaly detection model should be considered as an unsupervised regression problem that makes full use of the temporal correlation in an ADS-B time series. There are two reasons why previous methods, including FGSM and BIM, are not suitable for adversarial attacks against ADS-B anomaly detection models [14,15]. First, previous methods failed to take into account the temporal correlation of an ADS-B time series, thus resulting in a poor success rate of adversarial attacks. Second, previous adversarial attack methods failed to stabilize the update directions of adversarial samples and could not escape from poor local optimum during the process of iterating, thus leading to poor transferability of adversarial attacks.
In order to solve the problems mentioned above, this paper proposes a TNAI-FGSM to craft imperceptible adversarial samples to fool the victim deep-learning-based ADS-B time series unsupervised anomaly detection models. The main contributions of this paper are summarized as follows:
(I) The TNAI-FGSM is proposed to craft ADS-B adversarial samples. To the best of our understanding, we demonstrate, for the first time, the vulnerability of deep-learning-based ADS-B unsupervised anomaly detection models to adversarial attacks.
(II) Based on a time neighborhood mechanism, the TNAI-FGSM fully takes into account the temporal correlation of an ADS-B time series which can improve the success rate of adversarial attacks.
(III) Based on an accumulation iteration mechanism, the TNAI-FGSM can stabilize the update directions of adversarial samples and escape from poor local optimum during the process of iterating, thus improving the transferability of adversarial attacks.
(IV) The experimental results show that the TNAI-FGSM is superior to other adversarial attack methods in terms of success rate, stealthiness and transferability.
The rest of this paper is organized as follows. Section 2 introduces the basic knowledge about security solutions to ADS-B data, including deep learning for ADS-B anomaly detection, and briefly discusses adversarial attacks. Section 3 formalizes the adversarial attacks against deep-learning-based ADS-B unsupervised anomaly detection models and explains the FGSM, BIM and our proposed TNAI-FGSM for crafting adversarial samples. Section 4 shows the experimental results of adversarial attacks against four types of ADS-B anomaly detection models. The transferability property of adversarial attacks is evaluated. We also use adversarial training to defend against ADS-B adversarial attacks in this section. Section 5 provides a discussion about the limitations of this work and future work. Section 6 makes conclusions.

2. Related Works

2.1. ADS-B

ADS-B has become a core technology for the new generation ATM system. ADS-B has the advantages of wide coverage, high accuracy, low cost and support for information sharing. Figure 1 gives a general overview of an ADS-B system. Aircraft obtain precise position, heading and velocity information from the Global Navigation Satellite System (GNSS) and on-board devices. Aircraft equipped with ADS-B Out broadcast ADS-B data over the 1090 MHz Extended Squitter (1090ES) or Universal Access Transceiver (UAT) communication channel. Nearby aircraft equipped with ADS-B In can receive ADS-B data. The ground stations receive and process ADS-B data which are then sent to the Air Traffic Control (ATC) system. However, since ADS-B data are broadcast in plaintext format and lack an authentication mechanism, attackers can easily jam, eavesdrop, modify, inject and delete ADS-B messages [26].

2.2. Secure Solutions to ADS-B Data

At present, security solutions for ADS-B risks mainly include four categories. The four categories are encryption, physical layer information, multilateration and deep-learning-based ADS-B anomaly detection. The first three categories can be called traditional methods. Table 1 provides a comparison of the existing security solutions. In this paper, we focus on deep-learning-based ADS-B unsupervised anomaly detection models as the victim models. Therefore, in the following, we present less traditional methods and analyze more deep-learning-based ADS-B anomaly detection models.

2.2.1. Traditional Methods

(1) Encryption: An ADS-B encryption method uses keys to directly encrypt ADS-B plaintext or generate corresponding integrity information to provide an information basis for ADS-B security reinforcement [7,27,28]. The encryption method needs to modify the original ADS-B protocol, which hinders system compatibility and international interoperability.
(2) Physical layer information: When ADS-B data are transmitted by wireless communication, the data are equipped with physical fingerprint information, which is the basis for analyzing the probability of ADS-B anomaly data [8,29]. However, when attackers obtain prior knowledge of physical layer information through statistical analysis, the applicability of the method will need further verification for sophisticated ADS-B data attacks.
(3) Multilateration: The multilateration method usually uses Time Difference of Arrival (TDOA) to measure the position of the aircraft and compares it with the parsed ADS-B position [9,30]. If the difference between the two positions is too large, the ADS-B message is illegal. However, the multilateration method requires deploying multiple ground stations, which is not economical.

2.2.2. Deep Learning for ADS-B Anomaly Detection

Recently, deep learning has demonstrated tremendous success in the field of detecting ADS-B anomaly data [2,3,10]. ADS-B data have the characteristics of temporal correlation, maneuvering characteristics of ADS-B flight data, the lack of labeled ADS-B anomaly data in real environments and the characteristic of multidimensional attributes such as longitude, latitude, altitude, velocity and heading at any moment [17]. To achieve excellent performance, the deep learning method needs to take into account the above characteristics of ADS-B data. The deep learning method usually detects ADS-B anomaly data based on prediction errors and reconstruction errors [31]. Based on fully considering the characteristic of temporal correlation and the lack of labeled ADS-B anomaly data in real environments, LSTM and GRU are usually used to establish unsupervised ADS-B anomaly detection models and determine whether ADS-B data are anomalous based on prediction errors [11]. In order to improve the accuracy of anomaly detection, an LSTM-Encoder-Decoder is used to reconstruct ADS-B data and determine whether ADS-B data are anomalous based on reconstruction errors [12]. In order to fully consider the maneuvering characteristics of ADS-B flight data, a VAE is used to detect ADS-B anomaly data by analyzing reconstruction errors [13]. The LSTM-Encoder-Decoder and VAE also use LSTM and Gated Recurrent Unit (GRU) as the hidden layer to preserve the temporal correlation of ADS-B data [32]. In this paper, we focus on four representative deep-learning-based ADS-B unsupervised anomaly detection models as the victim models which are based on the VAE, LSTM and LSTM-Encoder-Decoder. The specific descriptions of the four victim models are shown in Section 3.2: Typical ADS-B training module.

2.3. Adversarial Attacks

Adversarial attacks are proposed for image recognition at first [14,33]. From the perspective of image recognition, adversarial examples refer to modifying the original image with small, imperceptible adversarial perturbations, causing the modified image to be classified incorrectly with high confidence [21,34]. Based on the idea above, some adversarial attack algorithms have been proposed, among which the most classic algorithms are the FGSM and BIM [14,15,35]. By utilizing the gradient information of the loss function, the FGSM generates adversarial examples that are designed to mislead the model’s prediction or classification [17,35]. The BIM is an extension of the FGSM. The BIM performs multiple small perturbations along the direction of an increasing gradient through iterative methods, while the FGSM only performs one step of perturbations [15,36]. Although the application of adversarial samples in image fields is becoming more and more popular, studies on adversarial samples in image fields, especially in time series prediction and regression problems, are vastly limited [23]. This also includes the lack of research on adversarial samples for ADS-B time series anomaly detection despite the increasing success of deep-learning-based ADS-B time series anomaly detection. In our previous work, we managed to craft adversarial examples to fool an ADS-B spoofing detection classifier based on a Manchester Encoding Attack (MEA) [24]. The biggest flaw of this work is that it assumes that the victim model is supervised and assumes ADS-B spoofing detection is a classification problem. However, there is scarce labeled ADS-B anomaly data in a real-world environment [25]. Therefore, the victim ADS-B anomaly detection model should be considered as an unsupervised regression problem that makes full use of the temporal correlation in ADS-B time series. There are two reasons why previous methods, including the FGSM and BIM, are not suitable for adversarial attacks against ADS-B anomaly detection models [14,15,35,36]. First, previous methods failed to take into account the temporal correlation of an ADS-B time series, thus resulting in a poor success rate of adversarial attacks. Second, previous adversarial attack methods failed to stabilize the update directions of adversarial samples and could not escape from poor local optimum during the process of iterating, thus leading to poor transferability of adversarial attacks. Table 2 summarizes the gaps of previous adversarial attack methods when they are considered to be applied to deep-learning-based ADS-B time series unsupervised anomaly detection models.

3. Adversarial Attacks against ADS-B Anomaly Detection Models

In this section, we formalize adversarial attacks against deep-learning-based ADS-B unsupervised anomaly detection models and explain our proposed TNAI-FGSM for crafting adversarial samples.

3.1. Formalization of Adversarial Attacks

Definition 1.
Let  X = [ X 1 , X 2 , , X t , , X M ]   be the original ADS-B multivariate time series received by ADS-B anomaly detection models.  M = | X |   is the length of  X . X t  is an N-dimensional vector at time  t  where  t [ 1 , M ] . Each dimension of   X t  represents the ADS-B feature, which includes latitude, longitude, altitude, velocity and heading.
Definition 2.
Let  f ( ) : R N × M F  represent a deep-learning-based ADS-B unsupervised anomaly detection model. This task can be represented as training a deep learning model in order to predict F from the input X = [ X 1 , X 2 , , X t , , X M ] . F can be normal or an anomaly.
Definition 3.
Let  X a d v = X + η  denote the attacked ADS-B time series.  η  is a small, imperceptible noise.  X a d v  is a perturbed version of  X  with the aim of  F F  and  X a d v X ε , where  f ( X ) = F ,  f ( X a d v ) = F  and  ε 0  is a very small magnitude. Therefore, we give the formalization of adversarial attacks against deep-learning-based ADS-B unsupervised anomaly detection models. Given a trained anomaly detection model f and original input ADS-B time series  X , crafting adversarial examples  X a d v  can be represented as an optimization problem with constraints as follows:
min X a d v X a d v X s . t . f ( X ) = F , f ( X a d v ) = F , F F
η = X a d v X is a small, imperceptible noise added to X . Figure 2 also shows the general architecture of adversarial attacks against ADS-B anomaly detection models. Below, we separately demonstrate a typical ADS-B training module, threshold determination module and adversarial attacks module as shown in Figure 2.

3.2. Typical ADS-B Training Module

3.2.1. Selection Criteria

In this work, we focus on four representative deep-learning-based ADS-B unsupervised anomaly detection models as the victim models which are based on a VAE, LSTM, GRU and LSTM-Encoder-Decoder [11,12,13]. The reasons for selecting these four victim models as targets include the following.
(1) In terms of scope, the four selected models include both types of unsupervised ADS-B anomaly detection methods, which are based on prediction errors and reconstruction errors.
(2) In terms of classic and state-of-the-art methods, models based on LSTM, GRU and LSTM-Encoder-Decoder are representative because later ADS-B anomaly detection methods are generally improved based on the four models to improve the accuracy of anomaly detection. Furthermore, an ADS-B anomaly detection model based on a VAE is relatively state-of-the-art and fully considers the maneuvering characteristic of ADS-B flight data.
(3) In fact, not only the four ADS-B anomaly detection models mentioned above but also all the other existing models have not considered adversarial attacks. We can reasonably infer that all other existing ADS-B anomaly detection models are vulnerable to adversarial attacks based on the transferability property of adversarial samples.

3.2.2. Selected ADS-B Anomaly Detection Baselines

We selected the following victim ADS-B anomaly detection models based on the above criteria. (1) ADS-B anomaly detection based on LSTM and GRU [11]: The core idea of an LSTM-based and GRU-based method is prediction errors. Specifically, the neural network composed of LSTM and GRUs is used for ADS-B predictive training. The threshold is set by calculating the difference value between the ADS-B predicted value and ADS-B actual value. If the difference value is greater than the threshold, the ADS-B data are an anomaly, and vice versa. (2) ADS-B anomaly detection based on an LSTM-Encoder-Decoder and VAE [12,13]: An LSTM-Encoder-Decoder and VAE detect ADS-B anomaly data based on reconstruction errors. In order to improve the accuracy of anomaly detection, an LSTM-Encoder-Decoder is used to reconstruct ADS-B data and determine whether ADS-B data are anomalous. In order to fully consider maneuvering characteristics of ADS-B flight data, a VAE is used to detect ADS-B anomaly data on the basis of reconstruction errors. An LSTM-Encoder-Decoder and VAE also use LSTM and GRUs as the hidden layer to preserve the temporal correlation of ADS-B data. The parameter settings of the four selected models are basically consistent with those in reference [11,12,13]. In order to solve the problem of poor adaptability of the threshold determined by manually analyzing the prediction/reconstruction errors, Support Vector Data Description (SVDD) is used to train the difference value [37].

3.3. Threshold Determination Module

This module compares prediction/reconstruction errors with the threshold of anomaly detection. If prediction/reconstruction errors are greater than the threshold, the ADS-B data are an anomaly, and vice versa. SVDD can solve the threshold adaptive problem and achieve the relatively optimal accuracy of ADS-B anomaly detection, so it is selected for determining the threshold [13,37]. The difference value between the predicted/reconstructed value and the actual value is put into SVDD for training, and then the threshold of anomaly detection can be obtained. SVDD can be expressed as the following optimization problem:
min H ( R , a ) = R 2 + C j n ξ j s . t . { D j a 2 R 2 + ξ j ,   ( j = 1 , 2 , , L ) ξ j 0
R is the radius of SVDD hypersphere, which is the threshold of anomaly detection. a is the center of the hypersphere. ξ j is the slack variable. C is the penalty coefficient which is used to control the volume of the hypersphere, and the value of C is usually 1. If the distance from D j to the center a is denoted as S ( D j ) , then S ( D j ) can be expressed as follows:
S ( D j ) = D j a = ( D j , D j ) 2 i = 1 L λ i ( D i , D j ) + i = 1 L k L λ i λ k ( D i , D k )
It means that the sample D j are anomaly data when S ( D j ) > R . The sample D j are normal data when S ( D j ) R .

3.4. Adversarial Attacks Module

In the adversarial attacks module, we first define and analyze adversarial attackers from three aspects which are capability, knowledge and goals. Then, we explain the FGSM, BIM and our proposed TNAI-FGSM for crafting ADS-B adversarial samples.

3.4.1. Definition of the Adversarial Attackers

Capabilities: We consider adversarial attackers whose objective is to reduce the effectiveness of the victim ADS-B anomaly detection models. Adversarial attackers could appear in the ground station, air traffic control center, airlines and aircraft. Adversarial attackers with Software Defined Radio (SDR) devices can compromise the ADS-B sensors and 1090 MHz Extended Squitter (1090ES) or Universal Access Transceiver (UAT) communication channel [4,5]. Based on the capabilities above, attackers can apply the adversarial samples by modifying ADS-B data.
Knowledge: To verify the vulnerability of ADS-B anomaly detection models, we launch a white-box attack and black-box attack, respectively. (1) ADS-B white-box attackers know all the information and parameters inside the anomaly detection models. Based on the gradient of the given model, adversarial samples are generated to attack the network. (2) ADS-B black-box attackers cannot obtain the internal structure and parameter information of anomaly detection models and can only attempt to disrupt the behavior of the models by observing ADS-B input data and the output results. In this work, we launch the transfer-based black-box attack against ADS-B anomaly detection models.
Goals: Attackers consider two situations: (a) normal to anomaly and (b) anomaly to normal. In (a), adversarial attackers generate the adversarial sample X a d v to make the models determine ADS-B normal data as anomaly data incorrectly, hence generating a false-positive. In (b), adversarial attackers create X a d v to make the models determine anomaly data as normal data incorrectly, hence generating a false-negative. Adversarial attacks can lead pilots to make incorrect decisions, including changing flight routes, adjusting flight altitude and velocity and even executing emergency landing procedures.

3.4.2. Adversarial Samples Generation

Adversarial Samples Generation

The FGSM attack was first proposed in image fields [14]. The FGSM calculates the gradient of the cost function relative to the input of ADS-B anomaly detection models. The FGSM is also known as the one-shot method as adversarial perturbations are generated by a single-step computation. The FGSM generates ADS-B adversarial examples X a d v as follows:
{ η = ε × s i g n ( x J f ( X ,   F ) ) X a d v = X + η
J f is the cost function of ADS-B anomaly detection model f , x refers to the gradient of the model f with respect to the original ADS-B time series X with the correct output F , sign refers to the sign function, ε is a very small magnitude and X a d v is the adversarial sample.
The BIM is a block-based iterative method [15]. The BIM extends the FGSM by applying it multiple times with a small step size and clipping the ADS-B time series after each step. Adversarial samples generated by the BIM are more similar to the original ADS-B time series, which means better stealthiness. This is because the BIM uses a multi-step iterative method with a small step size to generate ADS-B adversarial samples. However, due to the multi-step iterations, the BIM may fall into the local optimum during the process of iterating, resulting in the poor transferability of adversarial attacks. The BIM generates ADS-B adversarial examples X a d v as follows:
X i + 1 a d v = min { X i a d v + ε ,   max { X i a d v ε ,   X i a d v + α × s i g n ( x J f ( X i a d v ,   F ) ) } } s . t . 1 i I
X i a d v represents the adversarial sample of the i-th iteration. ε represents the amount of maximum perturbation. α represents the per step small perturbation. sign refers to the sign function. x refers to the gradient of the model f with respect to the original ADS-B time series X . J f is the cost function of ADS-B anomaly detection model f . I represents the number of iterations. The three parameters in the BIM usually need to satisfy α × I = ε . The BIM utilizes a min and max clipping function to generate adversarial samples.

Our Proposed TNAI-FGSM Generating ADS-B Adversarial Samples

In order to improve the success rate of adversarial attacks, the TNAI-FGSM utilizes a time neighborhood mechanism to craft adversarial ADS-B samples which fully takes into account the temporal correlation of ADS-B data. Additionally, to improve the transferability of adversarial attacks against different ADS-B anomaly detection models, the TNAI-FGSM integrates an accumulation iteration mechanism into the process of adversarial attacks. The accumulation iteration mechanism can stabilize the update directions of adversarial samples and escape from poor local optimum during the process of iterating, resulting in more transferable ADS-B adversarial attacks. By adding the time variable t, the TNAI-FGSM generates ADS-B adversarial examples X a d v as Equation (6) to Equation (8). Algorithm 1 also specifically shows the process of the TNAI-FGSM generating adversarial samples.
E L ( X i , t ) = 1 2 L + 1 k = L L x J f ( X i , t + k a d v ,   F )
G i + 1 , t = G i , t + E L ( X i , t ) E L ( X i , t ) 1
X i + 1 , t a d v = X i , t a d v + α × s i g n ( G i + 1 , t )
In Equations (6)–(8), i represents the i-th iteration. t denotes ADS-B data at time t.
Equation (6) represents the time neighborhood mechanism to craft adversarial ADS-B samples. x J f ( X i , t a d v ,   F ) represents the gradient of the loss function at time t. In order to fully take into account the temporal correlation of ADS-B data, E L ( X i , t ) is used to represent the average gradient from time t L to time t + L . L denotes the range of time neighborhood (Section 4.2.3: Selection of the range of time neighborhood L). The time neighborhood mechanism fully considers the gradient information of the historical and future moments of ADS-B data, and thus generates more stable and reliable adversarial samples with a higher success rate of adversarial attacks.
Equation (7) represents the accumulation iteration mechanism to craft adversarial ADS-B samples. Note that Equation (7) utilizes the accumulation iteration mechanism for crafting adversarial ADS-B samples while considering the time neighborhood mechanism. G i , t represents the cumulative gradient which gathers the gradients of the first i iterations at time t. In each iteration, the average gradient E L ( X i , t ) is normalized by the L1 distance (any distance measure is feasible). The accumulation iteration mechanism can memorize the gradients of previous iterations which helps to barrel through narrow valleys, small humps and poor local minima or maxima [38]. We apply the idea to generate ADS-B adversarial examples which can stabilize the update directions of adversarial samples and escape from poor local optimum during the process of iterating, resulting in more transferable ADS-B adversarial attacks. In Equation (8), sign refers to the sign function and α represents the per step small perturbation.
Algorithm 1. TNAI-FGSM generating ADS-B adversarial samples
Input:Original ADS-B time series X = [ X 1 , X 2 , , X t , , X M ] and its anomaly detection result F
Parameters:The number of iterations I, maximum perturbation ε , per step small perturbation α , the range of time neighborhood L
Output:ADS-B adversarial sample X a d v
//Initialize adversarial example X 0 , t a d v and cumulative gradient G 0 , t
1. X 0 , t a d v = X t , G 0 , t = 0
2. for i = 1   t o   I   do
//Time neighborhood mechanism calculates the average gradient
3.    E L ( X i , t ) = 1 2 L + 1 k = L L x J f ( X i , t + k a d v ,   F )
//Accumulation iteration mechanism gathers the gradients of the first i iterations
4.    G i + 1 , t = G i , t + E L ( X i , t ) E L ( X i , t ) 1
//Calculate the final adversarial example X i + 1 , t a d v
5.    X i + 1 , t a d v = X i , t a d v + α × s i g n ( G i + 1 , t )
6. end for

3.5. Evaluation Index

The confusion matrix for sample classification is given in Table 3. True Positive (TP) indicates that real normal data are detected as normal data. False Negative (FN) indicates that real normal data are detected as anomaly data. False Positive (FP) indicates that real anomaly data are detected as normal data. True Negative (TN) indicates that real anomaly data are detected as anomaly data.
False Positive Rate (FPR), False Negative Rate (FNR), Accuracy and F1_score are used as the evaluation index. FPR, FNR, Accuracy and F1_score are defined as follows:
{ FPR = FN TP + FN FNR = FP FP + TN Accuracy = TN + TP TP + FN + FP + TN F 1 _ score = 2 TP 2 TP + FN + FP

4. Experiments

4.1. Data Collection

For the experiment, ADS-B data from 50 flights were collected as training samples from OPENSKY which records ADS-B data from the real world [39]. In addition, ADS-B data from 40 flights were collected as test samples. In OPENSKY, ADS-B data from real world were recorded every 10 s. Each flight includes aircraft’s take-off, climb, cruise, turning and descent phase. Each flight contains ADS-B data from 200 to 1000. Considering the limitations of actual environments, ADS-B anomaly data were difficult to obtain. ADS-B anomaly data in the test dataset were generated by simulation. The simulated anomaly styles include random position deviation, velocity slow offset, Denial of Service (DOS) and altitude slow offset. In order to demonstrate the effect of adversarial attacks, a sample flight containing 242 ADS-B data was selected randomly for testing. Four types of ADS-B anomaly data were simulated based on original ADS-B data which is shown in Table 4 and Figure 3.

4.2. Experimental Analysis

4.2.1. Adversarial Attack against VAE

We first attacked VAE by using non-targeted FGSM, BIM and TNAI-FGSM with the size of adversarial perturbation ε = 0.05 , the per step small perturbation α = 0.001 , the number of iterations I = 50 , the range of time neighborhood L = 6 (Section 4.2.3: Selection of the range of time neighborhood L). In order to demonstrate the effect of adversarial attacks, a sample flight containing 242 ADS-B data was selected randomly for testing. The first half of ADS-B data contained the aircraft’s take-off, climb, turning and cruise phase. The last half of ADS-B data contained the aircraft’s cruise, turning and descent phase. The simulated ADS-B anomaly data began in the last half of the ADS-B data. When the anomaly type of ADS-B data was random position deviation, the result of adversarial attacks against VAE is depicted in Figure 4. We can analyze Figure 4 to obtain the following results. For ground truth (no adversarial attacks), in the first half of normal test samples, there were 5 samples whose anomaly values were greater than the threshold, thus FPR was 4.13%. In the last half of anomaly test samples, there were 9 samples whose anomaly values were less than the threshold, thus FNR was 7.44%. Accuracy was 94.22%. F1_score was 94.31%. For FGSM adversarial attack, in the first half of normal test samples, there were 29 samples whose anomaly values were greater than the threshold, thus FPR was 23.97%. In the last half of anomaly test samples, there were 8 samples whose anomaly values were less than the threshold, thus FNR was 6.61%. Accuracy was 84.71%. F1_score was 83.26%. For BIM adversarial attack, in the first half of normal test samples, there were 74 samples whose anomaly values were greater than the threshold, thus FPR was 61.16%. In the last half of anomaly test samples, there were 7 samples whose anomaly values were less than the threshold, thus FNR was 5.79%. Accuracy was 66.53%. F1_score was 53.71%. For TNAI-FGSM adversarial attack, in the first half of normal test samples, there were 121 samples whose anomaly values were greater than the threshold, thus FPR was 100%. In the last half of anomaly test samples, there are 2 samples whose anomaly values were less than the threshold, thus FNR was 1.65%. Accuracy was 49.17%. F1_score was 0%.
When the anomaly type of ADS-B data was velocity slow offset, DOS and altitude slow offset, the results of adversarial attacks against VAE are depicted from Figure 5, Figure 6 and Figure 7. The experiments tested 40 flights for adversarial attacks against VAE model, and the average values of FPR, FNR, Accuracy and F1_score of the 40 flights with four types of anomaly ADS-B data were taken as the experimental results. Table 5 shows the experimental results of adversarial attacks against VAE model. By analyzing Table 5, we can draw the following conclusions. (1) For FPR, a small, imperceptible adversarial perturbation can result in VAE model producing a false alarm, with TNAI-FGSM performing the best. In other words, TNAI-FGSM adversarial attack makes VAE model incorrectly interpret over 90% of ADS-B normal data as anomaly data. (2) For FNR, adversarial attacks helped VAE models detect ADS-B anomaly data. It is an interesting phenomenon that adversarial attacks against unsupervised anomaly detection models can lead to a decrease in FNR. This may be because adversarial attacks can increase the degree of anomaly in ADS-B data, making ADS-B anomaly data easier to detect. (3) For Accuracy and F1_score, TNAI-FGSM adversarial attack performed best. In other words, TNAI-FGSM is better than FGSM and BIM in fooling ADS-B anomaly detection models. The reason is that TNAI-FGSM improved FGSM and BIM by utilizing time neighborhood mechanism and accumulation iteration mechanism, which resulted in a better attack impact.
In Figure 8, we evaluated VAE model’s Accuracy with respect to the different amounts of perturbations. By analyzing Figure 8, we can draw the following conclusions. (1) For FGSM adversarial attack, larger values of ε made Accuracy gradually decrease. When ε was greater than 0.08, Accuracy was not significantly decreased. For BIM adversarial attack, when ε was greater than 0.07, Accuracy was not significantly decreased. For TNAI-FGSM adversarial attack, when ε was greater than 0.05, Accuracy was not significantly decreased. The results can provide guidance for attackers when selecting attack parameters. (2) We observed that for the larger value of ε , Accuracy gradually decreased and converged to around 50%. This was due to normal and anomaly data; each accounted for half of the ADS-B test data in this experiment. Adversarial attack makes VAE model incorrectly interpret almost all of ADS-B normal data as anomaly data with the larger value of ε . However, adversarial attacks against VAE model can gradually reduce FNR almost to 0 with the larger value of ε .

4.2.2. Adversarial Attack against LSTM, GRU and LSTM-Encoder-Decoder

As with LSTM, GRU and LSTM-Encoder-Decoder, we first used non-targeted FGSM, BIM and TNAI-FGSM with the size of adversarial perturbation ε = 0.05 , the per step small perturbation α = 0.001 , the number of iterations I = 50 and the range of time neighborhood L = 6 (Section 4.2.3: Selection of the range of time neighborhood L). A sample flight containing 242 ADS-B data was selected randomly, which was the same as VAE. When the anomaly type of ADS-B data was random position deviation, velocity slow offset, DOS and altitude slow offset, the results of adversarial attacks against LSTM, GRU and LSTM-Encoder-Decoder are depicted from Figure 9, Figure 10 and Figure 11. Table 6 shows the average FPR, FNR, Accuracy and F1_score results of adversarial attacks against LSTM, GRU and LSTM-Encoder-Decoder model. By analyzing Table 6, we can draw the following conclusions. For FPR, a small imperceptible adversarial perturbation can result in LSTM, GRU and LSTM-Encoder-Decoder model to produce false alarm, with TNAI-FGSM performing best. For FNR, adversarial attacks helped LSTM, GRU and LSTM-Encoder-Decoder model detect ADS-B anomaly data. For Accuracy and F1_score, TNAI-FGSM adversarial attack performed best with the size of adversarial perturbation ε = 0.05 . In other words, TNAI-FGSM was better than FGSM and BIM in fooling ADS-B anomaly detection models when the attack was stealthy. The reason was that TNAI-FGSM improved FGSM and BIM by utilizing time neighborhood mechanism and accumulation iteration mechanism, which resulted in a better attack impact.
From Figure 12, Figure 13 and Figure 14, we evaluated LSTM, GRU and LSTM-Encoder-Decoder model’s Accuracy with respect to the different amounts of perturbations allowed for crafting the adversarial ADS-B data. We can draw the following conclusions. (1) For FGSM, BIM and TNAI-FGSM adversarial attacks, the larger value of ε made Accuracy gradually decrease and converge to around 50%. (2) For FGSM attack, Accuracy of LSTM and GRU was not significantly decreased when ε was greater than 0.07. However, Accuracy of LSTM-Encoder-Decoder was not significantly decreased when ε was greater than 0.08. (3) For BIM attack, Accuracy of LSTM and GRU were not significantly decreased when ε was greater than 0.06. However, Accuracy of LSTM-Encoder-Decoder was not significantly decreased when ε was greater than 0.07. The results can provide guidance for attackers when selecting attack parameters. (4) The results also showed that TNAI-FGSM was better than FGSM and BIM in fooling ADS-B anomaly detection models when the attack was stealthy. TNAI-FGSM decreased Accuracy more when compared to FGSM and BIM especially when the attack was stealthy with ε not greater than 0.05. The reason was that TNAI-FGSM improved FGSM and BIM by utilizing time neighborhood mechanism and accumulation iteration mechanism, which resulted in a better attack impact.
In the experiments, the average time spent by TNAI-FGSM in generating ADS-B adversarial samples was recorded. TNAI-FGSM required 5.63 s to generate 100 ADS-B adversarial samples. ADS-B data was broadcast once per second on average. Therefore, TNAI-FGSM generated ADS-B adversarial samples in real-time compared to the frequency at which ADS-B data was broadcast. It proved that TNAI-FGSM was efficient in generating ADS-B adversarial samples.

4.2.3. Selection of the Range of Time Neighborhood L

In order to fully utilize temporal correlation of ADS-B data to craft adversarial samples, TNAI-FGSM needed to choose an appropriate range of time neighborhood L. If the range of time neighborhood is too short, TNAI-FGSM will lose a large amount of valid gradient information coming from time neighborhood, resulting in the poor performance of adversarial attacks. If the range of time neighborhood is too long, TNAI-FGSM will memorize too much invalid gradient information, resulting in the poor performance of adversarial attacks as well. In order to obtain an appropriate range of time neighborhood, this paper compares the decrease in Accuracy of anomaly detection under different ranges of time neighborhood. As shown in Figure 15, when the range of time neighborhood gradually increased to 6, the decrease in Accuracy of anomaly detection reached the optimum. When the range of time neighborhood exceeded 6, the decrease in Accuracy deteriorated. Therefore, an appropriate range of time neighborhood was selected as 6.

4.2.4. Transfer-Based Black-Box Attack

To evaluate the transferability of ADS-B adversarial attacks, we applied ADS-B adversarial examples crafted for LSTM model on the GRU, LSTM-Encoder-Decoder and VAE model. The size of adversarial perturbation we set was ε = 0.07 . Under this parameter setting, decrease in Accuracy of LSTM converged. Table 7 summarizes the obtained results on transferability. We observed that for all the GRU, LSTM-Encoder-Decoder and VAE models, ADS-B adversarial examples crafted for LSTM were transferable. For instance, ADS-B adversarial examples crafted using FGSM for LSTM model caused an 18.63, 17.41% and 14.24% decrease when transferred to the GRU, LSTM-Encoder-Decoder and VAE model. A similar trend was also observed, however, with a larger percentage increase when ADS-B adversarial examples crafted using BIM and TNAI-FGSM for LSTM model were transferred to GRU, LSTM-Encoder-Decoder and VAE model.
In addition, the results also showed that our proposed TNAI-FGSM was better than FGSM and BIM in fooling ADS-B anomaly detection models. TNAI-FGSM decreased Accuracy more when compared to the FGSM and BIM. The reason was that TNAI-FGSM integrated accumulation iteration mechanism into the process of adversarial attacks. Accumulation iteration mechanism can stabilize the update directions of adversarial samples and escape from poor local optimum during the process of iterating, resulting in more transferable ADS-B adversarial attacks.
Overall, the results showed that ADS-B adversarial examples can be transferred to different ADS-B anomaly detection models. This type of attack is known as transfer-based black-box attack, where attackers do not have access to the target model’s internal parameters, yet they are able to generate perturbed ADS-B time series that can fool ADS-B anomaly detection models.

4.2.5. Defense against Adversarial Attacks

The existing adversarial attack defense approaches applied in computer vision areas mainly include blocking the transferability [40], gradient hiding [41], defensive distillation [42] and adversarial training [14]. We analyzed the applicability of each defense method for ADS-B adversarial attacks.
(1) Blocking the transferability: To block the transferability, this method was proposed such that, as the input is more perturbed, the classifier smoothly outputs lower confidence on the original label and instead predicts that the input is invalid. In essence, this method augments the output class set with a NULL label and trains the classifier to reject the adversarial examples by classifying them as NULL. However, blocking the transferability required the dataset to be labeled, thus it was not applicable to unsupervised machine learning including deep-learning-based ADS-B unsupervised anomaly detection.
(2) Gradient hiding: This method constructed a model that did not have useful gradients, e.g., by using a nearest neighbor classifier instead of a Deep Neural Network (DNN). This method made it difficult to construct adversarial examples due to the absence of a gradient. However, it was not applicable to deep-learning-based ADS-B unsupervised anomaly detection models which are based on a DNN with the presence of the gradients.
(3) Defensive distillation: This method trained a DNN-based classifier that was more robust to input perturbations. Defensive distillation extracted additional knowledge about training points as class probability vectors produced by a DNN, which was fed back into the training regimen. To defend against such perturbations, defensive distillation attempts to reduce variations around the input, and consequently the amplitude of adversarial gradients. In other words, defensive distillation smooths the model learned by a DNN architecture during training by helping the classifier generalize better to samples outside of its training dataset. However, defensive distillation required the dataset to be labeled, thus it was not applicable to unsupervised machine learning including deep-learning-based ADS-B unsupervised anomaly detection.
(4) Adversarial training: The core idea of adversarial training is to use ADS-B adversarial samples as additional training data so that ADS-B anomaly detection models are more robust to adversarial perturbations. Adversarial training can be performed without the need for the training dataset to be labeled, thus it was suitable for ADS-B anomaly detection. Table 8 shows the accuracy performance of the four deep-learning-based anomaly detection models before adversarial attacks, after adversarial attacks and after adversarial training. For adversarial training, 20 of the 40 test flights processed by adversarial attacks were added to the training dataset, and the remaining 20 test flights were used as the test dataset. The results in Table 8 were obtained with the setting of adversarial perturbation ε = 0.05 , the per step small perturbation α = 0.001 , the number of iterations I = 50 and the range of time neighborhood L = 6 . As can be seen in Table 8, Accuracy of LSTM, GRU, LSTM-Encoder-Decoder and VAE were restored to more than 75% after adversarial training. The results verified that adversarial training is a generalized defense strategy. In other words, adversarial training is effective for the security defense of different deep-learning-based ADS-B unsupervised anomaly detection models suffering from different adversarial attacks.

5. Limitations and Future Work

This work verifies that ADS-B anomaly detection models are vulnerable to adversarial attacks and ADS-B adversarial attacks have the characteristic of transferability. However, there may be some limitations in our work, and future work will try to solve them.
(1) This work assumes that the transfer-based black-box attack can fool other ADS-B anomaly detection models. However, except for the VAE, LSTM, GRU and LSTM-Encoder-Decoder, other anomaly detection models which are based on Transformer [43] and Graph Neural Network (GNN) [44] have not been experimentally verified to determine to what extent they are prone to adversarial attacks. We can address this issue in two ways in the future. First, we could consider more efficient and general black-box adversarial attacks. Second, we can use other anomaly detection models which are based on Transformer and GNN as the victim models for experiments.
(2) Although we launched the transfer-based black-box attacks against ADS-B anomaly detection models, the decrease in accuracy was not the most efficient. To address this issue, we will further explore other advanced black-box attack methods including a query-based attack [45] and data-free black-box adversarial attack [46]. When considering a query-based attack against ADS-B anomaly detection models, the most important thing is how to reduce the number of queries and thus reduce the complexity. When considering data-free black-box adversarial attacks against ADS-B anomaly detection models, the most important thing is how to design an efficient Generative Adversarial Network (GAN) for obtaining the substitute model.
(3) For the security defense against ADS-B adversarial attacks, this work explores and discusses defense mechanisms including blocking the transferability, gradient hiding, defensive distillation and adversarial training. Blocking the transferability requires the dataset to be labeled, thus it is not applicable to unsupervised machine learning including deep-learning-based ADS-B unsupervised anomaly detection. Gradient hiding is applicable to the model that does not have useful gradients. Gradient hiding is applicable to the models without gradients, thus it is not suitable for ADS-B anomaly detection models which are based on a DNN with the presence of the gradients. Defensive distillation requires the dataset to be labeled, thus it is not applicable to unsupervised machine learning including deep-learning-based ADS-B unsupervised anomaly detection. Adversarial training can be performed without the need for the training dataset to be labeled, thus it can be used for ADS-B anomaly detection. However, adversarial samples are able to evolve continuously by adversarial attackers, thus adversarial training cannot solve the evolved adversarial samples. Due to the increasing threat of adversarial machine learning to deep-learning-based ADS-B unsupervised anomaly detection models, we will thus improve the robustness of ADS-B anomaly detection models from their initial designs which consider adversarial attacks in the future.

6. Conclusions

In this work, the concept of adversarial attacks against deep-learning-based ADS-B unsupervised anomaly detection models is considered. We define and formalize adversarial attacks against deep-learning-based ADS-B time series unsupervised anomaly detection models. We propose the TNAI-FGSM for adversarial sample generation which is based on the time neighborhood mechanism and accumulation iteration mechanism. The obtained results show the impact of adversarial samples on four types of ADS-B unsupervised anomaly detection models. The results also show that the TNAI-FGSM is better than the FGSM and BIM in fooling ADS-B anomaly detection models when the attack is stealthy. Additionally, ADS-B adversarial examples can be transferred to different ADS-B anomaly detection models. Furthermore, adversarial training is used to defend against ADS-B adversarial attacks. Our study differs from previous research because we focus on adversarial attacks against ADS-B time series unsupervised anomaly detection rather than the broader classification problem. Through our work, we want to raise awareness of the adversarial vulnerability of ADS-B anomaly detection models which is vital in safety-critical and cost-critical ATM. In the future, we plan to extend our work by utilizing other adversarial attacks against ADS-B anomaly detection models and investigate more defense strategies to mitigate ADS-B adversarial attacks.

Author Contributions

Conceptualization, P.L.; Data curation, P.L.; Formal analysis, J.T.; Investigation, P.L.; Methodology, B.W.; Project administration, B.W.; Resources, B.W.; Software, P.L.; Supervision, B.W.; Validation, J.T., C.L. and Y.Y.; Visualization, P.L.; Writing—original draft, P.L.; Writing—review & editing, B.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, Y.; Xiao, G.; Dai, Z. Integrated Display and Simulation for Automatic Dependent Surveillance–Broadcast and Traffic Collision Avoidance System Data Fusion. Sensors 2017, 17, 2611. [Google Scholar] [CrossRef] [PubMed]
  2. Leonardi, M.; Sirbu, G. ADS-B Crowd-Sensor Network and Two-Step Kalman Filter for GNSS and ADS-B Cyber-Attack Detection. Sensors 2021, 21, 4992. [Google Scholar] [CrossRef] [PubMed]
  3. Antoine, C.; Vernotte, A.; Legeard, B. CAE: Contextual auto-encoder for multivariate time-series anomaly detection in air transportation. Comput. Secur. 2022, 116, 102652. [Google Scholar]
  4. Strohmeier, M.; Schäfer, M.; Pinheiro, R.; Lenders, V.; Martinovic, I. On Perception and Reality in Wireless Air Traffic Communication Security. IEEE Trans. Intell. Transp. Syst. 2017, 18, 1338–1357. [Google Scholar] [CrossRef]
  5. Strohmeier, M.; Lenders, V.; Martinovic, I. On the Security of the Automatic Dependent Surveillance-Broadcast Protocol. IEEE Commun. Surv. Tutor. 2014, 17, 1066–1087. [Google Scholar] [CrossRef]
  6. Li, T.; Wang, B.; Shang, F.; Tian, J.; Cao, K. Online sequential attack detection for ADS-B data based on hierachical temporal memory. Comput. Secur. 2019, 87, 101599. [Google Scholar] [CrossRef]
  7. Wu, Z.; Guo, A.; Yue, M.; Liu, L. An ADS-B Message Authentication Method Based on Certificateless Short Signature. IEEE Aerosp. Electron. 2020, 56, 1742–1753. [Google Scholar] [CrossRef]
  8. Leonardi, M.; Strohmeier, M.; Lenders, V. On jamming attacks in crowdsourced air traffic surveillance. IEEE Aerosp. Electron. 2021, 36, 44–54. [Google Scholar] [CrossRef]
  9. Nijsure, Y.A.; Kaddoum, G.; Gagnon, G.; Gagnon, F.; Yuen, C.; Mahapatra, R. Adaptive Air-to-Ground Secure Communication System Based on ADS-B and Wide-Area Multilateration. IEEE Trans. Veh. Technol. 2016, 65, 3150–3165. [Google Scholar] [CrossRef]
  10. Yue, M.; Han, Z. GAN-LSTM-Based ADS-B Attack Detection in the Context of Air Traffic Control. IEEE Internet Things J. 2023, 10, 12651–12665. [Google Scholar] [CrossRef]
  11. Wang, J.; Zou, Y.; Ding, J. ADS-B spoofing attack detection method based on LSTM. EURASIP J. Wirel. Commun. 2020, 2020, 160. [Google Scholar] [CrossRef]
  12. Habler, E.; Shabtal, A. Using LSTM encoder-decoder algorithm for detecting anomalous ADS-B Messages. Comput. Secur. 2019, 78, 155–173. [Google Scholar] [CrossRef]
  13. Luo, P.; Wang, B.; Li, T.; Tian, J. ADS-B anomaly data detection model based on VAE-SVDD. Comput. Secur. 2021, 104, 102213. [Google Scholar] [CrossRef]
  14. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
  15. Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial Machine Learning at Scale. arXiv arXiv:1611.01236, 2016.
  16. Habler, E.; Shabtai, A. Analyzing Sequences of Airspace States to Detect Anomalous Traffic Condition. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 1843–1857. [Google Scholar] [CrossRef]
  17. Li, T.; Wang, B.; Shang, F.; Tian, J.; Cao, K. Dynamic temporal ADS-B data attack detection based on sHDP-HMM. Comput. Secur. 2020, 93, 101789. [Google Scholar] [CrossRef]
  18. Qiu, S.; Liu, Q.; Zhou, S.; Wu, C. Review of artificial intelligence adversarial attack and defense technologies. Appl. Sci. 2019, 9, 909. [Google Scholar] [CrossRef]
  19. Xu, H.; Ma, Y.; Liu, H.; Deb, D.; Liu, H.; Tang, J.; Jain, A. Adversarial attacks and defenses in images, graphs and text: A review. Int. J. Autom. 2019, 17, 151–178. [Google Scholar] [CrossRef]
  20. Yuan, X.; He, P.; Zhu, Q.; Li, X. Adversarial examples: Attacks and defenses for deep learning. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 2805–2824. [Google Scholar] [CrossRef]
  21. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  22. Goswami, G.; Ratha, N.; Agarwal, A.; Singh, R.; Vatsa, M. Unravelling robustness of deep learning-based face recognition against adversarial attacks. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  23. Fawaz, H.I.; Forestier, G.; Weber, J.; Idoumghar, L.; Muller, P.A. Adversarial attacks on deep neural networks for time series classification. In Proceedings of the 2019 International Joint Conference on Neural Networks, Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar]
  24. Shang, F.; Wang, B.; Li, T.; Tian, J.; Cao, K.; Guo, R. Adversarial Examples on Deep-Learning-Based ADS-B Spoofing Detection. IEEE Wirel. Commun. Lett. 2020, 9, 1734–1737. [Google Scholar] [CrossRef]
  25. Li, T.; Wang, B. Sequential collaborative detection strategy on ADS-B data attack. Int. J. Crit. Infrastruct. 2019, 24, 78–99. [Google Scholar] [CrossRef]
  26. Costin, A.; Francillon, A. Ghost in the Air Traffic: On insecurity of ADS-B protocol and practical attacks on ADS-B devices. In Proceedings of the Black Hat, Las Vegas, NV, USA, 21–26 July 2012; pp. 1–12. [Google Scholar]
  27. Yang, H.; Zhou, Q.; Yao, M.; Lu, R.; Li, H.; Zhang, X. A Practical and Compatible Cryptographic Solution to ADS-B Security. IEEE Internet Things J. 2019, 6, 3322–3334. [Google Scholar] [CrossRef]
  28. Kacem, T.; Wijesekera, D.; Costa, P. ADS-Bsec: A Holistic Framework to Secure ADS-B. IEEE Trans. Intell. Veh. 2018, 3, 511–521. [Google Scholar] [CrossRef]
  29. Strohmeier, M.; Lenders, V.; Martinovic, I. Intrusion Detection for Airborne Communication Using PHY-Layer Information. In Proceedings of the 12th International Conference on Detection of Intrusions and Malware, and Vulnerability Assessment, Milan, Italy, 9–10 July 2015; pp. 66–77. [Google Scholar]
  30. Monteiro, M. Detecting malicious ADS-B broadcasts using wide area multilateration. In Proceedings of the 2015 IEEE/AIAA 34th Digital Avionics Systems Conference, Prague, Czech Republic, 13–17 September 2015. [Google Scholar]
  31. Thill, M.; Konen, W.; Wang, H.; Bäck, T. Temporal convolutional autoencoder for unsupervised anomaly detection in time series. Appl. Soft Comput. 2021, 112, 107751. [Google Scholar] [CrossRef]
  32. Cho, K.; Merrienboer, B. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Doha, Qatar, 25–29 October 2014. [Google Scholar]
  33. Goel, A.; Singh, A.; Agarwal, A.; Vatsa, M.; Singh, R. Smartbox: Benchmarking adversarial detection and mitigation algorithms for face recognition. In Proceedings of the 2018 IEEE 9th International Conference on Biometrics Theory, Applications and Systems, Redondo Beach, CA, USA, 22–25 October 2018; pp. 1–7. [Google Scholar]
  34. Tian, J.; Wang, B. Adversarial Attacks and Defenses for Deep-Learning-Based Unmanned Aerial Vehicles. IEEE Internet Things J. 2022, 9, 22399–22409. [Google Scholar] [CrossRef]
  35. Jadidi, Z.; Pal, S.; Nayak, N.; Selvakkumar, A.; Chang, C.C.; Beheshti, M.; Jolfaei, A. Security of Machine Learning-Based Anomaly Detection in Cyber Physical Systems. In Proceedings of the 2022 International Conference on Computer Communications and Networks, Honolulu, HI, USA, 25–28 July 2022. [Google Scholar]
  36. Li, T.; Luo, W.; Shen, L.; Zhang, P.; Ju, X.; Yu, T.; Yang, W. Adversarial sample detection framework based on autoencoder. In Proceedings of the 2020 International Conference on Big Data & Artificial Intelligence & Software Engineering, Bangkok, Thailand, 30 October–1 November 2020. [Google Scholar]
  37. Tax, D.; Duin, R. Support Vector Data Description. Mach. Learn. 2004, 54, 45–66. [Google Scholar] [CrossRef]
  38. Duch, W.; Korczak, J. Optimization and global minimization methods suitable for neural networks. ACM Comput. Surv. 1998, 2, 163–212. [Google Scholar]
  39. Strohmeier, M.; Martinovic, I.; Fuchs, M.; Schäfer, M.; Lenders, V. OpenSky: A swiss army knife for air traffic security research. In Proceedings of the 2015 IEEE/AIAA 34th Digital Avionics Systems Conference, Prague, Czech Republic, 13–17 September 2015. [Google Scholar]
  40. Hosseini, H.; Chen, Y.; Kannan, S.; Zhang, B.; Poovendran, R. Blocking transferability of adversarial examples in black-box learning systems. arXiv arXiv:1703.04318, 2017.
  41. Papernot, N.; McDaniel, P.; Goodfellow, I.J.; Jha, S.; Celik, Z.B.; Swami, A. Practical black-box attacks against machine learning. In Proceedings of the 2017 ACM on Asia Conference on Computer and Communications Security, Abu Dhabi, United Arab Emirates, 2–6 April 2017; pp. 506–519. [Google Scholar]
  42. Papernot, N.; McDaniel, P.; Wu, X.; Jha, S.; Swami, A. Distillation as a defense to adversarial perturbations against deep neural networks. In Proceedings of the 2016 IEEE Symposium on Security and Privacy, San Jose, CA, USA, 22–26 May 2016; pp. 582–597. [Google Scholar]
  43. Tuli, S.; Casale, C.; Jennings, N.R. TranAD: Deep transformer networks for anomaly detection in multivariate time series data. Proc. VLDB Endow. 2022, 15, 1201–1214. [Google Scholar] [CrossRef]
  44. Deng, A.; Hooi, B. Graph Neural Network-Based Anomaly Detection in Multivariate Time Series. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021. [Google Scholar]
  45. Chen, J.; Michael, I. HopSkipJumpAttack: A Query-Efficient Decision-Based Attack. In Proceedings of the 2020 IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 18–21 May 2020; pp. 1277–1294. [Google Scholar]
  46. Zhang, J.; Li, B. Towards Efficient Data Free Blackbox Adversarial Attack. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 8–24 June 2022; pp. 15094–15104. [Google Scholar]
Figure 1. General overview of ADS-B.
Figure 1. General overview of ADS-B.
Sensors 24 03584 g001
Figure 2. General architecture of adversarial attacks against an ADS-B anomaly detection model.
Figure 2. General architecture of adversarial attacks against an ADS-B anomaly detection model.
Sensors 24 03584 g002
Figure 3. ADS-B anomaly flight track. (The anomaly type of the subgraph in the upper left corner is random position deviation. The upper right corner is velocity slow offset. The lower left corner is DOS. The lower right corner is altitude slow offset).
Figure 3. ADS-B anomaly flight track. (The anomaly type of the subgraph in the upper left corner is random position deviation. The upper right corner is velocity slow offset. The lower left corner is DOS. The lower right corner is altitude slow offset).
Sensors 24 03584 g003
Figure 4. Adversarial attacks against VAE (random position deviation).
Figure 4. Adversarial attacks against VAE (random position deviation).
Sensors 24 03584 g004
Figure 5. Adversarial attacks against VAE (velocity slow offset).
Figure 5. Adversarial attacks against VAE (velocity slow offset).
Sensors 24 03584 g005
Figure 6. Adversarial attacks against VAE (DOS).
Figure 6. Adversarial attacks against VAE (DOS).
Sensors 24 03584 g006
Figure 7. Adversarial attacks against VAE (altitude slow offset).
Figure 7. Adversarial attacks against VAE (altitude slow offset).
Sensors 24 03584 g007
Figure 8. Accuracy variation with respect to the amounts of perturbations.
Figure 8. Accuracy variation with respect to the amounts of perturbations.
Sensors 24 03584 g008
Figure 9. Adversarial attacks against LSTM. (The anomaly type of the subgraph in the upper left corner is random position deviation. The upper right corner is velocity slow offset. The lower left corner is DOS. The lower right corner is altitude slow offset).
Figure 9. Adversarial attacks against LSTM. (The anomaly type of the subgraph in the upper left corner is random position deviation. The upper right corner is velocity slow offset. The lower left corner is DOS. The lower right corner is altitude slow offset).
Sensors 24 03584 g009
Figure 10. Adversarial attacks against GRU. (The anomaly type of the subgraph in the upper left corner is random position deviation. The upper right corner is velocity slow offset. The lower left corner is DOS. The lower right corner is altitude slow offset).
Figure 10. Adversarial attacks against GRU. (The anomaly type of the subgraph in the upper left corner is random position deviation. The upper right corner is velocity slow offset. The lower left corner is DOS. The lower right corner is altitude slow offset).
Sensors 24 03584 g010aSensors 24 03584 g010b
Figure 11. Adversarial attacks against LSTM-Encoder-Decoder. (The anomaly type of the subgraph in the upper left corner is random position deviation. The upper right corner is velocity slow offset. The lower left corner is DOS. The lower right corner is altitude slow offset).
Figure 11. Adversarial attacks against LSTM-Encoder-Decoder. (The anomaly type of the subgraph in the upper left corner is random position deviation. The upper right corner is velocity slow offset. The lower left corner is DOS. The lower right corner is altitude slow offset).
Sensors 24 03584 g011
Figure 12. LSTM Accuracy variation with respect to the amounts of perturbations.
Figure 12. LSTM Accuracy variation with respect to the amounts of perturbations.
Sensors 24 03584 g012
Figure 13. GRU Accuracy variation with respect to the amounts of perturbations.
Figure 13. GRU Accuracy variation with respect to the amounts of perturbations.
Sensors 24 03584 g013
Figure 14. LSTM-Encoder-Decoder Accuracy variation with respect to the amounts of perturbations.
Figure 14. LSTM-Encoder-Decoder Accuracy variation with respect to the amounts of perturbations.
Sensors 24 03584 g014
Figure 15. Selection of the range of time neighborhood.
Figure 15. Selection of the range of time neighborhood.
Sensors 24 03584 g015
Table 1. Comparison of related ADS-B solutions.
Table 1. Comparison of related ADS-B solutions.
MethodsProtocol Modification Required or Additional Nodes RequiredAdvantages and Disadvantages
Traditional methodsEncryptionRequire modifying original ADS-B protocolDisadvantages: Encryption method needs to modify the original ADS-B protocol, which hinders system compatibility
Physical layer informationRequire ground stations and other entitiesDisadvantages: When attackers obtain prior knowledge of physical layer information through statistical analysis, the applicability of the method will need further verification for sophisticated ADS-B data attacks
MultilaterationRequire more than one sensor or entityDisadvantages: Multilateration method requires deploying multiple ground stations, which is not economical
Deep-learning-based anomaly detectionDo not require modifying protocol and additional nodesAdvantages: Deep learning method makes full use of ADS-B time series to detect anomaly data quickly and accurately
Table 2. Gaps of previous methods.
Table 2. Gaps of previous methods.
Name of the MethodGaps When the Methods Are Considered to Be Applied to ADS-B Anomaly Detection Models
MEA [24]The biggest flaw is that MEA assumes that the victim models are supervised and assumes ADS-B spoofing detection is a classification problem. However, there is scarce labeled ADS-B anomaly data in a real-world environment.
FGSM [14,35] and BIM [15,36]FGSM and BIM fail to take into account temporal correlation of ADS-B time series. Also, they cannot stabilize the update directions of adversarial samples and cannot escape from poor local optimum during the process of iterating.
Table 3. Confusion matrix for sample classification.
Table 3. Confusion matrix for sample classification.
Confusion MatrixDetected Normal DataDetected Anomaly Data
Real normal dataTPFN
Real anomaly dataFPTN
Table 4. Experiment dataset.
Table 4. Experiment dataset.
Anomaly TypesData Simulated Methods
Random position deviationInject Gaussian noise with a mean value 0 and a variance 0.5 into the actual latitude and longitude
Velocity slow offsetEnlarge the velocity with a multiple of 5 gradually
DOSThe aircraft track disappears in the air traffic surveillance system
Altitude slow offsetEnlarge the altitude with a multiple of 100 gradually
Table 5. Adversarial attack performance against VAE (%).
Table 5. Adversarial attack performance against VAE (%).
Evaluation IndexNo Adversarial AttackFGSMBIMTNAI-FGSM
FPR4.9325.7560.5991.05
FNR7.716.835.292.44
Accuracy93.6983.6867.0651.55
F1_score93.7981.9954.4713.59
Table 6. Adversarial attack performance against LSTM, GRU and LSTM-Encoder-Decoder (%).
Table 6. Adversarial attack performance against LSTM, GRU and LSTM-Encoder-Decoder (%).
Evaluation IndexNo Adversarial AttacksFGSMBIMTNAI-FGSM
LSTMFPR9.0852.1385.1194.46
FNR10.733.712.852.73
Accuracy90.0972.0856.0251.44
F1_score90.1763.1525.2910.29
GRUFPR9.0250.3283.3592.75
FNR10.893.842.932.82
Accuracy90.0373.1757.0951.97
F1_score90.1164.2426.2711.25
LSTM-Encoder-DecoderFPR7.9135.9967.3590.95
FNR8.607.145.483.55
Accuracy91.7578.4463.5952.74
F1_score91.7774.8247.2716.11
Table 7. Transferability of ADS-B adversarial attacks (% decrease in Accuracy).
Table 7. Transferability of ADS-B adversarial attacks (% decrease in Accuracy).
GRULSTM-Encoder-DecoderVAE
FGSM (LSTM)18.6317.4114.24
BIM (LSTM)23.5421.7417.48
TNAI-FGSM (LSTM)36.8735.3828.53
Table 8. Accuracy performance of ADS-B anomaly detection models (%).
Table 8. Accuracy performance of ADS-B anomaly detection models (%).
LSTMGRULSTM-Encoder-DecoderVAE
Before adversarial attacks90.0990.0391.7593.69
After FGSM attacks72.0873.1778.4483.68
After adversarial training84.2485.1386.9588.75
After BIM attacks56.0257.0963.5967.06
After adversarial training82.5882.8383.3283.09
After TNAI-FGSM attacks51.4451.9752.7451.55
After adversarial training75.4377.8578.2377.34
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Luo, P.; Wang, B.; Tian, J.; Liu, C.; Yang, Y. Adversarial Attacks against Deep-Learning-Based Automatic Dependent Surveillance-Broadcast Unsupervised Anomaly Detection Models in the Context of Air Traffic Management. Sensors 2024, 24, 3584. https://doi.org/10.3390/s24113584

AMA Style

Luo P, Wang B, Tian J, Liu C, Yang Y. Adversarial Attacks against Deep-Learning-Based Automatic Dependent Surveillance-Broadcast Unsupervised Anomaly Detection Models in the Context of Air Traffic Management. Sensors. 2024; 24(11):3584. https://doi.org/10.3390/s24113584

Chicago/Turabian Style

Luo, Peng, Buhong Wang, Jiwei Tian, Chao Liu, and Yong Yang. 2024. "Adversarial Attacks against Deep-Learning-Based Automatic Dependent Surveillance-Broadcast Unsupervised Anomaly Detection Models in the Context of Air Traffic Management" Sensors 24, no. 11: 3584. https://doi.org/10.3390/s24113584

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop