Next Article in Journal
Study on the Influence of Tip Clearance on Working Characteristics of High-Altitude Fan
Previous Article in Journal
Integrated Technologies for Anti-Deicing Functions and Structures of Aircraft: Current Status and Development Trends
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Trajectory Prediction Method Based on CNN, BiLSTM, and Multi-Head Attention Mechanism

1
School of Automation, Northwestern Polytechnical University, Xi’an 710129, China
2
School of Mechanical and Electrical Engineering, Lanzhou University of Technology, Lanzhou 730050, China
3
School of Engineering, Xi’an International University, Xi’an 710077, China
*
Author to whom correspondence should be addressed.
Aerospace 2024, 11(10), 822; https://doi.org/10.3390/aerospace11100822
Submission received: 7 September 2024 / Revised: 3 October 2024 / Accepted: 7 October 2024 / Published: 8 October 2024
(This article belongs to the Section Aeronautics)

Abstract

:
A four-dimensional (4D) trajectory is a multi-dimensional time series that embodies rich spatiotemporal features. However, its high complexity and inherent uncertainty pose significant challenges for accurate prediction. In this paper, we present a novel 4D trajectory prediction model that integrates convolutional neural networks (CNNs), bidirectional long short-term memory networks (BiLSTMs), and multi-head attention mechanisms. This model effectively addresses the characteristics of aircraft flight trajectories and the difficulties associated with simultaneously extracting spatiotemporal features using existing prediction methods. Specifically, we leverage the local feature extraction capabilities of CNNs to extract key spatial and temporal features from the original trajectory data, such as geometric shape information and dynamic change patterns. The BiLSTM network is employed to consider both forward and backward temporal orders in the trajectory data, allowing for a more comprehensive capture of long-term dependencies. Furthermore, we introduce a multi-head attention mechanism that enhances the model’s ability to accurately identify key information in the trajectory data while minimizing the interference of redundant information. We validated our approach through experiments conducted on a real ADS-B trajectory dataset. The experimental results demonstrate that the proposed method significantly outperforms comparative approaches in terms of trajectory estimation accuracy.

1. Introduction

With the rapid advancement of aviation, navigation, and unmanned technology, trajectory estimation has emerged as a crucial technology in these fields [1]. The accuracy and real-time performance of trajectory estimation are vital for ensuring navigation safety, enhancing navigation efficiency, and enabling intelligent autonomous navigation [2]. Traditional trajectory estimation methods, such as Kalman filtering and particle filtering, have demonstrated effectiveness in certain scenarios. However, their performance is often constrained when confronted with complex and dynamic navigation environments and uncertainties [3]. Consequently, the development of more advanced and efficient trajectory estimation methods has become a prominent research focus.
Recently, deep learning techniques [4], particularly CNNs [5] and long short-term memory networks (LSTMs) [6], have made significant strides in various domains, including image processing, natural language processing, and time series analysis [7]. The robust feature extraction and learning capabilities inherent in these network models provide distinct advantages in managing complex data patterns and sequence information [8]. Motivated by these advancements, numerous researchers have begun applying deep learning methods to the field of trajectory estimation, overcoming the limitations of traditional approaches and enhancing both the accuracy and robustness of trajectory predictions.
For instance, Rahman et al. [9] transformed trajectory data into images of varying sizes and subsequently classified these images using CNNs, achieving high accuracy on test sets involving two aircraft. Similarly, Wang et al. [10] optimized the trajectories of hypersonic aircraft during the re-entry phase by meticulously designing and constructing deep learning models. Their approach facilitated the efficient and rapid generation of optimal flight trajectories, thereby offering substantial technical support for stable re-entry under high-speed flight conditions.
Li et al. [11] developed a three-dimensional hypersonic trajectory prediction method leveraging deep neural networks (DNNs), addressing the challenge of deriving optimal flight paths under complex control scenarios. Han et al. [12] introduced a novel 4D trajectory prediction method utilizing LSTM, enabling the precise acquisition of 4D information related to predicted trajectories. Zeng et al. [13] employed regularization techniques to reconstruct historical trajectory data and proposed a data-driven LSTM model capable of effectively capturing the temporal dependencies of flight trajectories, resulting in high prediction accuracy.
Furthermore, Schimpf et al. [14] trained a recurrent neural network (RNN) on multiple flight path trajectories to enhance the accuracy of trajectory predictions. Liu et al. [15] developed a trajectory prediction model based on DNNs that leverages the dynamic characteristics of aircraft. This model generates extensive flight trajectory data through established dynamic equations, conducts offline training on the generated data, and subsequently applies it to real-time online predictions. This integration into flight control systems significantly improves aircraft safety.
While the aforementioned methods excel in the complex task of predicting aircraft tracks, single models often struggle to fully capture the spatial characteristics, temporal dependencies, and dynamic changes inherent in track data. This limitation can lead to inadequate prediction accuracy and generalization ability [16,17,18,19]. In contrast, composite models can comprehensively capture the features of trajectory data by leveraging the strengths of various models, thereby improving adaptability to new data and scenarios [20]. Consequently, researchers have started to implement composite models for trajectory estimation to enhance prediction accuracy and overall performance.
For instance, Wu et al. [21] developed a track prediction model that integrates CNNs, RNNs, and fully connected neural networks, validating the model with real data under severe convective weather conditions. Their results demonstrated that the prediction variance of the composite model was significantly lower than that of single models. Similarly, Tran et al. [22] achieved a 30% reduction in prediction error by combining encoder–decoder architectures, CNNs, and gated recurrent units (GRUs) for trajectory prediction. Furthermore, Shafienya et al. [23] proposed a hybrid deep learning model that incorporates CNNs, GRUs, and 3D-CNNs. This model specifically extracts spatial and temporal features from the input data using the CNN–GRU combination and employs a 3D-CNN for prediction. Experimental validation using automatic dependent surveillance–broadcast (ADS-B) data revealed that this method yielded lower measurement errors compared to traditional trajectory prediction techniques.
Although the above methods have shown certain advantages in trajectory prediction, there are still the following issues: (1) Traditional trajectory estimation methods often have limitations in processing spatial features, making it difficult to fully capture the complex spatial characteristics of trajectory data [24]. (2) When handling trajectory data, some anomalies or critical events may involve multiple positions in the sequence, and traditional single models often find it hard to simultaneously focus on these global dependencies [25]. In the method proposed in this paper, a CNN is used as an initial spatial feature extractor, efficiently capturing spatial features of the trajectory data, such as the aircraft’s position, speed, direction, and other key information. A BiLSTM not only processes forward temporal sequence information but also leverages backward temporal sequence information, thereby gaining a more comprehensive understanding of the temporal characteristics of trajectory data. The multi-head attention mechanism is used to further enhance the model’s ability to model global dependencies. It dynamically adjusts the importance of different features, allowing the model to focus more accurately on critical information during prediction. Specifically, the main contributions of this paper are as follows:
(1) A specific trajectory estimation framework has been proposed, which fully utilizes the modeling capability of BiLSTMs for capturing long-term dependencies in sequential data, especially considering the temporal continuity of trajectory data and its correlation with past and future states. Meanwhile, CNNs are employed to extract key local features such as curvature and velocity changes from the trajectory data, which are crucial for capturing the dynamic characteristics of the trajectories.
(2) To further enhance the accuracy of trajectory estimation, we introduced a multi-head attention mechanism into the framework to weight and fuse the extracted features, thereby highlighting key information within the trajectories. The advantage of this mechanism lies in its ability to allow the model to simultaneously focus on multiple important parts of the sequence across various representation subspaces, which is particularly critical for handling potential turning points or anomalies that may appear within the trajectories.
(3) In addition, the proposed trajectory estimation framework emphasizes not only high-accuracy predictions but also the improvement of real-time performance and robustness. Through carefully designed network architecture and parameter optimization, as well as the introduction of efficient parallel computing techniques, this framework can process large-scale trajectory data and respond quickly, while also demonstrating good noise resistance and generalization performance.
The remaining chapters of this paper are organized as follows: Section 2 provides a brief introduction to the fundamental theoretical concepts of CNNs, LSTMs, and BiLSTMs. Section 3 presents a detailed explanation of the proposed trajectory estimation method. Section 4 validates the effectiveness and accuracy of this method using real ADS-B trajectory data. Finally, Section 5 concludes the paper.

2. Theoretical Foundations

2.1. Convolutional Neural Networks (CNNs)

A CNN is a typical feed-forward neural network composed primarily of convolutional layers, pooling layers, and fully connected layers. The typical structure of a CNN is illustrated in Figure 1 [26].
The convolutional layer applies convolution operations to the feature map generated by the previous layer using convolutional kernels, resulting in a new feature map. The calculation process is as follows:
g j l = σ ( i = 1 n x i l 1 w j i l + b j l )
where i is the serial number of the input feature map, j is the serial number of the output feature map, l is the number of layers, x i l 1 is the ith input feature map of the (l – 1)th layer, w i j l is the weight of the convolution kernel, b j l is the bias term, is the sign of the convolution operation, and σ is the activation function.
The pooling layer is employed to reduce the number of internal parameters in a CNN and includes two methods: max pooling and average pooling [27]. This paper utilizes max pooling, and the calculation formula is as follows:
x l j = max ( x l 1 i )
where max ( ) is the maximum pooling function, which is used to calculate the maximum value of the feature map x l 1 i in the pooled region.
The fully connected layer transforms the two-dimensional (2D) feature maps generated by the convolutional and pooling layers into one-dimensional (1D) vectors, thereby completing the feature extraction task. The calculation formula is as follows:
o j l = i = 1 ( x i l 1 × w j i l + b j l )

2.2. Long Short-Term Memory Network (LSTM)

An LSTM is a specialized structure of an RNN designed to address the gradient vanishing and exploding problems that traditional RNNs encounter when processing long sequences [28]. By incorporating gating mechanisms, the LSTM has achieved significant success in various sequence processing tasks, allowing the network to effectively learn long-term dependencies within the sequence. Unlike traditional RNNs, LSTMs introduce three gating units: the input gate, oblivion gate, and output gate, as illustrated in Figure 2. The calculation process of LSTMs is as follows:
The input gate controls the addition of information and is calculated as follows:
i t = ς ( x t w x i + h t 1 w h i + b i )
The oblivion gate controls the discarding of information and is calculated as:
f t = ς ( x t w x f + h t 1 w h f + b f )
c t represents the internal cell state and is calculated as:
C ˜ t = tanh ( x t w x c + h t 1 w h c + b c )
C t = f t C t 1 + i t C ˜ t
The output gate controls the output of the message and is calculated as follows:
o t = ς ( x t w x o + h t 1 w h o + b o )
h t = O t tanh ( C t )
where i t is the input gate; f t is the oblivion gate; o t is the output gate; x t denotes the input at time t ; ς is the sigmoid activation function; w x i , w x f , w x c , w x o denote the weight matrix associated with x t ; w h i , w h f , w h c , w h o denote the weight matrix associated with h t 1 ; b i , b f , b c , b o are the bias vectors; and denotes the Hadamard product.

2.3. Bidirectional Long Short-Term Memory Network (BiLSTM)

A BiLSTM combines the principles of an LSTM with a bidirectional RNN. It consists of two parallel LSTM layers: a forward LSTM and a backward LSTM, as illustrated in Figure 3. At each time step, the forward LSTM processes the current part of the sequence while incorporating future information, whereas the backward LSTM processes the same part of the sequence while considering past information [29]. The outputs from these two LSTM layers are then merged using an activation function to produce a final output that accounts for both preceding and following information in the sequence.
This method can effectively capture the forward and backward information of trajectory data. Among them, h t and h t represent the hidden vectors of the forward LSTM layer and the backward LSTM layer at time t . They are independent of each other and only related to their respective LSTM layers. y t is obtained by weighted connections between these two hidden layers, and ς represents the activation function. The specific calculation process is shown in Equations (10) and (11).
h t = LSTM ( x t , h t 1 )
h t = LSTM ( x t , h t + 1 )

3. Proposed Method

3.1. CNN-BiLSTM Incorporating Multi-Head Attention Mechanism

When processing sequential data, particularly complex sequences with both spatial and temporal dependencies, a single network structure often struggles to capture all essential information. To address this challenge, we propose a new model that combines BiLSTMs and CNNs, integrating a multi-head attention mechanism. This network architecture aims to simultaneously capture both spatial and sequential features of the data while enhancing the focus on key information through the multi-head attention mechanism, thereby improving the overall performance of the model.
The proposed model structure is illustrated in Figure 4. We first employ a BiLSTM to capture bidirectional dependencies in the sequence data. The BiLSTM consists of a forward LSTM and a backward LSTM, which, respectively, capture forward and backward information in the sequence, fusing this information to obtain a more comprehensive representation of the sequence. Next, we use the output of the BiLSTM as the input for the CNN, leveraging the CNN’s convolution operation to capture the local spatial features of the data. The introduction of the CNN not only enhances the model’s understanding of the spatial structure of the data but also further enriches the model’s representational capability.
Although the combination of a BiLSTM and a CNN can effectively capture spatial features and bidirectional dependencies in sequences, certain parts of the data may hold greater importance than others in specific cases. To enhance the model’s focus on key information, we introduced a multi-head attention mechanism. This mechanism allows the model to concentrate on different segments of the input sequence from multiple perspectives and assign varying weights to each segment. By effectively highlighting key information and suppressing less relevant parts, the multi-head attention mechanism improves the overall performance of the model. A detailed explanation of the multi-head attention mechanism will be provided in the next section.

3.2. Multi-Head Attention Mechanism

Self-attention is a mechanism that enables models to allocate varying levels of attention to different positions in the input sequence when processing sequential data. In this mechanism, the model calculates the correlation between each position in the input sequence and all other positions, generating a new weighted representation based on these correlation scores.
The core idea of the self-attention mechanism is to capture the dependency relationships between different positions in the sequence by calculating the similarity score between each position in the input sequence and all other positions. Specifically, given an input sequence x i , each position in x i is first linearly transformed to generate three vectors: query (q), key (k), and value (v). The query vector is used to match other positions in the sequence, the key vector is used to calculate similarity with the query vector, and the value vector contains the information that needs to be extracted. The calculation formula is as follows:
{ q i = w q i · x i k i = w k i · x i v i = w v i · x i
where x i is the input sequence data; w q i , w k i , w v i are three different learnable weight matrices; and q i , k i , v i are the computed query vectors, key vectors and value vectors.
After obtaining the query vector and key vector, the self-attention mechanism calculates the dot product between the query vector and all key vectors in the sequence. These dot product values indicate the similarity or correlation between the query position and other positions in the sequence. To obtain normalized attention weights, the dot product values are transformed using a SoftMax function. The SoftMax function converts the original dot product values into a probability distribution, ensuring that the sum of weights for all positions equals 1. Consequently, the weight of each position represents its importance in the current context. The calculation formula is as follows:
a i , j = S o f t m a x ( q i k i )
where a i , j is the normalized attention weight, is the dot product operation symbol, and S o f t m a x is used to normalize the result of the dot product operation, which is calculated as follows:
S o f t m a x ( x i ) = e x i i = 1 m e x i ( 0 , 1 )
After obtaining the attention weights, the self-attention mechanism weights and sums these weights with the corresponding value vectors. Specifically, this involves multiplying the value vectors of each position by their corresponding weights and then summing these weighted value vectors, as illustrated in Figure 5. This process serves as a weighted aggregation of information from all positions in the sequence, with more important positions receiving greater weight. The final output representation is the result of this weighted aggregation for the current position, incorporating information from all positions in the sequence, filtered and fused according to the significant attention weights. The calculation formula is as follows:
O j = i = 1 m a j , i v i
The multi-head attention mechanism is an enhancement and extension of the self-attention mechanism. By introducing multiple attention heads, it enables the model to learn different aspects of the input data in parallel across various representation subspaces. This approach allows for a more effective capture of the dependency relationships between different positions in the input sequence, as illustrated in Figure 6.
Specifically, the multi-head attention mechanism first divides the input trajectory data into multiple smaller segments, each corresponding to a different attention head. Each head then independently performs self-attention calculations to generate its own output tensor. Finally, the outputs from all heads are concatenated and transformed into the final output representation through a linear transformation. Its expression is as follows:
M u l t i H e a d ( q , k , v ) = C o n c a t e n a t e ( h e a d 1 , h e a d 2 , , h e a d n ) w o
where h e a d i = A t t e n t i o n ( q w q i , k w k i , v w v i ) , w q i , w k i , w v i are the mapping matrix weights, and w o is the output weight matrix.

3.3. Network Structure Parameters

The structural parameters of the network are presented in Table 1. Its core components consist of two layers of LSTM and two layers of convolutional layers. The number of neurons in these layers is 128, 256, 32, and 64, respectively. This design facilitates the network’s ability to extract and abstract features from the input data in a layered manner.
To enhance the model’s performance, we introduced both Dropout and batch normalization (BN) layers. The Dropout layer operates by randomly “turning off” or “discarding” a portion of neurons during training, which means that in each iteration, the network learns with a slightly different structure. This approach compels the model to learn more robust features, as it must depend on the remaining active neurons for predictions. Consequently, this not only helps prevent overfitting but also increases the model’s diversity and improves its generalization ability on unseen data.
Conversely, the BN layer serves to standardize the inputs of each layer within the network, adjusting them to a relatively stable distribution. This helps mitigate the issue of internal covariate shift during training, making it easier for the network to learn the data distribution. Additionally, the BN layer can accelerate the training process by allowing the use of higher learning rates and alleviating the vanishing or exploding gradient problems. However, it is important to note that BN is not commonly used within LSTM layers, as LSTMs inherently manage changes in sequence data through their internal gating mechanisms. Therefore, we only apply a BN layer after the convolutional layers.
The hyperparameter settings of the network are detailed in Table 2. We employed sliding window technology, setting the window size to 10, to capture relevant contextual information when processing sequence data. To optimize the parameters within the network, we selected the Adam optimizer, which adaptively adjusts the learning rate based on gradient information during training. This choice accelerates the training process and enhances the model’s performance. For evaluating the model’s performance, we utilized the mean squared error (MSE) as the loss function, which quantifies the average difference between the model’s predicted and actual values.
In addition, we set the maximum number of iterations to 500 to ensure that the model has enough training epochs to converge to a better solution. To effectively utilize computational resources and accelerate the training process, we determined the optimal batch size and test set proportion for the model through experimentation, as shown in Figure 7. Specifically, the batch size is set to 32, meaning that the network will process 32 samples simultaneously in each iteration. Finally, regarding the dataset partitioning, we set the proportion of the test set to 0.3, which means we will use 70% of the input data as the training set to train the model, while the remaining 30% will be used as the test set to evaluate the model’s generalization ability.

3.4. Network Training Process

Based on the proposed att-CNN-BiLSTM model, combined with trajectory data sample partitioning and model testing, a complete trajectory prediction process can be obtained, as shown in Algorithm 1 and Figure 8. The main steps can be summarized as follows:
Algorithm 1. att-CNN-BiLSTM
Require: Trajectory data; learning rate: η; batch size: m; max epoch: θ; window_size: δ; optimizer: Adam.
While epoch ≤ max epoch:
1:  for i in range(len(data) − window_size):
2:    X.append(data[i:i + window_size])
3:    y.append(data[i + window_size])
4:    i += 1
5:  Normalization of input trajectory data;
6:  Split the dataset into a training set and a testing set, where the testing set is 30%;
7:  Reshaping the input data to match the input shape of the LSTM;
8:  Construct att-CNN-BiLSTM model and input training set data for training;
9:  The forward and reverse information of the trajectory data is captured through Equations (10) and (11);
10: Enhancing the model’s focus on key information through Equations (12)–(16);
11: Local spatial features in the data are captured through Equation (1);
12: Prediction using testing set;
13: Output MSE, RMSE, MAE, R2;
14: epoch += 1
end while
(1) Divide the preprocessed trajectory into training and test sets using a sliding time window based on the test set share determined in Figure 7.
(2) Construct an att-CNN-BiLSTM model based on the structural parameters of each layer in Table 1. Specifically, using BiLSTMs as the first two layers of the model to capture the bidirectional dependencies of the sequence. After the BiLSTMs, add a multi-head attention mechanism layer to calculate the attention weights of different parts in the input sequence. Next, use CNN layers to capture local features.
(3) Set the learning rate, optimizer, and loss function of the model.
(4) Input the training set data into the att-CNN-BiLSTM model for training, update the parameters in the network using the Adam optimizer, calculate the loss of the model using the MSE loss function, and end the training when the loss value is less than or equal to the set value.
(5) Judge whether the number of training times reaches the set epoch threshold; if not, repeat step 4; if yes, end the training.
(6) Finally, the trained model is predicted using the test set and several performance evaluation metrics are output, including mean square error (MSE), root mean square error (RMSE), mean absolute error (MAE), and coefficient of determination (R2). These metrics can comprehensively assess the prediction accuracy and generalization ability of the model and help us understand the real performance of the model.

4. Experimental Validation

In this section, we validate the performance of the proposed method through experimental verification using real ADS-B trajectory data. The experiments were conducted on a computer equipped with an Intel i7-10750H CPU, an Nvidia GTX 1650Ti GPU, and 16 GB of memory. The network was successfully trained using software platforms, including Python 3.7, TensorFlow 2.3, and Keras 2.7.

4.1. Dataset Description

The ADS-B dataset [25] represents a crucial advancement in air traffic management technology. It offers real-time insights into air traffic for managers by automatically transmitting essential information, such as an aircraft’s position, speed, and heading. This automatic broadcasting surveillance system enhances the efficiency and safety of air traffic operations while also serving as a valuable resource for aviation research and development.
In this study, we analyzed the actual flight trajectory data of the aircraft with registration number B1520 from May 2024. To gain a deeper understanding of the aircraft’s flight characteristics, we selected a comprehensive dataset that includes various flight-related information, such as aircraft identification codes, timestamps, altitude, speed, heading angle, longitude, and latitude.
For our model training and experimental analysis, we concentrated on five core parameters: altitude, speed, heading angle, longitude, and latitude. These parameters not only illustrate the aircraft’s dynamic flight state but are also essential for understanding the interplay of various factors throughout the flight process.
Table 3 presents detailed statistical information for these five parameters, including key indicators such as minimum, maximum, mean, and standard deviation. These data provide an overall view and the distribution characteristics of the dataset. Additionally, we created visual representations of these data (as shown in Figure 9) to intuitively illustrate the distribution patterns and trends.

4.2. Experimental Results and Analysis

4.2.1. Effect of Different Numbers of Neurons on Modeling

When constructing a deep learning model, the number of neurons is a critical hyperparameter that significantly influences the model’s capacity for representation and learning. Different configurations of neuron counts can profoundly affect the overall performance of the model.
In the case of a BiLSTM layer, a small number of neurons limits the model’s ability to represent complex patterns and capture long-term dependencies within sequence data. With fewer neurons, the more the model’s information processing capacity is restricted, which may lead to underfitting, characterized by poor performance on the training set. Conversely, as the number of neurons increases, the model’s representation capacity improves. However, an excessive number of neurons can lead to overfitting, where the model performs well on the training set but struggles on the test set due to its tendency to learn noise and outliers present in the training data.
Similarly, for CNN layers, a limited number of neurons may hinder the extraction of essential features from the input data. Just like in the BiLSTM layer, having too many neurons in the CNN layer can also result in overfitting.
Therefore, to determine the optimal neuron count configuration, we compared the model performance under different configurations through experimental methods, and, thus, selected the optimal neuron count setting. Specifically, we fixed the number of layers of the BiLSTM and CNN to 2 and referred to the literature [30] to choose from the commonly used parameters such as 32, 64, 128, and 256. Due to too many considerations, we set the number of neurons in the latter layer of the BiLSTM and CNN to twice that of the previous layer. The experimental results are shown in Table 4. Among them, {32, 64, 128, 256} indicates that the number of neurons in the two layers of the BiLSTM is 32 and 64, respectively, and the number of neurons in the two layers of the CNN is 128 and 256, respectively. We chose four metrics, MSE, RMSE, MAE, and R2, to evaluate the performance of the model. The calculation formula for these four indicators is as follows:
If the true value is y = { y 1 , y 2 , , y n } and the predicted value is y ^ = { y ^ 1 , y ^ 2 , , y ^ n } , then
M S E = 1 n i = 1 n ( y ^ i y i ) 2
MSE is generally used to detect the deviation between the predicted and true values of a model. With a range of [ 0 , + ) , the larger the value, the greater the error.
R M S E = 1 n i = 1 n ( y ^ i y i ) 2
RMSE stands for the root of MSE, which represents the sample standard deviation of the difference between the predicted value and the true value. The range [ 0 , + ) is equal to 0 when the predicted value completely matches the true value.
M A E = 1 n i = 1 n | y ^ i y i |
With a MAE range of [ 0 , + ) , the smaller the value, the better accuracy the predictive model possesses.
R 2 = 1 i = 1 n ( y ^ i y i ) 2 i = 1 n ( y ^ i y ¯ ) 2 y ¯ = 1 n i = 1 n y i
Based on the value of R 2 , the quality of the model can be judged. The larger the value, the better the fitting effect of the model.
In Table 4, we evaluate the five signals—height, speed, angle, longitude, and latitude—using four performance indicators: MSE, RMSE, MAE, and R2. To emphasize the best performance for each signal across these indicators, we present the optimal values in bold. The model exhibits optimal performance when the number of neurons is set to {128, 256, 32, 64}. Consequently, we selected these neuron configurations for our model, and the detailed parameters are outlined in Table 1.

4.2.2. Comparative Analysis of Different Methods

To further validate the effectiveness of the proposed method, we selected several related techniques for comparative analysis, specifically: (1) real data: ADS-B real trajectories; (2) proposed: our method; (3) encoder–decoder [22]; (4) CNN-3DCNN-GRU [23]; (5) TCN-Informer [20]; and (6) attention-TCN-GRU [31]. The experimental results are presented in Figure 10. Panels (a) to (e) illustrate the comparison between predicted and actual trajectories for each model across five different signals: altitude, speed, angle, longitude, and latitude. The predicted trajectory of our proposed method aligns closely with the overall trend of the actual trajectory, whereas the other methods exhibit noticeable errors. This further demonstrates the effectiveness and robustness of our approach.
To provide a clearer evaluation of the performance of various methods, we plotted the results of the proposed method alongside those of the comparative methods across four indicators: MSE, RMSE, MAE, and R2, as illustrated in Figure 11. For indicators (a) to (c), larger values correspond to greater errors, indicating that the proposed method exhibits smaller prediction errors across five datasets: altitude, speed, angle, longitude, and latitude. Conversely, for indicator (d), larger values signify higher prediction accuracy. The proposed method demonstrated superior prediction accuracy for all five datasets, further confirming its enhanced trajectory prediction performance compared to the other methods.

4.2.3. Ablation Experiment

To verify the effectiveness of the improvements presented in this paper, we designed a series of ablation experiments, outlined as follows: (1) proposed: our method; (2) CNN-BiLSTM: removal of the multi-head attention mechanism; (3) att-CNN: replacing the BiLSTM in the model with a CNN; (4) att-BiLSTM: replacing the CNN in the model with a BiLSTM; (5) att-CNN-LSTM: replacing the BiLSTM in the model with an LSTM; (6) att1-CNN-BiLSTM: replacing the multi-head attention mechanism in the model with a self-attention mechanism.
The experimental results are illustrated in Figure 12. First, we focused on the three indicators shown in panels (a) to (c), which measure the magnitude of prediction error. In the ablation experiments, we systematically removed each key component or improvement point from the method and re-evaluated the model’s performance. By comparing the error metrics before and after the removal of each component, we found that eliminating any key element led to a significant increase in prediction error. This clearly indicates that each improvement in our method plays a crucial role in reducing prediction errors.
Next, we analyzed the indicator shown in panel (d), which represents the accuracy of the model’s predictions. During the ablation experiments, we observed that as key components of our method were gradually reintroduced, the value of this indicator also steadily increased, demonstrating continuous improvement in the model’s prediction accuracy. This highlights the effectiveness of our enhanced method in boosting prediction accuracy.
From this series of ablation experiments, we concluded that the proposed method demonstrates superior predictive performance across five datasets: altitude, speed, angle, longitude, and latitude. Whether evaluated from the perspective of reducing prediction errors or enhancing prediction accuracy, our method significantly outperforms the comparative methods. This further validates that our improvements to existing trajectory prediction methods are effective and possess significant practical application value.

5. Conclusions

In this paper, we propose an innovative trajectory estimation method that integrates CNNs, BiLSTMs, and multi-head attention mechanisms. This method not only highlights the advantages of CNNs in accurately capturing local features (such as curvature and velocity changes) within trajectory data, but also fully utilizes the capability of BiLSTMs in modeling long-term dependencies, especially considering the temporal characteristics of trajectory data and its associations with past and future states. Additionally, by incorporating a multi-head attention mechanism, the method can globally focus on key information within the trajectories, thereby enhancing the accuracy and robustness of trajectory estimation.
To validate the effectiveness of the proposed method, we conducted experiments using a real ADS-B trajectory dataset. The experimental results demonstrate that this method significantly outperforms traditional approaches in both accuracy and robustness of trajectory estimation. It not only overcomes the limitations faced by conventional trajectory estimation methods when dealing with complex trajectory data but also brings substantial technical breakthroughs to the field of trajectory estimation.
In the future, we will continue to explore and improve the proposed trajectory estimation method. On one hand, we will attempt to introduce more deep learning techniques and strategies, such as temporal convolutional networks and graph neural networks, to further enhance the performance of the model. On the other hand, we will also focus on the challenges and needs in practical applications to promote the effective application and further development of the proposed method in the field of trajectory estimation.

Author Contributions

Methodology, Y.X.; software, B.H.; writing—original draft preparation, Y.X.; writing—review and editing, Z.W.; supervision, Z.W. and Q.P.; data curation, B.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was partially supported by the National Natural Science Foundation of China (Grant Nos. 61790552, 62203358, 62233014).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

ADS-B datasets can be downloaded from https://flightadsb.variflight.com/track-data (accessed on 24 February 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hu, Y.; Yi, J.; Cheng, F.; Wan, X.; Hu, S. 3-D Target Tracking for Distributed Heterogeneous 2-D–3-D Passive Radar Network. IEEE Sens. J. 2023, 23, 29502–29512. [Google Scholar] [CrossRef]
  2. Bartusiak, E.R.; Jacobs, M.A.; Chan, M.W.; Comer, M.L.; Delp, E.J. Predicting Hypersonic Glide Vehicle Behavior with Stochastic Grammars. IEEE Trans. Aerosp. Electron. Syst. 2024, 60, 1208–1223. [Google Scholar] [CrossRef]
  3. Hashemi, S.M.; Botez, R.M.; Grigorie, T.L. New Reliability Studies of Data-Driven Aircraft Trajectory Prediction. Aerospace 2020, 7, 145. [Google Scholar] [CrossRef]
  4. Chuya-Sumba, J.; Alonso-Valerdi, L.M.; Ibarra-Zarate, D.I. Deep-Learning Method Based on 1D Convolutional Neural Network for Intelligent Fault Diagnosis of Rotating Machines. Appl. Sci. 2022, 12, 2158. [Google Scholar] [CrossRef]
  5. Ruan, D.; Wang, J.; Yan, J.; Gühmann, C. CNN Parameter Design Based on Fault Signal Analysis and Its Application in Bearing Fault Diagnosis. Adv. Eng. Inform. 2023, 55, 101877. [Google Scholar] [CrossRef]
  6. Kosova, F.; Altay, Ö.; Ünver, H.Ö. Structural Health Monitoring in Aviation: A Comprehensive Review and Future Directions for Machine Learning. Nondestruct. Test. Eval. 2024, 1–60. [Google Scholar] [CrossRef]
  7. Zhu, Z.; Lei, Y.; Qi, G.; Chai, Y.; Mazur, N.; An, Y.; Huang, X. A Review of the Application of Deep Learning in Intelligent Fault Diagnosis of Rotating Machinery. Measurement 2023, 206, 112346. [Google Scholar] [CrossRef]
  8. Huang, J.; Li, Z.; Liu, D.; Yang, Q.; Zhu, J. An Adaptive State Estimation for Tracking Hypersonic Glide Targets with Model Uncertainties. Aerosp. Sci. Technol. 2023, 136, 108235. [Google Scholar] [CrossRef]
  9. Rahman, S.; Lapasset, L.; Mothe, J. Aircraft Conflict Resolution Using Convolutional Neural Network on Trajectory Image. In Intelligent Systems Design and Applications, Proceedings of the 21st International Conference on Intelligent Systems Design and Applications (ISDA 2021), Online, 13–15 December 2021; Springer: Cham, Switzerland, 2022. [Google Scholar]
  10. Wang, J.; Wu, Y.; Liu, M.; Yang, M.; Liang, H. A Real-Time Trajectory Optimization Method for Hypersonic Vehicles Based on a Deep Neural Network. Aerospace 2022, 9, 188. [Google Scholar] [CrossRef]
  11. Li, H.; Chen, H.; Tan, C.; Jiang, Z.; Xu, X. Fast Trajectory Generation with a Deep Neural Network for Hypersonic Entry Flight. Aerospace 2023, 10, 931. [Google Scholar] [CrossRef]
  12. Han, P.; Yue, J.; Fang, C.; Shi, Q.; Yang, J. Short-Term 4D Trajectory Prediction Based on LSTM Neural Network. In Proceedings of the Second Target Recognition and Artificial Intelligence Summit Forum, Changchun, China, 20–22 August 2019. [Google Scholar]
  13. Zeng, W.; Quan, Z.; Zhao, Z.; Xie, C.; Lu, X. A Deep Learning Approach for Aircraft Trajectory Prediction in Terminal Airspace. IEEE Access 2020, 8, 151250–151266. [Google Scholar] [CrossRef]
  14. Schimpf, N.; Wang, Z.; Li, S.; Knoblock, E.J.; Li, H.; Apaza, R.D. A Generalized Approach to Aircraft Trajectory Prediction via Supervised Deep Learning. IEEE Access 2023, 11, 116183–116195. [Google Scholar] [CrossRef]
  15. Liu, Z.; Yan, J.; Ai, B.; Fan, Y.; Luo, K.; Cai, G.; Qin, J. An Online Generation Method of Terminal-Area Trajectories for Wave-Rider Using Deep Neural Networks. Aerospace 2023, 10, 654. [Google Scholar] [CrossRef]
  16. Zhang, A.; Zhang, B.; Bi, W.; Mao, Z. Attention Based Trajectory Prediction Method under the Air Combat Environment. Appl. Intell. 2022, 52, 17341–17355. [Google Scholar] [CrossRef]
  17. Liu, Z.; Wang, Z.; Yang, Y.; Lu, Y. A Data-Driven Maneuvering Target Tracking Method Aided with Partial Models. IEEE Trans. Veh. Technol. 2024, 73, 414–425. [Google Scholar] [CrossRef]
  18. Zeng, W.; Chu, X.; Xu, Z.; Liu, Y.; Quan, Z. Aircraft 4D Trajectory Prediction in Civil Aviation: A Review. Aerospace 2022, 9, 91. [Google Scholar] [CrossRef]
  19. Pepper, N.; Thomas, M. Learning Generative Models for Climbing Aircraft from Radar Data. J. Aerosp. Inf. Syst. 2024, 21, 474–481. [Google Scholar] [CrossRef]
  20. Dong, Z.; Fan, B.; Li, F.; Xu, X.; Sun, H.; Cao, W. TCN-Informer-Based Flight Trajectory Prediction for Aircraft in the Approach Phase. Sustainability 2023, 15, 16344. [Google Scholar] [CrossRef]
  21. Wu, Y.; Yu, H.; Du, J.; Liu, B.; Yu, W. An Aircraft Trajectory Prediction Method Based on Trajectory Clustering and a Spatiotemporal Feature Network. Electronics 2022, 11, 3453. [Google Scholar] [CrossRef]
  22. Tran, P.N.; Nguyen, H.Q.V.; Pham, D.-T.; Alam, S. Aircraft Trajectory Prediction With Enriched Intent Using Encoder-Decoder Architecture. IEEE Access 2022, 10, 17881–17896. [Google Scholar] [CrossRef]
  23. Shafienya, H.; Regan, A.C. 4D Flight Trajectory Prediction Using a Hybrid Deep Learning Prediction Method Based on ADS-B Technology: A Case Study of Hartsfield–Jackson Atlanta International Airport (ATL). Transp. Res. Part C Emerg. Technol. 2022, 144, 103878. [Google Scholar] [CrossRef]
  24. Tong, Q.; Hu, J.; Chen, Y.; Guo, D.; Liu, X. Long-Term Trajectory Prediction Model Based on Transformer. IEEE Access 2023, 11, 143695–143703. [Google Scholar] [CrossRef]
  25. Yang, Z.; Kang, X.; Gong, Y.; Wang, J. Aircraft Trajectory Prediction and Aviation Safety in ADS-B Failure Conditions Based on Neural Network. Sci. Rep. 2023, 13, 19677. [Google Scholar] [CrossRef] [PubMed]
  26. Cao, H.; Shao, H.; Zhong, X.; Deng, Q.; Yang, X.; Xuan, J. Unsupervised Domain-Share CNN for Machine Fault Transfer Diagnosis from Steady Speeds to Time-Varying Speeds. J. Manuf. Syst. 2022, 62, 186–198. [Google Scholar] [CrossRef]
  27. Hu, B.; Liu, J.; Zhao, R.; Xu, Y.; Huo, T. A New Fault Diagnosis Method for Unbalanced Data Based on 1DCNN and L2-SVM. Appl. Sci. 2022, 12, 9880. [Google Scholar] [CrossRef]
  28. Smagulova, K.; James, A.P. A Survey on LSTM Memristive Neural Network Architectures and Applications. Eur. Phys. J. Spec. Top. 2019, 228, 2313–2324. [Google Scholar] [CrossRef]
  29. Gao, Y.; Wang, C.; Shen, J.; Wang, Z.; Liu, Y.; Chai, Y. Systematic Review and Network Meta-Analysis of Machine Learning Algorithms in Sepsis Prediction. Expert Syst. Appl. 2024, 245, 122982. [Google Scholar] [CrossRef]
  30. Jia, P.; Chen, H.; Zhang, L.; Han, D. Attention-LSTM Based Prediction Model for Aircraft 4-D Trajectory. Sci. Rep. 2022, 12, 15533. [Google Scholar] [CrossRef]
  31. Ma, L.; Meng, X.; Wu, Z. Data-Driven 4D Trajectory Prediction Model Using Attention-TCN-GRU. Aerospace 2024, 11, 313. [Google Scholar] [CrossRef]
Figure 1. CNN structure.
Figure 1. CNN structure.
Aerospace 11 00822 g001
Figure 2. LSTM structure.
Figure 2. LSTM structure.
Aerospace 11 00822 g002
Figure 3. BiLSTM structure.
Figure 3. BiLSTM structure.
Aerospace 11 00822 g003
Figure 4. Structure of the proposed model.
Figure 4. Structure of the proposed model.
Aerospace 11 00822 g004
Figure 5. Self-attention mechanism.
Figure 5. Self-attention mechanism.
Aerospace 11 00822 g005
Figure 6. Multi-head attention mechanism.
Figure 6. Multi-head attention mechanism.
Aerospace 11 00822 g006
Figure 7. Batch size and test set percentage settings.
Figure 7. Batch size and test set percentage settings.
Aerospace 11 00822 g007
Figure 8. Flowchart of network training.
Figure 8. Flowchart of network training.
Aerospace 11 00822 g008
Figure 9. Five signals over time. (a) Height. (b) Speed. (c) Angle. (d) Longitude. (e) Latitude.
Figure 9. Five signals over time. (a) Height. (b) Speed. (c) Angle. (d) Longitude. (e) Latitude.
Aerospace 11 00822 g009
Figure 10. Comparison of predicted and true trajectories of different methods. (a) Height. (b) Speed. (c) Angle. (d) Longitude. (e) Latitude.
Figure 10. Comparison of predicted and true trajectories of different methods. (a) Height. (b) Speed. (c) Angle. (d) Longitude. (e) Latitude.
Aerospace 11 00822 g010
Figure 11. Performance of different methods under four metrics. (a) MSE. (b) RMSE. (c) MAE. (d) R2.
Figure 11. Performance of different methods under four metrics. (a) MSE. (b) RMSE. (c) MAE. (d) R2.
Aerospace 11 00822 g011
Figure 12. Results of ablation experiments. (a) MSE. (b) RMSE. (c) MAE. (d) R2.
Figure 12. Results of ablation experiments. (a) MSE. (b) RMSE. (c) MAE. (d) R2.
Aerospace 11 00822 g012
Table 1. Structural parameters of the network layers.
Table 1. Structural parameters of the network layers.
LayersNetwork ParametersOutput Shape#Param
bidirectional_1LSTM (128, activation = ‘relu’, return_sequences = True),
input_shape = 10 × 1
10 × 256133,120
dropout_10.210 × 2560
bidirectional_2LSTM (256, activation = ‘relu’, return_sequences = True)10 × 5121,050,624
dropout_20.210 × 5120
Multi-head attention-10 × 5120
conv1d_1filters = 32, kernel_size = 3, activation = ‘relu’8 × 3249,184
BN_3-8 × 32128
dropout_30.28 × 320
max_pooling1d_1pool_size = 24 × 320
conv1d_2filters = 64, kernel_size = 3, activation = ‘relu’2 × 646208
BN_4-2 × 64256
dropout_40.22 × 640
max_pooling1d_2pool_size = 21 × 640
dense_132, activation = ‘relu’1 × 322080
dense_211 × 133
Table 2. Network hyperparameter settings.
Table 2. Network hyperparameter settings.
ParametersValue
window_size10
optimizerAdam
loss functionMSE
epochs500
batch_size32
testing_split0.3
Table 3. Trajectory data statistics.
Table 3. Trajectory data statistics.
ParametersMinimum ValuesMaximum ValuesMeanStandard Deviation
Height (m)312.4210,393.688424.5622372.032
Speed (m/s)37.04850.068745.8833124.0183
Angle (°)111329288.186231.27066
Longitude (°)87.46532114.073101.8867.601263
Latitude (°)33.993944.1185337.811953.028017
Table 4. The impact of different numbers of neurons on the model (Bold is the best value).
Table 4. The impact of different numbers of neurons on the model (Bold is the best value).
Number of NeuronsErrorHeightSpeedAngleLongitudeLatitude
{32, 64, 32, 64}MSE411.156212.22711.9360.3756.614
RMSE20.27714.5683.4540.6132.571
MAE19.41914.1693.2700.5952.562
R20.3540.4010.4000.632−5.270
{32, 64, 64, 128}MSE105.54079.56915.6542.1031.198
RMSE10.2738.9203.9561.4501.095
MAE10.0028.7203.8621.4490.063
R20.5270.4040.214−1.063−0.136
{32, 64, 128, 256}MSE187.838117.38114.5411.4572.599
RMSE13.70510.8343.8131.2071.612
MAE13.38110.6493.4571.1561.607
R20.7050.1210.270−0.429−1.464
{64, 128, 32, 64}MSE271.5761487.10511.35875.85840.001
RMSE16.48038.5633.3708.9986.324
MAE16.00414.9763.1868.0086.324
R20.5730.8880.4290.206−36.920
{64, 128, 64, 128}MSE188.3213803.36524.5530.5260.278
RMSE13.72361.6714.9550.7250.527
MAE13.51559.8104.7650.6510.496
R20.7040.715−0.2320.4840.736
{64, 128, 128, 256}MSE60.829133.1166.5072.2382.083
RMSE7.78011.5372.5511.4961.443
MAE7.58111.2691.4011.4761.439
R20.9040.8040.673−1.196−0.974
{128, 256, 32, 64}MSE40.90067.99014.8730.7650.026
RMSE6.3958.2463.8560.8750.162
MAE5.8948.1243.5470.8650.277
R20.9360.7910.8540.9480.907
{128, 256, 64, 128}MSE134.479100.1295.75642.8080.097
RMSE11.59710.0062.3996.5420.311
MAE11.1379.6131.5866.3060.109
R20.7890.6510.7110.8920.975
{128, 256, 128, 256}MSE50.7913685.08510.937111.7090.534
RMSE7.12760.7043.30710.5690.731
MAE6.88657.0983.11210.1160.658
R20.9200.7240.451−0.0950.493
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, Y.; Pan, Q.; Wang, Z.; Hu, B. A Novel Trajectory Prediction Method Based on CNN, BiLSTM, and Multi-Head Attention Mechanism. Aerospace 2024, 11, 822. https://doi.org/10.3390/aerospace11100822

AMA Style

Xu Y, Pan Q, Wang Z, Hu B. A Novel Trajectory Prediction Method Based on CNN, BiLSTM, and Multi-Head Attention Mechanism. Aerospace. 2024; 11(10):822. https://doi.org/10.3390/aerospace11100822

Chicago/Turabian Style

Xu, Yue, Quan Pan, Zengfu Wang, and Baoquan Hu. 2024. "A Novel Trajectory Prediction Method Based on CNN, BiLSTM, and Multi-Head Attention Mechanism" Aerospace 11, no. 10: 822. https://doi.org/10.3390/aerospace11100822

APA Style

Xu, Y., Pan, Q., Wang, Z., & Hu, B. (2024). A Novel Trajectory Prediction Method Based on CNN, BiLSTM, and Multi-Head Attention Mechanism. Aerospace, 11(10), 822. https://doi.org/10.3390/aerospace11100822

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop