Next Article in Journal
A Cross-View Geo-Localization Algorithm Using UAV Image and Satellite Image
Previous Article in Journal
Statistical Parameters Extracted from Radar Sea Clutter Simulated under Different Operational Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Method for Predicting Inertial Navigation System Positioning Errors Using a Back Propagation Neural Network Based on a Particle Swarm Optimization Algorithm

1
Wuhan Second Ship Research and Design Institute, Wuhan 430205, China
2
School of Intelligent Science and Engineer, Harbin Engineering University, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(12), 3722; https://doi.org/10.3390/s24123722
Submission received: 29 April 2024 / Revised: 4 June 2024 / Accepted: 4 June 2024 / Published: 7 June 2024
(This article belongs to the Section Navigation and Positioning)

Abstract

:
In order to reduce the position errors of the Global Positioning System/Strapdown Inertial Navigation System (GPS/SINS) integrated navigation system during GPS denial, this paper proposes a method based on the Particle Swarm Optimization–Back Propagation Neural Network (PSO-BPNN) to replace the GPS for positioning. The model relates the position information, velocity information, attitude information output by the SINS, and the navigation time to the position errors between the position information output by the SINS and the actual position information. The performance of the model is compared with the BPNN through an actual ship experiment. The results show that the PSO-BPNN can obviously reduce the position errors in the case of GPS signal denial.

1. Introduction

In the integrated navigation system, the primary navigation system operates continuously. Concurrently, other navigation systems provide correction data to enhance the positioning accuracy of the system and continuously supply navigation parameters [1,2,3,4]. The most common combination for integrated navigation systems is the employment of SINS as the primary navigation system, with the GPS serving as the auxiliary navigation system. When the GPS is denied, the positioning errors of the SINS will diverge rapidly over time. Currently, the introduction of the BPNN as a substitution for the GPS is a conventional approach to assist in the prediction and compensation of SINS position errors during GPS denial. The ability of the BPNN to handle complex nonlinear relationships enables the precise fitting of the relationship between the SINS position errors and variables associated with them. This helps with accurate prediction and the compensation of the SINS positioning errors.
Artificial neural networks have been widely applied in practical engineering. N. Shaukat et al. [5] proposed enhancing Radial Basis Function Neural Networks (RBFNN) to compensate for integrated navigation systems. Considering the variable time delays between GPS measurement intervals, a self-regressive neural network fusion architecture was introduced in [6]. Reference [7] proposed a method for the error feedback correction of Micro-Electro-Mechanical Systems—Strapdown Inertial Navigation System (MEMS-SINS) based on neural network prediction during GPS denial. It employed an RBF Neural Network, utilizing the outputs of the MEMS-SINS gyroscopes and accelerometers as inputs and the position, velocity, and attitude errors of the MEMS-SINS as outputs. The method is validated through ground vehicle tests, demonstrating that the MEMS-SINS can achieve high-precision navigation information through the neural network, enabling continuous navigation. In reference [8], to address the issue of the filter accuracy divergence caused by system uncertainty errors, a novel algorithm was introduced that uses a Model Predictive Filter (MPF) to predict and correct the error propagation in a feedforward neural network. This novel feedforward algorithm fits and learns the nonlinear relationships of the system, and through simulation data, it was verified that its application effects in integrated navigation were significantly superior to those of the Extended Kalman Filter (EKF) algorithm. In the study presented in [9], a Multilayer Perceptron (MLP) Neural Networks was suggested to provide pseudo-GPS positions for compensating INS errors in the case of GPS denial. To address the limitations of GPS and WIFI signals in complex indoor environments, the study in [10] proposed an indoor positioning technology that combined the real-time capabilities of inertial navigation with the high precision of machine learning. The validation demonstrated that this method can significantly reduce the positioning time while ensuring the accuracy of the location determination. In [11], a fuzzy neural network topology for an integrated GPS/INS/odometer system was constructed. Additionally, the work made predictions for integrated navigation sequences using adaptive fuzzy neurons, fuzzy inference systems, Recurrent Neural Networks (RNNs), and Wavelet Neural Networks. Reference [12] employed a multilayer BPNN instead of Kalman filtering for transfer alignment in inertial navigation systems, achieving comparable accuracy to Kalman filtering while effectively reducing the computational time and improving the real-time capability of the system. The BPNN is likely to become entangled in local optima, thereby failing to attain global optimal solutions. To address this issue, the authors of [13] proposed a novel fusion algorithm containing additional momentum and hidden unit competition to optimize the BPNN. It can efficiently predict the increment in the position and compensate for the INS error accumulation during a GPS outage. Southeast University has proposed an improved BPNN adaptive algorithm that integrates a Genetic Algorithm (GA) and has also optimized the training process of the GA [14]. In [15], the GA-BPNN was utilized to predict a ship’s trajectory and showed a higher prediction accuracy than traditional BPNN. In addition to the GA, there are other types of artificial intelligence methods, such as ensemble models [16] and Support Vector Machine (SVM) [17], which have been employed to overcome the propensity of neural networks to become entangled in local optima. However, these algorithms often need complex feature engineering and exhibit low computational efficiency. To overcome the drawbacks of the BPNN in handling long-sequence data and retaining information over extended periods, scholars have employed the Long Short-Term Memory (LSTM) Neural Network in handling inertial navigation problems. Huazhong University of Science and Technology [18] proposed a filtering technique for gyroscope random errors based on the LSTM Neural Network. By utilizing this model, the standard deviation of the filtered gyroscope random errors was reduced compared to the original data, thereby enhancing the measurement precision of the gyroscope. Y.X. Zhang [19] proposed an integration method based on the LSTM Neural Network, which can adaptively utilize historical data without the need for additional inputs.
As the output of a BPNN is entirely determined by the input at the current moment, this network is characterized by a more rapid response to anomalous changes in the vehicle’s trajectory when compared to RNNs such as an LSTM Neural Network in navigation problems. An INS exhibits complex strong nonlinear characteristics, and its navigation parameters show a high dimensionality. The BPNN demonstrates a strong capability in fitting complex nonlinear relationships and has the advantages of straightforward principles. Synthesizing these attributes, this paper intends to employ the BPNN to assist the integrated navigation system. According to the theory proposed in article [13], the characteristics of the INS errors include their variability during the training and predicting process, and based on the method, this work also adds navigation time into the input of the neural network. However, the BPNN is likely to become entangled in local optima, which can lead to significant prediction errors. Although many optimization algorithms were proposed in the article mentioned above, they were exclusively utilized for the optimization of the BPNN. Moreover, the GA algorithm exhibits a long computation time and a low searching efficiency, which cannot meet the real-time requirements of the inertial navigation system. The PSO algorithm can easily converge to the global optimum, and its convergence rate is faster than that of the GA. It also does well in identifying optimal values in high-dimensional data spaces. Considering the aforementioned advantages of the PSO and its ability to effectively satisfy the characteristics and practical needs of the inertial navigation systems during actual usage, this work employed the PSO algorithm to optimize the BPNN. By utilizing the weights obtained during each training iteration of the BPNN, the PSO algorithm was applied to identify the optimal weights. This process enables the BPNN to escape from the local optima, thereby enhancing the model’s generalization capability and subsequently optimizing the performance of the BPNN. Based on actual ship data, the study conducted comparative validation experiments between the PSO-BPNN model and the traditional BPNN model.
This paper addresses the diverging positioning accuracy of a SINS in integrated navigation systems during GPS denial, by optimizing the BPNN through the PSO algorithm. This approach aims to solve the application problems of integrated navigation systems in special environments.
The structural arrangement of this paper is as follows:
Section 1 elucidates the significance of the research presented in this paper. Section 2 exposits the error model of the SINS. Section 3 introduces the proposed BPNN, as well as the PSO algorithm designed for optimizing the BPNN. Section 4 details the environment and outcomes of the actual ship experiments. Section 5 offers a summary of the experimental results.

2. Error Equations of Strapdown Inertial Navigation [20]

This section presents the error equations of the SINS, explaining the propagation process of position errors.

2.1. Attitude Error Equation

The attitude error equation is as follow:
Φ ˙ n = Φ n × ( ω i e n + ω e n n ) + δ ω i e n + δ ω e n n ε n
where ω i e n is the rotational angular velocity of the navigation coordinate system relative to the Earth in the n frame, ω e n n is the rotational angular velocity of the Earth relative to the inertial space in the n frame, δ ω i e n and δ ω e n n correspond to the computational errors for ω i e n and ω e n n , Φ n represents the attitude error angle in the n frame, and ε n is the gyro drift in the n frame.

2.2. Velocity Error Equation

The velocity error equation is as follow:
δ V ˙ e n = f n × Φ n ( 2 δ ω i e n + δ ω e n n ) × V e n ( 2 ω i e n + ω e n n ) × δ V e n + δ g n + n
where f n represents the measured values of the accelerometer in the n frame, V e n is the velocity of the vehicle relative to the Earth in the n frame, δ V e n is the computational error of the V e n , δ g n is the error in gravitational acceleration, and n is the zero-offset error of the accelerometer in the n frame.

2.3. Position Error Equation

By applying the total differentiation on both sides of the position computation equation, the position error equation can be expressed as follows:
{ δ L ˙ = δ V y R M δ λ ˙ = δ V x R N c o s L + V x R N c o s L t a n L δ L
where L is the latitude, λ is the longitude, R N and R M correspond to the meridian radius of curvature and the prime vertical radius of curvature, and V x and V y represent the eastward and northward velocities.

2.4. Kalman Filtering

This work applied a loose integration model to the GPS and SINS. In this model, the Kalman filter receives the different output values of the same navigation parameter from the two navigation systems. After filtering calculations, it estimates the error quantities. The estimated values of the SINS’s errors are then used to correct the parameters output by the SINS, thereby obtaining the optimal estimation of the navigation parameters. Figure 1 shows the working process of the Kalman filter.
The Kalman filter employs a recursive approach for a priori calculation of state parameters, utilizing the state estimation value and the mean square error matrix of the state estimation from the previous moment to predict the state parameters. It combines the current moment’s observation values to calculate the state estimation value and the mean square error matrix of the state estimation for the current moment. The process follows a “prediction–correction” formulation for recursion, thereby enabling real-time estimation of the system’s state parameters.
The state equation and measurement equation of the Kalman filter are as follows:
{ X k = Φ k , k 1 X k 1 + ω k 1 z k = X k + v k
where X k and z k are the state vector and measurement vector at time instance k ; Φ k , k 1 is the state transition matrix, and H k is referred to the measurement matrix; ω k 1 and v k are the process noise vector and measurement noise vector. They are uncorrelated error vectors, and both are sequences of zero-mean Gaussian white noise vectors.
The more general expressions for the state one-step prediction and the state one-step prediction mean squared error matrix of the state estimation that can be obtained through derivation are as follows:
{ X ^ k , k 1 = Φ k , k 1 X ^ k 1 P k , k 1 = Φ k , k 1 P k 1 Φ k , k 1 T + Q k 1
where Q k 1 is the variance matrix of the system noise, which is required to be a known positive semi-definite matrix in the Kalman filter. Based on the recursive least squares estimation method, the current moment’s state estimation and the mean square error matrix of the state estimation can be derived as follows:
{ K k = P k , k 1 H k T ( H k P k , k 1 H k T + R k ) 1 X ^ k = X ^ k , k 1 + K k ( z k H k X ^ k , k 1 ) P k = ( I K k H k ) P k , k - 1 ( I K k H k ) T + K k R k K k T
where R k represents the variance matrix of the measurement noise, which is required to be a known positive definite matrix in the Kalman filter.
In (6), the first equation is the Kalman gain equation, and the matrix derived from it is the Kalman gain matrix. It is determined under the constraint criterion of minimizing the state variance and dictates the weight of the current estimation between one-step prediction and measurement. The second equation is the state estimation calculation equation, and the third equation is the estimation error covariance equation.

3. Optimization Method for the Backpropagation Neural Network

3.1. Principle of the Backpropagation Neural Network

The name of the Backpropagation Neural Network comes from the “backpropagation of error” process within its algorithm. Its structure can be divided into three components: the input layer, the hidden layer, and the output layer. Each layer consists of multiple neurons. The propagation process involves each neuron receiving inputs from the preceding layer of neurons, multiplying these inputs by their corresponding weights, summing them with the bias, and then undergoing a nonlinear transformation through the neuron’s activation function. The transformed result is subsequently passed on to the next layer of neurons.
The Backpropagation Neural Network employs a gradient descent method for self-learning and updating. When a set of desired outputs and input data are fed into the BPNN, the response values of the “neurons” propagate from the input layer through the hidden layer to the output layer. After being processed by the activation function, the actual output is obtained at the output layer. By comparing the difference between the actual output and the desired output, with the goal of minimizing this discrepancy, the weights and biases between neurons are updated layer by layer, starting from the output layer to the hidden layer and then to the input layer. This process of “forward computation of output–backward propagation of error” continues iteratively until the error is corrected to within a preset allowable error range, at which point the self-learning phase of the BPNN concludes. The Figure 2 shows the structure of a common BPNN.
Define W i j ( l ) as the weight between the j -th neuron of l 1 -th layer and the i -th neuron of l -th layer, and let b i ( l ) represent the bias of the i -th neuron in the output layer. The process can be expressed as follows:
n e t i ( l ) = j = 1 M l 1 W i j ( l ) h j ( l 1 ) + b i ( l ) h j ( l ) = f ( n e t i ( l ) )
where n e t i ( l ) is the input to i -th neuron in the output layer, f ( ) is the activation function, h j ( l 1 ) is the output of the j -th neuron of l 1 -th layer, and M l 1 is the number of neurons of l 1 -th layer.
Define the error function as E , with the minimum error ε and the maximum number of iterations set as M . Suppose there exists an m-dimensional input vector x = { x ( 1 ) , x ( m ) } ; let d ( p ) represent the network’s expected output for the p -th set of input data and y ( p ) represent the actual output of the network for the p -th set of input. Let n represent the number of neurons in the output layer. The symbol k represents the k -th neuron. The error function can be expressed as
E = 1 2 m p = 1 m k = 1 n ( d k ( p ) y k ( p ) ) 2
The partial derivative of the error function with the weights and biases for the p -th data set at the l -th layer can be expressed as follows:
E ( p ) W i j ( l ) = δ i ( l ) h j l 1 E ( p ) b i ( l ) = δ i ( l )
δ i ( l ) can be expressed as follows:
δ i ( l ) = k = 1 M l + 1 W k i ( l + 1 ) δ k ( l + 1 ) f ( x ) | x = n e t i ( l ) δ k ( l + 1 ) = ( d k ( p ) y k ( p ) ) f ( x ) | x = n e t i ( l + 1 )
Upon obtaining the partial derivatives, the weights and biases can be calculated. Subsequently, the error function E is determined and compared with the minimum error to assess whether the error function has fallen below it. The learning process concludes once the error meets the specified condition or the number of training iterations reaches the maximum number of iterations. If the criteria are not met, the learning continues.

3.2. The Backpropagation Neural Network Based on Particle Swarm Optimization Algorithm

The propensity of the BPNN to become entangled in local optima, coupled with the potential for larger errors, leads to less ideal prediction performance. Utilizing the PSO algorithm to optimize the BPNN, this method treats the weights generated during the neural network training process as particles for training. The optimal weights are identified when the maximum number of iterations is reached or the error reduces to a certain level. Apart from the strengths of the BPNN, the optimized network model overcomes the propensity of the BPNN to become entangled in local optima and the generation of larger errors. Consequently, this enhancement leads to an improvement in the prediction performance of the BPNN.

3.2.1. The Principle of the Particle Swarm Optimization Algorithm

Particle Swarm Optimization is a swarm intelligence optimization algorithm. In this algorithm, each particle represents a potential solution to the problem. The fitness function is used to calculate the adaptability of each particle. Based on the calculation results, the velocity and position of the particles can be updated. The velocity of the particles dictates their movement direction and distance, thereby achieving the optimization of individuals in the solution space.
Suppose a D -dimensional space contains a swarm of n particles denoted as X = ( X 1 , X 2 , , X n ) . Let X i = [ x i 1 , x i 2 , , x i D ] T represent the position of the i -th particle, V i = [ v i 1 , v i 2 , , v i D ] T represent its velocity, P i = [ P i 1 , P i 2 , , P i D ] T represent the individual best values achieved by the i -th particle, and P g = [ P g 1 , P g 2 , , P g D ] T represent the global best values obtained by the entire swarm. The steps of the PSO algorithm are listed in Algorithm 1.
Algorithm 1 Steps of the PSO algorithm
Step 1: Initialize the position and velocity vectors of the particle swarm.
Step 2: Calculate the fitness value corresponding to each particle’s position based on the objective function, and evaluate the quality of the solutions based on the magnitude of the fitness values.
Step 3: Identify the individual extreme values and the global extreme values of the particle swarm based on the fitness values.
Step 4: Update the velocity and position of the particle swarm using the following formulas:
V i d k + 1 = ω V i d k + c 1 r 1 ( P i d k X i d k ) + c 2 r 2 ( P g d k X i d k )
X i d k + 1 = X i d k + V i d k + 1
where ω represents the inertia weight, d = 1 , 2 , , D , i = 1 , 2 , , n ; k represents the current iteration number; V i d represents the velocity of the particle; c 1 and c 2 are non-negative constants, referred to as acceleration factors; r 1 and r 2 are random numbers distributed in the range [ 0 , 1 ] .
Step 5: Determine whether the termination criteria are met. If the criteria are satisfied, output the global optimum and conclude the algorithm; otherwise, revert to step 2 and continue with further iterations.

3.2.2. PSO-Based Optimization Algorithm for Backpropagation Neural Network

The PSO algorithm can optimize the weights of the BPNN, thereby enhancing the network’s nonlinear fitting capability. The method of the PSO-BP algorithm [21] is listed in Algorithm 2.
Algorithm 2 Steps of the PSO-BP algorithm
Step 1: Transform the initial weights of the BPNN into particles within the framework of the PSO algorithm, with random initialization of the particles’ velocities and positions.
Step 2: Establish the number of training set samples, test set samples, hidden layer nodes, particle swarm size, iteration count, inertia weight, and acceleration factors for the network.
Step 3: Calculate the fitness values for each particle.
Step 4: When the conditions for fitness value updating are satisfied, update the positions and velocities of each particle according to (11) and (12), and record the best positions of each particle.
Step 5: Record the global best position.
Step 6: Assess whether the termination criteria are satisfied; if so, output the global optima, obtain the optimal weights, and conclude the algorithm; otherwise, reinitialize the velocities and positions of the particles and reiterate the process.

3.3. Error Prediction and Compensation Methods for Integrated Navigation Systems

The PSO-BPNN proposed in this paper takes the position information, velocity information, attitude information output by the SINS and the navigation time as inputs. The position errors between the position information output by the SINS and the actual position information is used as the output.
When GPS is available, the PSO-BPNN is trained. The training algorithm iteratively adjusts the parameters of each layer to minimize the mean square error, thereby achieving the optimal neural network model. During GPS denial, still employing the position information, velocity information, attitude information output by the SINS and the navigation time as inputs to the already trained PSO-BPNN, one can predict the position errors of the SINS. Figure 3 shows the training and predicting process of the method proposed in this article.

4. Experimental Validation on an Actual Ship and Result Analysis

4.1. The Design of the Actual Ship Experiment

This work employed actual ship data to conduct a validation experiment. The experimental conditions were as follows:
The experimental ship was outfitted with a SINS/GPS integrated navigation system constructed from a self-developed fiber-optic SINS and GPS. Concurrently, a fiber-optic SINS known as PHINS, developed by the French company iXblue (Saint-Germain en Laye, France), served as the benchmark. When PHINS is in the integrated navigation mode, it is capable of providing high-precision attitude information (with an error not exceeding 0.01°). The performance metrics of the self-developed fiber-optic Strapdown Inertial Navigation System are characterized by a gyroscopic constant drift rate of approximately 0.01°/h, an accelerometer zero bias of about 50 μ g , and an inertial sampler frequency of 100 Hz.
The experiment utilized 14 h of actual ship data, with 10 h of data allocated for training and 4 h designated for prediction. The performance of two types of neural networks in assisting the prediction of the SINS positioning errors during GPS denial was compared: the BPNN and the PSO-BPNN. Each model employed the same depth and parameters to ensure a fair comparison. The metrics for comparison included the latitude error and the longitude error.

4.2. Results and Analysis of the Actual Ship Experiment

To quantitatively compare the performance of the two methods, this work selected the Absolute Mean (AM) and the Root Mean Square Error (RMSE) as evaluation metrics.
{ A M = ( | x 1 | + | x 2 | + + | x n | ) n R M S E = 1 n i = 1 n ( y i y ^ i ) 2
where | x i | represents the absolute value of the i -th data point, n is the number of data points in the dataset, y ^ i is the expected output of the model, and y i is the actual output of the model.
The loss function curves of the two methods are shown in Figure 4. The positioning error curves of the SINS are shown in Figure 5. They are the difference between the position information output by the SINS and the actual position data. The positioning error curves for both methods are shown in Figure 6. They are the difference between the predicting position errors and the actual position errors.
Through the comparation of the loss function curves, it is evident that the loss function curve of the PSO-BPNN descends more rapidly than that of the BPNN. This indicates that the BPNN model optimized by the PSO algorithm exhibits a faster learning rate and higher learning efficiency. Ultimately, it exhibits a smaller mean square error, which indicates that the PSO has successfully optimized the BPNN by identifying the optimal weights for the neural network model.
Analysis of Figure 5 indicates that during periods of GPS denial, the positioning errors of the SINS gradually diverge over time, and Table 1 shows that the AM of the SINS longitude error is 4070.6 m, and the latitude error is 11,950 m. The error is too excessive to locate precisely, which fails to satisfy the requirements for navigation accuracy.
According to the Figure 6 and Table 2, the AM of the longitude error for the BPNN is 7.9319 m, the RMSE is 9.7788 m, the maximum value is 45.03 m, and the minimum value is −42.06 m; for the latitude error, the AM is 11.4662 m, the RMSE is 14.6223 m, the maximum value is 70.84 m, and the minimum value is −77.69 m. It is evident that the BPNN is capable of suppressing the divergence of errors over time, suggesting that it can serve as a substitute for a GPS in assisting the integrated navigation system during GPS signal denial.
The AM of the longitude error for the PSO-BPNN is 3.3339 m, the RMSE is 4.3494 m, the maximum value is 23.29 m, and the minimum value is −25.35 m; for the latitude error, the AM is 3.4728 m, the RMSE is 4.3309 m, the maximum value is 16.85 m, and the minimum value is −16.16 m. Compared to the BPNN, the PSO-BPNN achieves a reduction of 57.97% in the AM of longitude error, 55.52% in the RMSE of longitude error, 69.71% in the AM of latitude error, and 70.38% in the RMSE of latitude error, and the fluctuation range of the data has also been reduced. These demonstrate that the predictive performance of the PSO-BPNN is superior to that of the BPNN. It is also capable of suppressing the divergence of errors over time and can confine the errors to a small range of fluctuation.

5. Conclusions

This paper proposes a method for predicting the positioning errors of SINS based on a Particle Swarm Optimization–Back Propagation Neural Network model. When the GPS signal is available, the output information of the SINS, navigation time, and the position errors between the position information output by the SINS and the actual position information are collected and trained by the PSO-BPNN module. During GPS denial, the well-trained model will provide the position errors. Through comparative experiments, the model proposed in this paper showed a higher accuracy on predicting the position over the traditional BPNN model. It provides an effective way to solve the application challenges of integrated navigation in specialized environments. In the future study, the state of the navigation system, such as the motion status will be considered to affect the predictive performance of the model. More experiments will be conducted based on this foundation.

Author Contributions

Conceptualization, Y.W. and Y.B.; software, T.W. and Z.G.; validation, T.W. and Z.G.; writing—original draft preparation, R.J. All authors have read and agreed to the published version of the manuscript.

Funding

This article was supported by the National Natural Science Foundation of China under Grant 51979047, the Natural Science Foundation of Hubei Province of China No. 2022CFB865, and the Natural Science Foundation of Heilongjiang Province of China under Grant YQ2021E011.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, Q.; Ben, Y.; Yu, F.; Tan, J. Transversal strapdown INS based on reference ellipsoid for vehicle in the polar region. IEEE Trans. Veh. Technol. 2016, 65, 7791–7795. [Google Scholar] [CrossRef]
  2. Jiang, C.; Chen, S.; Chen, Y.; Zhang, B.; Feng, Z.; Zhou, H.; Bo, Y. A MEMS IMU de-noising method using long short term memory recurrent neural networks (LSTM-RNN). Sensors 2018, 18, 3470. [Google Scholar] [CrossRef]
  3. Badu, R.; Wang, J. Comparative study of interpolation techniques for ultra-tight integration of GPS/INS/PL sensors. J. Glob. Position. Syst. 2005, 4, 192–200. [Google Scholar]
  4. Chen, J.; Yang, Z.; Ye, S.; Lu, Q. Design of Strapdown Inertial Navigation System Based on MEMS. Acad. J. Comput. Inf. Sci. 2020, 3, 110–121. [Google Scholar]
  5. Shaukat, N.; Ali, A.; Javed Iqbal, M.; Moinuddin, M.; Otero, P. Multi-sensor fusion for underwater vehicle localization by augmentation of RBF neural network and error-state Kalman filter. Sensors 2021, 21, 1149. [Google Scholar] [CrossRef]
  6. Jaradat, M.A.; Abdel-Hafez, M.F. Non-linear autoregressive delay-dependent INS/GPS navigation system using neural networks. IEEE Sens. 2017, 17, 1105–1115. [Google Scholar] [CrossRef]
  7. Yang, Y.; Gao, S.; Zhong, Y.; Gu, C. Model Predictive Forward Neural Network Algorithm and Its Application in Integrated Navigation. J. Inert. Technol. China 2014, 4, 222–226. [Google Scholar]
  8. Cao, J.; Fang, J.; Wei, S.; Bai, H. Research on the Error Feedback Correction Method of MEMS-SINS Based on Neural Network Prediction During GPS Signal Loss. J. Astronaut. 2009, 30, 2231–2236+2264. [Google Scholar]
  9. Zhang, Y.; Wang, L. A hybrid intelligent algorithm DGP-MLP for GNSS/INS integration during GNSS outages. Navigation 2019, 72, 375–388. [Google Scholar] [CrossRef]
  10. Wang, Q.; Li, H. Research on Indoor Positioning Technology Based on Machine Learning and Inertial Navigation. Electron. Meas. Technol. 2016, 6. [Google Scholar]
  11. Zhang, G.P. Time series forecasting using a hybrid ARIMA and neural network model. Neurocomputing 2003, 50, 159–175. [Google Scholar] [CrossRef]
  12. Wang, X.; Zhao, G.; Gao, Q. Application of BP Neural Network in the Transfer Alignment of Inertial Navigation System. J. Nav. Aeronaut. Eng. Coll. 2008, 5, 489–492. [Google Scholar]
  13. Wang, G.; Xu, X.; Yao, Y.; Tong, J. A Novel BPNN-Based Method to Overcome the GPS Outages for INS/GPS System. IEEE Access 2019, 7, 82134–82143. [Google Scholar] [CrossRef]
  14. Yan, F.; Zhao, W.; Wang, X.; Wan, K.; Wang, Y. Research on master-slave filtering of Celestial Navigation System/Inertial Navigation System. J. Phys. Conf. Ser. 2021, 1732, 012189. [Google Scholar] [CrossRef]
  15. Ma, S.; Liu, S.; Meng, X. Optimized BP Neural Network Algorithm for Predicting Ship Trajectory; IEEE: New York, NY, USA, 2020. [Google Scholar] [CrossRef]
  16. Fan, Y.; Wang, Z.; Lin, Y.; Tan, H. Enhance the Performance of Navigation: A Two-Stage Machine Learning Approach. In Proceedings of the 2020 6th International Conference on Big Data Computing and Communications (BIGCOM), Deqing, China, 24–25 July 2020; pp. 212–219. [Google Scholar]
  17. Liu, N.; Zhao, H.; Su, Z.; Qiao, L.; Dong, Y. Integrated navigation on vehicle based on low-cost SINS/GNSS using Deep Learning. Wirel. Pers. Commun. 2021, 126, 1–22. [Google Scholar] [CrossRef]
  18. Yu, X. Research on Gyroscope Random Error Modeling and Filtering Technology Based on LSTM Neural Network. Ph.D. Thesis, Huazhong University of Science and Technology, Wuhan, China, 2022. [Google Scholar]
  19. Milad, M.; Bryan, M.; Ilya, A. Unsupervised anomaly detection in flight data using convolutional variational auto-encoder. Aerospace 2020, 7, 115. [Google Scholar] [CrossRef]
  20. Chen, Z. Principles of Strapdown Inertial Navigation Systems. Master’s Thesis, Beijing Aerospace Press, Beijing, China, 1986. [Google Scholar]
  21. Wang, X.; Shi, F. MATLAB Neural Network: Analysis of 43 Cases. Master’s Thesis, Beihang University Press, Beijing, China, 2013. [Google Scholar]
Figure 1. GPS/SINS pine integrated navigation system model.
Figure 1. GPS/SINS pine integrated navigation system model.
Sensors 24 03722 g001
Figure 2. The three-layer backpropagation neural network structure.
Figure 2. The three-layer backpropagation neural network structure.
Sensors 24 03722 g002
Figure 3. The training and prediction process of the PSO-BP neural network.
Figure 3. The training and prediction process of the PSO-BP neural network.
Sensors 24 03722 g003
Figure 4. The loss function curves of the two methods.
Figure 4. The loss function curves of the two methods.
Sensors 24 03722 g004
Figure 5. The position error curves of the SINS.
Figure 5. The position error curves of the SINS.
Sensors 24 03722 g005
Figure 6. The position error curves of the two methods.
Figure 6. The position error curves of the two methods.
Sensors 24 03722 g006
Table 1. The statistics of the SINS positioning errors.
Table 1. The statistics of the SINS positioning errors.
ErrorAMRMSE
Longitude error (m)4070.630.459
Latitude error (m)11,95021.540
Table 2. Analysis of the experimental results.
Table 2. Analysis of the experimental results.
MethodBPNNPSO-BPNN
Error IndexAMRMSEMAXMINAMRMSEMAXMIN
Longitude error (m)7.93199.778845.03−42.063.33394.349423.29−25.35
Latitude error (m)11.466214.622370.84−77.693.47284.330916.85−16.16
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Jiao, R.; Wei, T.; Guo, Z.; Ben, Y. A Method for Predicting Inertial Navigation System Positioning Errors Using a Back Propagation Neural Network Based on a Particle Swarm Optimization Algorithm. Sensors 2024, 24, 3722. https://doi.org/10.3390/s24123722

AMA Style

Wang Y, Jiao R, Wei T, Guo Z, Ben Y. A Method for Predicting Inertial Navigation System Positioning Errors Using a Back Propagation Neural Network Based on a Particle Swarm Optimization Algorithm. Sensors. 2024; 24(12):3722. https://doi.org/10.3390/s24123722

Chicago/Turabian Style

Wang, Yabo, Ruihan Jiao, Tingxiao Wei, Zhaoxing Guo, and Yueyang Ben. 2024. "A Method for Predicting Inertial Navigation System Positioning Errors Using a Back Propagation Neural Network Based on a Particle Swarm Optimization Algorithm" Sensors 24, no. 12: 3722. https://doi.org/10.3390/s24123722

APA Style

Wang, Y., Jiao, R., Wei, T., Guo, Z., & Ben, Y. (2024). A Method for Predicting Inertial Navigation System Positioning Errors Using a Back Propagation Neural Network Based on a Particle Swarm Optimization Algorithm. Sensors, 24(12), 3722. https://doi.org/10.3390/s24123722

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop