1. Introduction
The SINS is an autonomous navigation system, with the advantages of strong anti-interference ability, good real-time performance and no geographical restrictions, which has become the main navigation system for all kinds of carriers [
1,
2]. However, due to the accumulation of position, speed and attitude errors over time, the SINS is difficult to use as an independent high-precision navigation system [
3,
4]. The CNS has the advantages of the non-accumulation of errors over time, strong autonomy, anti-interference ability and high reliability, but it is vulnerable to being affected by sky background noise and has disadvantages such as a low data rate and weak real-time performance and continuity [
5,
6,
7]. Modern information-based navigation warfare has seriously threatened all kinds of radio navigation systems represented by satellite navigation systems [
8]. The complex battlefield electromagnetic environment requires that navigation system must have strong autonomy and anti-interference ability. The SINS and CNS can not only meet the aforementioned requirements, but also have strong complementarities with each other [
9,
10]. Therefore, in recent years, SINS/CNS integrated navigation has been widely considered and rapidly developed. With great development potential and broad application prospects, the SINS/CNS has become an important direction of integrated navigation technology development [
9].
A ship CNS works on the earth’s surface; due to harsh astronomical observation conditions, it will be interfered with by strong sky background noise. Even the high-resolution CNS with a small field of view has to face the problem of a sharp decline of observation accuracy and even observation interruption under the impact of atmospheric turbulence, clouds and other meteorological factors [
11,
12]. It is an urgent problem for SINS/CNS integrated navigation to keep the CNS in a stable working state and provide reliable navigation reference information for the SINS.
In recent years, recurrent neural networks (RNNs) have been widely used in dealing with time series problems, and they provide the inspiration for the investigation in this article. RNNs evolved from the Hopfield neural network proposed by Saratha Sathasivam in 1982, which embodies the idea that “human cognition is based on past experience and memory” [
13,
14]. The difference with ordinary Deep Neural Networks [
15,
16] and Convolutional Neural Networks [
17,
18] is that RNNs not only take into account the influence of the output of the previous moment on the current moment, but also have the memory of past content. Just like the human brain, they do not process and judge current information independently but make optimal choices based on previous knowledge and experience [
19]. Therefore, RNNs have significant applicability in processing sequence data and solving serialization-related problems. At present, RNNs have been successfully applied in many artificial intelligence fields, such as Google voice search, Baidu deep speech recognition and Amazon-related products. However, as the length of the time series becomes longer, the memory ability of the RNN will rapidly deteriorate, resulting in “gradient vanishing” or “gradient explosion” [
20,
21]. As an evolutionary version of RNNs, long short-term memory (LSTM) adopts a gating mechanism and memory cells, which can flexibly control the retention degree of the historical information of time series and effectively solve the problem of “gradient vanishing” or “gradient explosion” [
22,
23,
24]. At present, LSTM neural networks are being widely used in the field of navigation. Brandon Wagstaff et al. presented a method to improve the accuracy of a zero velocity-aided inertial navigation system (INS) by replacing the standard zero velocity detector with an LSTM neural network [
25]. Edoardo Topini et al. developed an LSTM-based dead reckoning approach to estimate surge and sway body-fixed frame velocities, without canonically employing the DVL sensor [
26]. Lv et al. proposed a position correction model based on hybrid gated RNNs that does not need to establish a motion model like typical navigation algorithms, to avoid modeling errors in the navigation process [
27]. Guang et al. proposed a method using LSTM to estimate position information based on inertial measurement unit (IMU) data and Global Positioning System (GPS) position information [
28].
Ships are affected by external factors such as wind, waves, currents and surges, making their attitude change constantly. However, from the perspective of the time dimension, this change can be described by a cosine curve with periodicity. Therefore, the attitude change in a ship is a typical time series problem, and there is a strong long-term correlation between the sequence information. It is highly suitable for the use of LSTM to forecast the ship’s attitude. With the aforementioned considerations, this article is devoted to investigating a method of ship SINS/CNS integrated navigation aided by the forecasted attitude of LSTM. The contributions of this article are as follows:
- (1)
The proposed method expands the application boundary of traditional ship SINS/CNS integrated navigation and effectively improves the robustness of the integrated system;
- (2)
The SINS/CNS integrated model is derived based on an attitude solution of the CNS, which provides more favorable feature data for LSTM learning;
- (3)
The LSTM neural network is developed to accomplish a high-precision ship attitude forecast.
This article is organized as follows. The derivation of the SINS/CNS integrated model based on the attitude solution is presented in
Section 2.
Section 3 describes the modeling method of the LSTM neural network in detail. In
Section 4, an experiment is conducted to verify the performance of the proposed method and the results are presented with a detailed discussion.
Section 5 summarizes the whole article.
2. SINS/CNS Integrated Navigation Based on Attitude Solution
The technical proposal of SINS/CNS integrated navigation based on a CNS attitude solution can be described as follows. First, the asynchronous multi-star vectors are transformed into a same star sensor frame. Then, the ship attitude relative to the inertial frame is calculated based on the transformed and inertial space multi-star vectors. Finally, the measurement model is derived based on the inertial attitude solved by the CNS for information fusion with the SINS.
The inertia-based integrated navigation equation with a Kalman filter is well known in the navigation field; it is not repeated in this article as the focus of this chapter is the CNS solution model.
2.1. Transformation for Asynchronous Multi-Star Vectors
With high optical resolution and strong background noise suppression ability, the small field of view CNS is widely used in carriers near the surface of the earth, such as ships and submarines, in poor celestial observation conditions. The small field of view CNS can track only one star at one time, which means that the multi-star vectors needed for every celestial navigation solution are not observed simultaneously in the same star sensor frame. Therefore, the asynchronous multi-star vectors must be corrected for a CNS solution.
It is assumed that the CNS needs to observe m (m ≥ 3) stars to complete a navigation solution once. Let
denote the multi-star vectors in the star sensor frame s, and
denote the corresponding multi-star vectors in inertial frame i: the transformation of the two can be expressed as
where
b denotes the body frame,
is the transformed matrix from
i to
b, which is also named the inertial attitude matrix and
is the transformed matrix from
b to
s, which is also named the installation error matrix.
In the process of practical celestial navigation,
is calibrated in the factory and the calibration residual can be estimated by Kalman filtering (KF),
represents the observation vectors changing with time and acquired by the star sensor and
represents the constant reference vectors obtained with celestial apparent position calculation [
29]. Let the superscript “~” denote the corresponding variable containing the measurement error, so (1) can be converted to
where
is provided by the SINS,
denotes the observation time of any star and
denotes the observation time of the last star. Frame b has linear or angular motion relative to the inertial space, so it needs to be labeled with time. Since
i is a fixed frame, there is no need to label it with time;
is usually treated as a constant matrix, so there is no need to time it either.
represents the multi-star vectors in s corresponding to
, while
represents the multi-star vectors in s corresponding to
. The transformation of the two can be derived from (2) and (3) as
where the right superscript “
T” is a matrix transpose operator. According to (4), all the asynchronous multi-star vectors can be unified into the star sensor frame corresponding to
.
2.2. Inertial Attitude Solved by CNS
After obtaining the multi-star vectors
, (3) can be converted to
where
is the inertial attitude matrix of frame
s with respect to frame
i at
. This can be obtained by the algorithm of the multi-vector attitude solution [
29] as
where
is the optimal estimation function for the multi-vector attitude solution and
is the weighting coefficient of each star vector. By correcting the installation error matrix
, the ship’s inertial attitude matrix
can be described as
Remark 1. Compared with the yaw or position solutions of the traditional ship CNS, the proposed CNS model based on an attitude solution will provide more comprehensive navigation information, which is also more suitable for LSTM learning.
2.3. Measurement Model for SINS/CNS
The measurement model base on
for SINS/CNS can be described as [
10]
where
is the transformed function from the matrix to the rotation vector,
e denotes the earth frame and n denotes the navigation frame (geographical frame is used as the navigation frame in this article).
, named the time matrix, is the transformed matrix from
i to
e, which is calculated based on the high-precision observed time and can be regarded as an error-free quantity.
, named the position matrix, is the transformed matrix from
e to
n.
, named the attitude matrix, is the transformed matrix from n to b. Both
and
are provided by the SINS.
is the misalignment angle and
is the position error vector presented in the form of latitude error, longitude error and height error. The matrix
can be expressed as
where
L is the latitude.
3. LSTM Attitude Forecast Model
The diagram of the technical framework for the LSTM attitude forecast model is summarized in
Figure 1. The input set is constructed by the attitude
output from the SINS independent solution module. The target set is constructed by the attitude
output from the SINS/CNS integrated module. The input and target sets are processed by the LSTM encoder to be converted to a data type suitable for the LSTM model training. Then, the LSTM begins learning to determine the hyperparameters and model parameters through iterative training. If the loss continues to decrease, the training continues; otherwise, the forecast starts. The forecasted result is converted to attitude
by the LSTM decoder. With the forecasted attitude as a reference, the attitude
is detected from the CNS module. If the attitude
is invalid, a fault warning is issued and the CNS attitude
will be replaced with the LSTM forecasted attitude
for information fusion with the SINS; otherwise, the training continues.
Remark 2. In the practice of LSTM model training, the input attitude, the target attitude and the forecasted attitude are all constructed in the form of three-dimensional Euler angles, but in order to facilitate understanding and expression, the attitude is described in the form of a matrix in this article.
3.1. Selection of Input Set and Target Set
The target set refers to the dataset composed of the parameters used as the forecasted target, and the input set refers to the dataset composed of the parameters that can accurately and fully reflect the features of the target. Obviously, the selection of input is determined by the target; whether the input matches the target is the key factor affecting the forecasted performance of the LSTM network.
According to the theory of optimal estimation, the precision of navigation parameters after information fusion with integrated navigation is better than that of each subsystem. Therefore, the attitude output from the SINS/CNS is selected to construct the target set. The input should not only reflect the attributes of the target attitude, but also have the anti-interference ability without the influence of CNS rejection. Consequently, the SINS independent solution module is added into the SINS/CNS integrated navigation system, and the attitude output from this module is selected to construct the input set.
3.2. LSTM Encoder Based on Seq2Seq
The function of the LSTM encoder is to reassemble, segment and normalize the datasets and make them suitable for network training [
21]. According to the different constructures of the input and target, the LSTM time series task can be classified as one-to-one, one-to-many, many-to-one and many-to-many [
22]. In order to give full play to the advantage of LSTM in processing the long-term sequence, the data structure of Seq2Seq (many-to-many) is used to construct the LSTM encoder.
The method of the second-order overlapping shift is proposed to encode the dataset in this article. The first order means that the target sequence is aligned with the input sequence in the form of an overlapping shift, where the shift step is just the forecasted step (delay). The other order means that the input sequence and the target sequence are divided into a series of subsequences in the form of overlapping shifts.
As shown in
Figure 2,
L is the length of the input and target sequences,
s is the length of the input and target subsequences and
d is the forecasted step. With
s as the sliding window and 1 as the sliding step, the original input and target sequences are divided into
L-
s + 1 subsequences, respectively. On this basis, a sample
set is constructed with
L-
s + 1 samples, as shown in
Figure 3. The sample set
is divided into a training set, a validation set and a test set according to a certain ratio (usually 6:2:2, 7:2:1 or 8:1:1).
3.3. Bayesian Hyperparameter Optimization
The performance of a deep learning model is mainly determined by two types of parameters, model parameters and model hyperparameters. Model parameters are a set of internal parameters used to determine a deep learning model, such as the weight coefficient matrix and the bias vector in the various gate structures of the LSTM neural network. Hyperparameters are a set of external parameters used to design the network structure and training strategy to optimize the training effect of the model. Model parameters can be obtained independently with deep learning for sample without manual intervention. However, the learning of the model parameters is directly affected by the hyperparameters, so the focus is on the problem of hyperparameter optimization as follows.
According to the application characteristics of the ship attitude forecast, the loss function based on the algorithm of mean squared error (MSE) is used as the objective function for hyperparameter optimization.
where
represents a set of hyperparameter sampling values,
represents the target attitude in the validation set,
represents the forecasted attitude and
is the number of samples. Hyperparameter optimization aims to find a set of ideal hyperparameters to minimize the objective function. This process can also be simply considered as finding a set of hyperparameters
to minimize the attitude forecast error of the LSTM model, as shown in (11).
where
is the value range of the hyperparameters to be optimized.
At present, Bayesian optimization has been applied more and more widely in solving the black box function and has become the mainstream method for hyperparameter optimization [
30]. Bayesian hyperparameter optimization mainly consists of two technical modules, a learning surrogate model and constructing an acquisition function [
31].
3.3.1. Learning Surrogate Model
The objective function belongs to the black box function. It is difficult to obtain the function’s specific expression, and the calculation cost is expensive. Therefore, Bayesian optimization must first choose a surrogate model for the objective function. The Gaussian process regression (GPR) model is a widely used and efficient surrogate model. On the basis of assuming that the objective function conforms to the Gaussian distribution, the mathematical expectation
and covariance matrix
are calculated by sampling the input and output
k times, and then a Gaussian process regression model
is trained to replace the objective function
.
satisfies the following Gaussian process (GP):
where
3.3.2. Constructing Acquisition Function
In order toget the approximate global optimal value of the objective function as soon as possible, it is necessary to construct the acquisition function
to select the next sampling point
(that is, another set of hyperparameters) optimally, where
is expressed as
where
represents the
k sampling points that are used to initialize the surrogate model and
represents the loss values corresponding to
.
represents another n sampling points after
and
represents the loss values corresponding to
. The surrogate model will continue to learn and update with the subsequent sampling, so the acquisition function
can also be considered as the distribution of sampling points based on the given surrogate model, and satisfies
where
and
are the calculated functions for the mathematical expectation
and covariance matrix
, respectively [
32]. Because the explicit expressions of the two functions are relatively complicated, and they can be directly calculated by the Bayesian optimization library, this article does not expand the description.
The expected improvement (EI) model is used to construct the acquisition function. To define the improvement of sampling points
where
is the approximate optimal solution of the objective function after the sampling points
.
represents the value space of the next sampling point near
. If
is smaller than
,return the descent degree; otherwise, return 0. Following this law, the acquisition function is constructed as
where
represents the mathematical expectation of
. The probability distribution of
can be determined by (17)–(19), so
is the next sampling point that maximizes the EI of the acquisition function, that is
3.3.3. Bayesian Hyperparameter Optimization Algorithm
To sum up, Bayesian hyperparameter optimization can be summarized as Algorithm 1.
Algorithm 1: Bayesian hyperparameter optimization |
| , selecting surrogate model (GPR) and acquisition function |
| } |
1 | Begin |
2 | |
3 | |
4 | do |
5 | |
6 | |
7 | |
8 | End |
9 | } |
10 | End |
3.4. Training for LSTM Model
The training for the LSTM model can be summarized as Algorithm 2.
Algorithm 2: Training for LSTM model |
| |
| Output: LSTM model |
1 | Begin |
2 | Step 1: Initializing the default hyperparameters, setting the value space for the hyperparameters to be optimized |
3 | Step 2: Instantiating LSTM model |
4 | do |
5 | |
6 | |
7 | epochs of training |
8 | ) |
9 | End |
10 | End |
11 | |
12 | End |
In fact, the core target of the LSTM model training can be understood as the process of determining the model hyperparameters and model parameters based on the training set and validation set with the loss function as the evaluation index. As mentioned above, Algorithm 2 “training for LSTM model” is closely related to Algorithm 1 “Bayesian hyperparameter optimization”. In practical applications, Algorithm 2 is often nested in Algorithm 1. One of output of Algorithm 2, , is just the in Algorithm 1.
5. Conclusions
This article presented a complete set of technical solutions for ship SINS/CNS integrated navigation aided by LSTM-RNN. The LSTM model accomplished a high-precision attitude forecast for ships. In the case of CNS signal abnormality or even failure, the LSTM forecasted attitude is used as reference information fused with the SINS, which has ensured the continuous and stable output of high-precision navigation information, expanded the application boundary of ship SINS/CNS integrated navigation and effectively improved the robustness of the integrated system.
Compared with the traditional SINS/CNS integrated navigation method, although the technical scheme proposed in this paper increases the computational load of the system through the optimized design of the LSTM network structure and training strategy, the mapping relationship between the input and output of the forecasted model is much simpler. With the support of the integrated computing environment, the increased computational burden is completely acceptable. In addition, in our future research, we will study more accurate and efficient SINS/CNS integrated navigation models that are more suitable for deep learning algorithms.