Next Article in Journal
Therapeutic Effect Enhancement by Dual-Bias High-Voltage Circuit of Transmit Amplifier for Immersion Ultrasound Transducer Applications
Next Article in Special Issue
Automatic Building Extraction from Google Earth Images under Complex Backgrounds Based on Deep Instance Segmentation Network
Previous Article in Journal
A Polarization Independent Quasi-TEM Metamaterial Absorber for X and Ku Band Sensing Applications
Previous Article in Special Issue
A Deep Learning Approach on Building Detection from Unmanned Aerial Vehicle-Based Images in Riverbank Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Online Prediction of Ship Behavior with Automatic Identification System Sensor Data Using Bidirectional Long Short-Term Memory Recurrent Neural Network

1
Navigation College, Dalian Maritime University, Dalian 116026, China
2
Collaborative Innovation Research Institute of Autonomous Ship, Dalian Maritime University, Dalian 116026, China
3
Key Laboratory of Navigation Safety Guarantee of Liaoning Province, Dalian 116026, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(12), 4211; https://doi.org/10.3390/s18124211
Submission received: 10 October 2018 / Revised: 12 November 2018 / Accepted: 27 November 2018 / Published: 30 November 2018
(This article belongs to the Special Issue Deep Learning Remote Sensing Data)

Abstract

:
The real-time prediction of ship behavior plays an important role in navigation and intelligent collision avoidance systems. This study developed an online real-time ship behavior prediction model by constructing a bidirectional long short-term memory recurrent neural network (BI-LSTM-RNN) that is suitable for automatic identification system (AIS) date and time sequential characteristics, and for online parameter adjustment. The bidirectional structure enhanced the relevance between historical and future data, thus improving the prediction accuracy. Through the “forget gate” of the long short-term memory (LSTM) unit, the common behavioral patterns were remembered and unique behaviors were forgotten, improving the universality of the model. The BI-LSTM-RNN was trained using 2015 AIS data from Tianjin Port waters. The results indicate that the BI-LSTM-RNN effectively predicted the navigational behaviors of ships. This study contributes significantly to the increased efficiency and safety of sea operations. The proposed method could potentially be applied as the predictive foundation for various intelligent systems, including intelligent collision avoidance, vessel route planning, operational efficiency estimation, and anomaly detection systems.

1. Introduction

For a ship to avoid collision, it must predict the behaviors of other ships in order to estimate the collision risk. High-precision real-time prediction of ships’ navigational behaviors can effectively increase the reliability of collision avoidance decisions and decrease the risk of collisions.
Navigation behavior prediction algorithms can be divided into two categories: offline and online prediction. Numerous offline prediction algorithms are employed in the fields of trajectory estimation and data restoration. In offline prediction, trajectory data are input to a fixed formula or trained model. These algorithms are insufficiently flexible for adaptation to actual predicted data; they also lack timeliness and high efficiency.
Examples of offline prediction studies include work by Han et al. [1] on the prediction of a ship’s trajectory by establishing a state-switching model; they proposed a conflict-free four-dimensional aircraft prediction method based on hybrid system theory. For predicting ship positions, Xu et al. [2] trained a three-layer back-propagation (BP) neural network to accept the ship’s direction and speed as inputs and yield differences in the latitude and longitude as output. Xu et al. [3] used a Kalman filter to modify automatic identification system (AIS) data and least-squares estimation to determine the system state. Liu et al. [4] first used a discrete wavelet transform to preprocess the ship’s trajectory and then applied grey systems theory to predict the ship’s navigation position. Zhen et al. [5] constructed a three-layer BP neural network in which AIS data from the previous three points were used as the input and the fourth point was given as the output to predict the navigation behavior of the ship. Zhao et al. [6] used an improved Kalman filter algorithm to predict ship trajectories. Tong et al. [7] examined the law of curved navigation and the reliability of ship trajectories and proposed a geographic information system (GIS)-based prediction method for the AIS data of island navigation. Kim and Hong [8] used a nonlinear Kalman filter in a curved-motion interactive multimodel (IMM) algorithm to track vehicles. Mehrotra and Mahapatra [9] applied a four-state Kalman filter and “jerk” model to track predictions via motion models in a three-dimensional (3D) space. Vaidehi et al. [10] used an ordinary Kalman filter with neural network elements and an auxiliary Kalman filtering scheme to track highly maneuverable multi-targets. Gan et al. [11] grouped historical trajectories by the k-means algorithm and then used artificial neural network (ANN) models to predict ship trajectories. Cai and Zhang [12] constructed an offline model using a hybrid particle swarm optimization evolutionary algorithm (PSO-EA) for time series prediction.
Previous online prediction algorithms could not learn or retain the experience of prior data repairs. Examples of online prediction studies include a ship trajectory prediction model based on Gaussian process regression, proposed by Mao et al. [13], in which a 24 min ship trajectory was simulated and predicted by extrapolating from the ship’s existing data. Perera et al. [14] analyzed and tracked multiple ship conditions, combined ship trajectory detection with shipping state estimation, and simulated ship trajectories. Berker et al. [15] applied a two-dimensional (2D) obstacle motion-tracking module to a dynamometer tracking algorithm to improve data quality for navigation purposes. Stubberud and Kramer [16] used a neural-extension Kalman filter to dynamically predict a target state online, thus improving the state estimation capability of existing models. Sang et al. [17] built a prediction model by using change of speed (COS), rate of turn (ROT), speed over ground (SOG), and course over ground (COG) to develop the closest point of approaching (CPA) searching method. Nagai and Watanabe [18] proposed a ship-position prediction model for a path following the ship’s performance on a curved path. Borkowski [19] presented an algorithm of ship-movement trajectory prediction, which considered measurements of the ship’s current position from a number of doubled autonomous devices through data fusion. Zissis et al. [20] employed machine learning and specifically ANNs as tools to add predictive capability to vessel traffic-monitoring information systems (VTMISs). Yin et al. [21] proposed an online sequential extreme learning machine (OS-ELM)-based gray prediction approach for online ship roll prediction. Zhang et al. [22] predicted ship maneuvering in regular waves. Based on a two-time scale model, the total ship motion can be divided into low-frequency maneuvering motion and high-frequency wave-induced motion. Breda [23], in the simulator, created critical vessel traffic scenarios and predicted maneuvering margins of the vessel, indicating the predicted boundaries of safe operation.
A recurrent neural network (RNN) is a deep learning technology designed to deal with time series data [24]. RNNs can achieve high precision for machine learning tasks with time series attribute data. For example, in 2015, Google substantially improved speech recognition in Android phones and other devices by using RNNs. Apple’s iPhone also uses an RNN framework in Siri. Microsoft uses RNNs for speech recognition, virtual-dialogue image generation, and programming code. RNNs have been applied in a wide range of research areas, including translation, document abstraction, speech recognition, image recognition, disease prediction, click-through rate prediction, stock forecasting, synthetic music, and e-commerce fraud detection. In an RNN, feedback is provided from the output variable to the input [25,26]. The feedback variable contains a time delay network. However, training an RNN is complex and convergence cannot be guaranteed. AIS data regarding a ship’s trajectory are given as a time series. In an RNN, increasing the data length may induce a gradient error, or a gradient explosive may occur when error parameters are back-propagated. The gradient explosion phenomenon does not meet the training objectives, and a typical RNN does not provide satisfactory results. Therefore, in a bidirectional RNN structure [27], based on the traditional RNN concept, forward and backward propagation are used to effectively update the weights, thereby increasing the contextual relevance and prediction precision. Instead of a hidden-layer unit, the bidirectional RNN uses a long short-term memory (LSTM) unit. The LSTM unit effectively filters key features, performs selective memory, and solves the problem of RNN processing for long-time data. The LSTM algorithm was proposed by Sak [28] to establish long-term correlations between input values. By replacing the hidden-layer network in an RNN with an LSTM unit [29,30], the problem of gradient dispersion is solved and a new storage unit is created to selectively forget or remember a given operation. In this study, we developed a bidirectional LSTM RNN (BI-LSTM-RNN) combining historical experience and real-time adjustment flexibility that can be utilized to predict the navigation behavior of an unmanned ship during collision avoidance measures against a manned ship. This approach can support a human operator’s better understanding of complex conditions at sea and enhance decision-making to avoid danger.

2. Navigation Behavior Learning Model

2.1. RNN Structure

RNNs use directional loops to address problems in the context of input nodes. The RNNs overcome the connection between a traditional neural network structure layer and the hidden layer. The transition between each layer node is no longer the input of a hidden layer. An RNN is a sequence-to-sequence model [25,26,31] that can suitably process sequence data of any length. In processing AIS data, the current state is related only to the previous ship states. The basic network structure of an RNN is depicted in Figure 1. In the figure, t represents the time, x is input data, O is output data, S is the network state, W is the update weight, V is the weight between the cell and the output, and U is the weight between the input and cell.
The idea of the RNN structure is to make full use of the information in the previous sequence, which is common in traditional neural networks. It is assumed that all inputs or outputs are independent of each other. However, many natural language processing tasks are judged by context. Therefore, this assumption has its limitations. An RNN is directly translated into a circulating neural network or RNN because different inputs pass through the same neural network, and the difference is concealed by the previous state of the hidden layer. As shown in Figure 1, the unrolled RNN is on the left side and the network structure of the expanded RNN is on the right side.
An RNN contains input units labeled { x 0 , x 1 , , x t , x t + 1 } , output units labeled { y 0 , y 1 , , y t , y t + 1 } , and hidden units labeled { s 0 , s 1 , , s t , s t + 1 } . The hidden units perform the most important work [32]. A one-way flow of information occurs from the input to the hidden units to the output units (Figure 2). In some cases, the RNN breaks the one-way nature of the second flow, and the boot information is returned to the hidden unit from the output unit. This process is called back-projection, and the input to the hidden layer includes the state of the previous hidden layer. The nodes in the layer can be connected to the next layer or to other units.

2.2. LSTM Cell Structure

An LSTM unit is a special RNN unit that can solve the long-term dependency problem. In an LSTM unit, the cell state controls the discarding and adding of information through the gate to achieve forgetting and memorizing functions [33,34,35]. An LSTM performs selective operations on the knowledge learned using the three gate structures of input, output, and forget gate. The LSTM structure (Figure 3) allows self-looping and real-time updating of weights to prevent gradient disappearance and gradient expansion.
The training process of the LSTM network is as follows:
(1)
State initialization: The number of neural nodes in the input layer, number of output nodes, and number of each cell unit (k) are determined. The initial state {S} of each cell unit is equal to 0, and the link weight ( ω i j ) of each layer is equal to 0. The average value (±1) is a randomly generated range. The offset θ j is initialized to 0.1. W represents the weight of the matrix.
(2)
The output data (H) are calculated from the input layer, according to ω i j and θ j from Equation (1):
H ˜ j = i = 1 n ω i j x i + θ j .
(3)
LSTM unit calculation: The output of the unit above the forget gate and the input of this unit are used as inputs for the sigmoid function (Figure 4), which adds to the degree of forgetting used to control the previous unit.
f t = sigmoid ( W f · [ h t 1 , x t ] + b f )
(4)
The input gate integrates the Ct−1 of the previous state with the Ct of the current state to update the cell unit state.
C t = f t · C t 1 + i t · C t ˜
C t = tanh ( W C · [ h t 1 , x t ] + b C )
(5)
The output value ( o t ) from the output gate is passed to the status value ( h t ) of the next unit to complete the training procedure.
o t = sigmoid ( W o · [ h t 1 , x t ] + b o )
h t = o t · tanh ( C t )
(6)
Error calculation: The prediction error ( e ) is calculated according to O and the expected output ( y ) of the RNN prediction error, returning the error number of each batch.
e k = y k 2 o k 2 k = 1 , 2 , 3 m
(7)
Weight updating: The random gradient descent method is used to optimize the error depending on the value. For each update parameter, it is unnecessary to traverse all of the training sets; only one value is used to update a parameter. Such an algorithm is more suitable for big data and has fast searching capability.
w k w k = w k η m j C X j w k
θ l θ l = θ l η m j C X j w k

2.3. Bidirectional Structure

If the following information can be accessed in advance, the analysis of the current sequence information becomes relatively easy [27,36,37]. As standard RNN processes have time series sequences, the following information is usually ignored. A delay is commonly added between the input and the target to predict future information. However, in practical applications, an excessively long delay yields an inferior prediction result. An inferior prediction result will cause the network to devote excessive resources to memorizing large amounts of input information, thus diminishing its ability to model the combined knowledge of different input vectors. Therefore, the size of the delay must be adapted manually. In a bidirectional RNN, the training sequence involves both forward and backward RNNs, which are connected to an output layer. Figure 5 illustrates the structure of a bidirectional RNN.
The structure of the bidirectional RNN provides complete historical and future information for each point in the input sequence to the output layer. Figure 5 displays the bidirectional RNN spread over time. Six unique weights are used repeatedly in each time step: (w1, w3) correspond to inputs to the forward and backward hidden layers, (w2, w5) correspond to hidden layers to the hidden layer itself, and (w4, w6) correspond to forward and backward hidden layers to the output layer.
No information flow occurs between the forward and backward hidden layers, ensuring that the expanded graph is acyclic for the weight relationship (Figure 6).

2.4. Batch Training Structure

In this study, the LSTM unit was added to the time series data to solve the gradient disappearance problem. Moreover, a bidirectional structure was implemented to enhance the context correlation, yielding the BI-LSTM-RNN. Batch training was performed with batches comprising 10 sets of data each. The AIS data were arranged in a matrix and randomly batched, as illustrated in Figure 7, to prevent overfitting, eliminate any self-correlation of the data, and improve the learning efficiency.

3. Navigation Behavior Prediction Model

To verify the validity of the model, 2015 AIS trajectory data from Tianjin Port for 11,032 ships were selected. The data included 36,807,928 coordinate points occupying 8.58 GB. The AIS data from January to October included 29,277,849 points and comprised the training group. The AIS data from November to December included 7,530,103 points and formed the verification group. A flowchart of the complete prediction model is displayed in Figure 8.

3.1. Parameter Analysis

For an unmanned ship meeting another ship, it must predict the behavior of the incoming ship and adopt an effective collision avoidance strategy. In our BI-LSTM-RNN, the current and historical AIS ship data are taken as input values, and the future ship position is taken as the output value of the network; the output can then be compared with actual ship position data. The AIS data are multidimensional and multiparametric to characterize ship behavior; for example, the data include the ship’s direction, position, and speed modified over time. In the test AIS data, each ship was subdivided according to its Maritime Mobile Service Identity (MMSI). The ships were sorted using timestamps. The storage structure of the AIS information is displayed in Table 1; the AIS data for June are illustrated in Figure 9.
To avoid overfitting, the position information is comprehensive and includes information regarding speed. Therefore, information related to the position, heading, and time is selected for learning. The input layer ship behavior data can be expressed as
I ( t ) = { l o n t 1 , l a t t 1 , t 1 , h e a d i n g t 1 ,     l o n t , l a t t , t , h e a d i n g t ,     l o n t + 1 , l a t t + 1 , t + 1 , h e a d i n g t + 1 ,     t + 2 }
Batch standardization is required before data are entered into each layer network:
x i ^ = x i μ B σ B 2 + ε
y i = γ x i ^ + β
where μ B is the batch mean, σ B 2 is the batch variance, and γ and β are the learned parameters.
The output layer ship position data O ( t + 2 ) can be expressed as
O ( t + 2 ) = { l o n t + 2 , l a t t + 2 , h e a d i n g t + 2 }
The error function is
l o s s = | h e a d i n g _ p r e h e a d i n g t + 2 | × ( l o n _ p r e l o n t + 2 ) 2 + ( l a t _ p r e l a t t + 2 ) 2 / s i n ( π | h e a d i n g _ p r e h e a d i n g t + 2 | / 2 )
Under the machine learning framework of Google TensorFlow, the BI-LSTM-RNN was implemented in the Python language. The learning network structure contained one input layer, two hidden layers, two LSTM unit layers, and one output layer. The overall structure of the network training, automatically generated through TensorBoard, is depicted in Figure 10.

3.2. AIS Trajectory Data Value Filter

When the compression algorithm compresses the trajectory data, the algorithm attempts to establish a criterion for judging the value of the ship’s trajectory [38] (Figure 11). It removes data with low trajectory values and retains data with extreme trajectory values. This operation achieves compression and must be retained. Data with high trajectory values are called ship trajectory feature points and key feature points. At feature points, the original trajectory is strong in the AIS ship trajectory data. If such a point is lost, the ability to restore the original trajectory considerably decreases. Points that are not key feature points can be simplified to achieve the compression effect. Eliminating some track data inevitably causes distortion; the threshold is determined to be between the compression and distortion rates.

4. Results

Four points from the AIS data were used to form a training sample. The first data point was fixed, and the second, third, and fourth points were intercepted at intervals of three to five points. This selection disrupted the personality association between the data and increased the sensitivity of the data to time parameters. The problems of gradient explosion and gradient disappearance were solved through batch training. The trained network model was highly versatile and directly usable; it did not require retraining for specific areas. The BI-LSTM-RNN network continued to learn online and adjusted the network in the actual application scenario. By using three historical data points at a time, six future ship position points could be predicted. Furthermore, by adjusting the network parameters according to the continuously generated ship data, six new future ship position points could be predicted (Figure 12).
The data from Tianjin Port for January–October were used as the training group to train the neural network parameters. The error was stable at approximately 90 m. The training error is displayed in Figure 13.
Previously published trajectory prediction algorithms had two major disadvantages: (1) The prediction algorithms mostly included multiple input points and eigenvalues, but only one output point, thus limiting their prediction capacity and consequently restricting their applicability and accessibility in navigation. In this study, the BI-LSTM-RNN structure used three historical ship position data points, and six points could be continuously predicted. Thereafter, the prediction accuracy gradually decreased. (2) Conventional prediction algorithms possess limited versatility, are based on a fixed mathematical model, and are only valid for learning the training data of a ship in advance. Conventional algorithms cannot be adaptively changed, and precise prediction and judgment of the targeted object are impossible, thus the versatility is weak. The BI-LSTM-RNN can be applied to improve versatility by retaining the navigational habits of ships included in AIS big data and using the forget operation on the individual cases of single-ship data. The object is forecasted online, and the characteristics of the current ship can be memorized for a short period of time. Thus, the navigation behavior of a ship can be precisely predicted.
Trajectory prediction involves grasping the movement of the ship and obtaining reliable collision avoidance decisions in advance. The trajectory prediction algorithm requires accuracy and timeliness. Greater consistency with reality improves the user’s likelihood of obtaining the correct conclusion. The time available for making collision avoidance decisions is very limited. Therefore, obtaining the prediction result quickly is necessary. To prove the superiority of the proposed prediction algorithm, AIS data from Tianjin Port from November to December, which included 7,530,103 data points, were used as the verification group. The data were not intercepted when the network was being trained, and all the data were directly taken as input for the navigation behavior of the ship. The ship used the BI-LSTM-RNN, BI-RNN, and LSTM-RNN to compare prediction errors. Figure 14 displays a comparison of the convergence effects of the three algorithms.
It can be clearly seen from Figure 14 that the neural network predicts the behavior of the ship. When predicted for a period of time, the data of a single ship in the verification dataset are limited, and the data after verification are replaced with the data of the other ship to continue the prediction and verification. Although the prediction accuracy deteriorates suddenly, it always converges in a short time and stabilizes. The convergence velocity, oscillation amplitude, and prediction accuracy of the BI-LSTM-RNN are superior to those of the LSTM-RNN and bidirectional RNN. The experimental results indicate that the BI-LSTM-RNN is trained to predict the position of a single ship in a short time, as illustrated in Figure 15 and Figure 16.
It can be seen from Figure 16 that although each conversion of the new ship data image will cause the rebound phenomenon to occur, it can quickly converge and stabilize in a short time. After about 15 batches of training, the predictive performance of the BI-LSTM-RNN appeared to be stable. The reliability of the navigation behavior prediction was 10 m or less within the accuracy of GPS positioning. The BI-LSTM-RNN was then applied as a prediction module.

5. Conclusions

We inferred the following from our experiment:
  • Selecting an RNN for the time-series data characteristics of AIS big data allows training regarding the general rules of ship maneuvering and motion characteristics.
  • Adding the LSTM unit improves the gradient loss caused by infinite-sequence data in the loop training. An RNN can remember the common features of the AIS big data and forget personality differences. Thus, the RNN has an autonomous choice to remember or forget.
  • By incorporating a two-way RNN structure, the network can learn the information provided by historical data and optimize the network by using future data. The current prediction can establish a strong correlation related to the context.
  • The trained BI-LSTM-RNN can accurately predict future ship navigation behavior and adjust parameters in real time with existing data as input.
The experimental results proved the reliability of the model. Trajectory prediction with the BI-LSTM-RNN can provide security for navigation and assist in trajectory planning and risk monitoring. This research provides a theoretical basis for the design of innovative intelligent collision avoidance systems for unmanned ships, prediction of the navigation behavior of other ships, and mitigation of the risks of vessel traffic service (VTS) management. This study also provides a theoretical basis for subsequent research. The BI-LSTM-RNN is suitable for shipment tracking, ship classification, and ship identification. Moreover, the determined vessel state can be used to estimate the future navigation trajectory, which ultimately assists in predicting the future sailing behavior and maneuvering intentions of a ship and generating an accurate collision avoidance strategy.

Author Contributions

Software, Writing-Original Draft Preparation, Writing-Review & Editing, M.G.; Supervision, G.S.; Validation, S.L.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 51709165; the Natural Science Foundation of Liaoning Province, grant number 20170540090; and the Fundamental Research Funds for the Central Universities, grant number 3132018306.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Han, Y.X.; Tang, X.M.; Han, S.C. Conflict-Free 4D Trajectory Prediction Based on Hybrid System Theory. J. Southwest Jiaotong Univ. 2012, 47, 1069–1074. [Google Scholar]
  2. Xu, T.; Liu, X.; Yang, X. Ship Trajectory Online Prediction Based on BP Neural Network Algorithm. In Proceedings of the International Conference on Information Technology, Computer Engineering and Management Sciences, Nanjing, China, 24–25 September 2011; Volume 1, pp. 103–106. [Google Scholar]
  3. Xu, T.; Cai, F.J.; Hu, Q.Y.; Chun, Y. Research on estimation of AIS vessel trajectory data based on Kalman filter algorithm. Mod. Electron. Tech. 2014, 5, 97–100. [Google Scholar]
  4. Liu, X.L.; Ruan, Q.S.; Gong, Z.Q. New Prediction Model of Ships GPS Navigation Positioning Trajectory. J. Jiangnan Univ. (Nat. Sci. Ed.) 2014, 13, 686–692. [Google Scholar]
  5. Zhen, R.; Jin, Y.X.; Hu, Q.Y.; Shi, C.J.; Wang, S.Z. Vessel Beavior Prediction Based on AIS Data and BP Neural Network. Navig. China 2017, 40, 6–10. [Google Scholar]
  6. Zhao, S.B.; Tang, C.; Liang, S.; Wang, D.J. Track prediction of vessel in controlled waterway based on improved Kalman filter. J. Comput. Appl. 2012, 32, 3247–3250. [Google Scholar] [CrossRef]
  7. Tong, X.; Chen, X.; Sang, L.; Mao, Z.; Wu, Q. Vessel trajectory prediction in curving channel of inland river. In Proceedings of the International Conference on Transportation Information and Safety (ICTIS), Wuhan, China, 25–28 June 2015; pp. 706–714. [Google Scholar]
  8. Kim, Y.S.; Hong, K.S. An imm algorithm for tracking maneuvering vehicles in an adaptive cruise control environment. Int. J. Control Autom. Syst. 2004, 2, 613–625. [Google Scholar]
  9. Mehrotra, K.; Mahapatra, P.R. A jerk model for tracking highly maneuvering targets. IEEE Trans. Aerosp. Electron. Syst. 1997, 33, 1094–1105. [Google Scholar] [CrossRef]
  10. Vaidehi, V.; Chitra, N.; Chokkalingam, M.; Krishnan, C.N. Neural network aided kalman filtering for multitarget tracking applications. Comput. Electr. Eng. 2001, 27, 217–228. [Google Scholar] [CrossRef]
  11. Gan, S.; Liang, S.; Li, K.; Deng, J.; Cheng, T. Ship trajectory prediction for intelligent traffic management using clustering and ANN. In Proceedings of the 2016 UKACC International Conference on Control (CONTROL), Belfast, UK, 31 August–2 September 2016; pp. 1–6. [Google Scholar]
  12. Cai, X.; Zhang, N.; Venayagamoorthy, G.K. Time series prediction with recurrent neural networks trained by a hybrid PSO-EA algorithm. Neurocomputing 2007, 70, 2342–2353. [Google Scholar] [CrossRef]
  13. Mao, C.H.; Pan, C.; Yin, B.; Lu, Y.N.; Xu, X.Q. Ship navigation trajectory prediction based on Gaussian process regression. Technol. Innov. Appl. 2017, 4, 28–29. [Google Scholar]
  14. Perera, L.P.; Oliveira, P.; Soares, C.G. Maritime Traffic Monitoring Based on Vessel Detection, Tracking, State Estimation, and Trajectory Prediction. IEEE Trans. Intell. Transp. Syst. 2012, 13, 1188–1200. [Google Scholar] [CrossRef]
  15. Becker, M.H.; Richard, K.; SaschaMacek, K.S.; Roland, J. 2D laser-based probabilistic motion tracking in urban-like environments. J. Braz. Soc. Mech. Sci. Eng. 2009, 31, 83–96. [Google Scholar] [CrossRef] [Green Version]
  16. Stubberud, S.C.; Kramer, K.A. Kinematic prediction for intercept using a neural kalman filter. IFAC Proc. Vol. 2005, 38, 144–149. [Google Scholar] [CrossRef]
  17. Sang, L.; Yan, X.; Wall, A. CPA Calculation Method based on AIS Position Prediction. J. Navig. 2016, 69, 1409–1426. [Google Scholar] [CrossRef]
  18. Nagai, T.; Watanabe, R. Applying position prediction model for path following of ship on curved path. In Proceedings of the 2016 IEEE Region 10 Conference (TENCON), Singapore, 22–25 November 2016; pp. 3675–3678. [Google Scholar]
  19. Borkowski, P. The Ship Movement Trajectory Prediction Algorithm Using Navigational Data Fusion. Sensors 2017, 17, 1432. [Google Scholar] [CrossRef] [PubMed]
  20. Zissis, D.; Xidias, E.K.; Lekkas, D. Real-time vessel behavior prediction. Evol. Syst. 2016, 7, 29–40. [Google Scholar] [CrossRef]
  21. Yin, J.C.; Zou, Z.J.; Xu, F.; Wang, N.N. Online ship roll motion prediction based on grey sequential extreme learning machine. Neurocomputing 2014, 129, 168–174. [Google Scholar] [CrossRef]
  22. Zhang, W.; Zou, Z.J.; Deng, D.H. A study on prediction of ship maneuvering in regular waves. Ocean Eng. 2017, 137, 367–381. [Google Scholar] [CrossRef]
  23. Breda, V.L. Capability Prediction: An Effective, Way to Improve Navigational Performance. J. Navig. 2000, 53, 343–354. [Google Scholar] [CrossRef]
  24. Potter, C.; Venayagamoorthy, G.K.; Kosbar, K. RNN based MIMO channel prediction. Signal Process. 2010, 90, 440–450. [Google Scholar] [CrossRef]
  25. Williams, R.J.; Zipser, D.A. Learning algorithm for continually running fully recurrent neural networks. Neural Comput. 2014, 1, 270–280. [Google Scholar] [CrossRef]
  26. Han, M.; Xi, J.; Xu, S.; Yin, F.L. Prediction of chaotic time series based on the recurrent predictor neural network. IEEE Trans. Signal Process. 2004, 52, 3409–3416. [Google Scholar] [CrossRef]
  27. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 2002, 45, 2673–2681. [Google Scholar] [CrossRef]
  28. Sak, H.; Senior, A.; Beaufays, F. Long short-term memory recurrent neural network architectures for large scale acoustic modeling. arXiv, 2014; arXiv:1402.1128. [Google Scholar]
  29. Yi, J.; Wen, Z.; Tao, J.; Ni, H.; Liu, B. CTC Regularized Model Adaptation for Improving LSTM RNN Based Multi-Accent Mandarin Speech Recognition. J. Signal Process. Syst. 2017, 90, 1–13. [Google Scholar]
  30. Gers, F.A.; Schmidhuber, E. LSTM recurrent networks learn simple context-free and context-sensitive languages. IEEE Trans. Neural Netw. 2001, 12, 1333–1340. [Google Scholar] [CrossRef] [PubMed]
  31. Blanco, A.; Delgado, M.; Pegalajar, M.C. A real-coded genetic algorithm for training recurrent neural networks. Neural Netw. 2001, 14, 93–105. [Google Scholar] [CrossRef]
  32. Pearlmutter, B.A. Learning state space trajectories in recurrent neural networks. Neural Comput. 1988, 1, 263–269. [Google Scholar] [CrossRef]
  33. Kumar, J.; Goomer, R.; Singh, A.K. Long Short Term Memory Recurrent Neural Network (LSTM-RNN) Based Workload Forecasting Model for Cloud Datacenters. Procedia Comput. Sci. 2018, 125, 676–682. [Google Scholar] [CrossRef]
  34. Liu, C.; Wang, Y.; Kumar, K.; Gong, Y. Investigations on speaker adaptation of LSTM RNN models for speech recognition. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing, Shanghai, China, 20–25 March 2016; pp. 5020–5024. [Google Scholar]
  35. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  36. Agarap, A.F.; Grafilon, P. Statistical analysis on e-commerce reviews, with sentiments classification using bidirectional recurrent neural network (RNN). arXiv, 2018; arXiv:1805.03687. [Google Scholar]
  37. Sundermeyer, M.; Alkhouli, T.; Wuebker, J.; Ney, H. Translation Modeling with Bidirectional Recurrent Neural Networks. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Doha, Qatar, 25–29 October 2014. [Google Scholar]
  38. Gao, M.; Shi, G.Y.; Li, W.F. Online compression algorithm of AIS trajectory data based on improved sliding window. J. Traffic Transp. Eng. 2018, 18, 218–227. [Google Scholar]
Figure 1. Basic recurrent neural network (RNN) structure.
Figure 1. Basic recurrent neural network (RNN) structure.
Sensors 18 04211 g001
Figure 2. Cell unit interconnection diagram.
Figure 2. Cell unit interconnection diagram.
Sensors 18 04211 g002
Figure 3. Long short-term memory (LSTM) cell unit structure diagram.
Figure 3. Long short-term memory (LSTM) cell unit structure diagram.
Sensors 18 04211 g003
Figure 4. Internal schematic of an LSTM cell unit.
Figure 4. Internal schematic of an LSTM cell unit.
Sensors 18 04211 g004
Figure 5. Bidirectional recurrent neural network (BI-RNN) structure diagram.
Figure 5. Bidirectional recurrent neural network (BI-RNN) structure diagram.
Sensors 18 04211 g005
Figure 6. BI-RNN weight relationship diagram.
Figure 6. BI-RNN weight relationship diagram.
Sensors 18 04211 g006
Figure 7. Bidirectional long short-term memory recurrent neural network (BI-LSTM-RNN) batch-training diagram.
Figure 7. Bidirectional long short-term memory recurrent neural network (BI-LSTM-RNN) batch-training diagram.
Sensors 18 04211 g007
Figure 8. Flowchart of the navigation prediction model using BI-LSTM-RNN.
Figure 8. Flowchart of the navigation prediction model using BI-LSTM-RNN.
Sensors 18 04211 g008
Figure 9. Automatic identification system (AIS) data diagram for the Tianjin Port area in 2015.
Figure 9. Automatic identification system (AIS) data diagram for the Tianjin Port area in 2015.
Sensors 18 04211 g009
Figure 10. BI-LSTM-RNN structure diagram.
Figure 10. BI-LSTM-RNN structure diagram.
Sensors 18 04211 g010
Figure 11. Improved sliding-window compression algorithm.
Figure 11. Improved sliding-window compression algorithm.
Sensors 18 04211 g011
Figure 12. BI-LSTM-RNN trajectory prediction diagram.
Figure 12. BI-LSTM-RNN trajectory prediction diagram.
Sensors 18 04211 g012
Figure 13. BI-LSTM-RNN training error change diagram.
Figure 13. BI-LSTM-RNN training error change diagram.
Sensors 18 04211 g013
Figure 14. Convergence effects of three prediction models: BI-LSTM-RNN (red line), BI-RNN (green line), and LSTM-RNN (blue line).
Figure 14. Convergence effects of three prediction models: BI-LSTM-RNN (red line), BI-RNN (green line), and LSTM-RNN (blue line).
Sensors 18 04211 g014
Figure 15. Schematic of ship behavior prediction.
Figure 15. Schematic of ship behavior prediction.
Sensors 18 04211 g015
Figure 16. Single-ship prediction error diagram for verification group.
Figure 16. Single-ship prediction error diagram for verification group.
Sensors 18 04211 g016
Table 1. Automatic identification system (AIS) data structure storage. MMSI, Maritime Mobile Service Identity.
Table 1. Automatic identification system (AIS) data structure storage. MMSI, Maritime Mobile Service Identity.
MMSIHeading (°)Course (°)Speed (Kn)UnixTime (ms)Lon_d (°)Lat_d (°)
1009000005.05.62.41,422,721,505117.99797338.669028
1009000002.02.52.41,422,721,943117.99826538.673763
100900000208.0208.12.91,422,723,323117.99990038.677282
100900000177.0177.32.71,422,728,073118.00074038.682742
100900000183.0183.82.71,422,753,718118.01162038.683072
100900000185.0185.42.61,422,814,193117.99673538.675307
100900000216.0216.42.91,422,820,375118.00024238.684793
100900000184.0241.53.41,422,845,041117.99377338.679000
100900000187.0184.62.51,422,851,329117.99159338.677032
100900000180.0187.72.81,422,925,499117.99161738.654512

Share and Cite

MDPI and ACS Style

Gao, M.; Shi, G.; Li, S. Online Prediction of Ship Behavior with Automatic Identification System Sensor Data Using Bidirectional Long Short-Term Memory Recurrent Neural Network. Sensors 2018, 18, 4211. https://doi.org/10.3390/s18124211

AMA Style

Gao M, Shi G, Li S. Online Prediction of Ship Behavior with Automatic Identification System Sensor Data Using Bidirectional Long Short-Term Memory Recurrent Neural Network. Sensors. 2018; 18(12):4211. https://doi.org/10.3390/s18124211

Chicago/Turabian Style

Gao, Miao, Guoyou Shi, and Shuang Li. 2018. "Online Prediction of Ship Behavior with Automatic Identification System Sensor Data Using Bidirectional Long Short-Term Memory Recurrent Neural Network" Sensors 18, no. 12: 4211. https://doi.org/10.3390/s18124211

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop