Next Article in Journal
Dynamic Zero Current Method to Reduce Measurement Error in Low Value Resistive Sensor Array for Wearable Electronics
Next Article in Special Issue
Position and Attitude Determination in Urban Canyon with Tightly Coupled Sensor Fusion and a Prediction-Based GNSS Cycle Slip Detection Using Low-Cost Instruments
Previous Article in Journal
Fusion Graph Representation of EEG for Emotion Recognition
Previous Article in Special Issue
Autonomous Driving Control Based on the Technique of Semantic Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Predictive Model of a Driver’s Target Trajectory Based on Estimated Driving Behaviors

The Institute of Industrial Science, The University of Tokyo, Tokyo 153-8505, Japan
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(3), 1405; https://doi.org/10.3390/s23031405
Submission received: 24 December 2022 / Revised: 18 January 2023 / Accepted: 21 January 2023 / Published: 26 January 2023
(This article belongs to the Special Issue On-Board and Remote Sensors in Intelligent Vehicles)

Abstract

:
With the development of automated driving, inferring a driver’s behavior can be a key element for designing an Advanced Driver Assistance System (ADAS). Current research is focused on describing and predicting a driver’s behaviors as labels, e.g., lane shifting, lane keeping, etc., during driving. In our work, we consider that predicting a driver’s behavior can be described as predicting a trajectory the driver may follow in the near future. The target trajectory can be calculated through certain polynomial functions. Via the data set collected by a Driving Simulator experiment covering nine volunteers, we proposed a model based on a deep learning network which is capable of predicting the corresponding coefficients of polynomial functions and then generating the trajectories in the next few seconds. The results also discussed and analyzed some possible factors affecting the prediction error. In conclusion, the model proved to be effective in predicting the target trajectory of a driver.

1. Introduction

Driving is becoming a complex activity with the increasing number of road vehicles. According to a report from the World Health Organization [1,2], every year over one million people are injured or killed due to traffic accidents. Some researchers noticed that many crashes occur during merging or changing lanes, and human errors were one of the major reasons [3,4]. As a result, the Advanced Driver Assistance System (ADAS) has become a promising system for reducing driving workload and helping drivers deal with complex situations. One classic function of an ADAS is the Lane-Keeping Assist System (LKAS), which can prevent unintentional lane departure by alert or vibration [5,6]. Some researchers also explored its application in assisting drivers to keep or change lanes [7,8].
ADASs show the potential of improving driving safety and efficiency. Predicting a driver’s behavior will help the ADAS to serve the driver more efficiently. Supposing that the LKAS is mistakenly activated when a driver tries to change lanes, he/she will feel interference from the ADAS [7]. Previous research also suggests that early detection of potential human errors can help the system to take corrective action such as braking, therefore reduce collision possibility [9]. In general, a driver’s behaviors can be classified into labels, such as changing or keeping lanes, merging, turning left, etc. Researchers have made a lot of effort in improving behavior prediction accuracy. Kumar et al. proposed a method based on the Support Vector Machine (SVM) and the Bayesian filter [10]. The model comprised vehicle dynamic input to detect the intention of lane shifting and predict probabilities of different behaviors as results. The data set was established on road experiments, but the selected features for prediction did not include any driver’s states such as eye or head movements. Kim et al. considered a preprocessing method based on road conditions and vehicle kinematic data to detect a driver’s intention [11]. Their data set was collected via a Driving Simulator (DS) experiment. Kim also explored different combinations in influencing prediction accuracy. Lethaus et al. proved that the gaze movements of a driver alone can be used for predicting behaviors [4]. They divided a driver’s vision into five areas: windshield, left and right window, rear-view mirror and speed indicator. By estimating the time that a driver’s gaze stayed in a certain place, they compared the performance of Neural Network, Bayesian Networks and Naïve Bayes Classifiers for intention prediction. Moreover, in [12], Dang et al. defined the behavior classification problem as a regression problem of when the ego vehicle will cross the borderline. They proved that a Deep-Neural-Network-based model was capable of predicting the time of crossing.
The researchers used yaw angle, lane departure distance, etc., to distinguish driving behaviors. As an example, a lane-changing behavior can be defined as when the vehicle crossed a lane boundary. However, there are in fact no clear boundaries for differentiating such behaviors. For a driver, driving behaviors such as changing or keeping lanes are tracking a continuous trajectory, and the labels are manually tagged without clear rules. A more ideal way is predicting a driver’s target trajectory in the upcoming few seconds. The previous research suggests in comparison with giving a classification label, a predicted target trajectory could be more helpful for ADAS [13,14,15]. A target trajectory contains more information than a simple label and can help ADASs in a more efficient way.
To be more specific, a target trajectory prediction problem can be described as shown in Figure 1, which asks the following question: at the current time t0, what is the trajectory that the driver may follow in the next Tp seconds?
In [16], Khakzar et al. proposed a method of prediction derived from the Second Strategic Highway Research Program (SHRP2) Naturalistic Driving Study (NDS) [17]. The researchers proposed a dual learning model and risk map for vehicle trajectory prediction. Khakzar et al. built a Deep Neural Network (DNN) to predict the behaviors in the upcoming future. They also analyzed the factors affecting prediction error such as drivers’ profiles. The conclusion suggested that self-reported questionnaires showed no significant difference in driving performance information. Liu used a kinematic model to predict the trajectories of the preceding vehicle [18]. They described the trajectory by calculating the corresponding coefficients of a cubic polynomial function. Other kinematic models such as Constant Turn Rate or Constant Velocity were also discussed and utilized for the prediction [19]. On the other hand, in [20], Hu et al. established a Gaussian mixture model based on a DNN to predict the probability distribution areas where the vehicle is occupied. Some researchers pointed out that such DNN-based methods consider the interactions vehicle-to-vehicle and environment-to-vehicle effectively. The fusion of information can help to improve behavior prediction accuracy [21,22].
Clearly, great achievements have been made in understanding and predicting the driver’s behavior. Previous research suggested that inferring the trajectories of a vehicle could be helpful for designing a more powerful ADAS. In this work, we proposed a method by considering a driver’s observing behavior, kinematic parameters and environmental factors. By considering those factors, our model can predict a driver’s target trajectory. We also compared the different combinations of features for reducing prediction error. The finding of effective features can be used for designing an accurate prediction model. The testing results suggest that our model can give an accurate prediction of the target trajectory.
This paper is organized as follows. Section 2 tells the methodology of describing and predicting the trajectory. Section 3 describes the Driving Simulator experiment, including the scenario, conditions, apparatus, participants and procedure. Section 4 provides the results of the experiment, and is followed by Section 5, where the performance of the system is discussed. Finally, the conclusion is presented in Section 6.

2. Trajectory Based on Polynomial Functions

As described in Section 1, if the current time is t0, our target is to predict the future positions of the ego vehicle in the next Tp seconds, as shown in Figure 1.
Polynomial functions have been proven effective in predicting and planning a smooth local trajectory [23,24]. Moreover, higher polynomials can be used to describe motions in lateral and longitudinal directions independently.
According to [23,25], the lateral motion is performed as
y r ( t ) = b 0 + b 1 t + b 2 t 2 + b 3 t 3 + b 4 t 4 + b 5 t 5
For the lateral motion, we assume the initial state is [ y r 0   y r 0 ˙   y r 0 ¨ ] , representing the lateral position, velocity, and acceleration at the beginning. Similarly, the final state is given as [ y r f   y r f ˙   y r f ¨ ].
The longitudinal motion is
x r ( t ) = a 0 + a 1 t + a 2 t 2 + a 3 t 3 + a 4 t 4
The initial state is defined as [ x r 0   x r 0 ˙   x r 0 ¨ ] , representing the longitudinal position, velocity, and acceleration at the beginning. The final state is given as [ x r f ˙   x r f ¨ ] , since the final longitudinal position is unknown a priori.
Further, x r 0 can be obtained by substituting t = 0 into (2):
x r 0 = a 0 + a 1 0 + a 2 0 + a 3 0 + a 4 0
Taking the derivative of (2), x r 0 ˙ can be obtained by substituting t = 0 into (4):
x r ( t ) ˙ = a 1 + 2 a 2 t + 3 a 3 t 2 + 4 a 4 t 3
x r 0 ˙ = a 1 + 2 a 2 0 + 3 a 3 0 + 4 a 4 0
Similarly, we have
x r 0 ¨ = 2 a 2 + 3 a 3 0 + 4 a 4 0
Substituting t = Tp, the prediction horizon as the final state, we have
x r f ˙ = a 1 + 2 a 2 T p + 3 a 3 T p 2 + 4 a 4 T p 3
x r f ¨ = 2 a 2 + 6 a 3 T p + 12 a 4 T p 2
For the lateral direction, we can obtain [ y r 0   y r 0 ˙   y r 0 ¨ ] and [ y r f   y r f ˙   y r f ¨ ] via the same way.
If the coefficients in (1) and (2) are known, the trajectories in lateral and longitudinal directions can be defined by time t. With the initial and final states, the coefficients in the polynomials can be obtained by solving the following equations:
[ a 0     a 1     a 2     a 3     a 4 ] T = A 1 X r
[ b 0     b 1     b 2     b 3     b 4     b 5 ] T = B 1 Y r
where the matrix A is given by (3), (5), (6), (7) and (8) (same for matrix B). Xr and Yr are the initial and final states, as follows:
A = [ 1 0 0 0 0 0 1 0 0 0 0 0 2 0 0 0 1 2 T p 3 T p 2 4 T p 3 0 0 2 6 T p 12 T p 2 ]
B = [ 1 0 0 0 0 0 0 1 0 0 0 0 0 0 2 0 0 0 1 T p T p 2 T p 3 T p 4 T p 5 0 1 2 T p 3 T p 2 4 T p 3 5 T p 4 0 0 2 6 T p 12 T p 2 20 T p 3 ]
X r = [ x r 0 x r 0 ˙ x r 0 ¨ x r f ˙ x r f ¨ ]
Y r = [ y r 0 y r 0 ˙ y r 0 ¨ y r f y r f ˙ y r f ¨ ]
Therefore, we have:
{ a 0 = x r 0 a 1 = x r 0 ˙ a 2 = x r 0 ¨ / 2 a 3 = 3 x r 0 ˙ 3 x r f ˙ + 2 T p x r 0 ¨ + T p x r f ¨ 3 T p 2 a 4 = 2 x r 0 ˙ 2 x r f ˙ + T p x r 0 ¨ + T p x r f ¨ 4 T p 3
{ b 0 = y r 0 b 1 = y r 0 ˙ b 2 = y r 0 ¨ 2 b 3 = ( 20 y r 0 20 y r f + 12 T p y r 0 ˙ + 8 T p y r f ˙ + 3 T p 2 y r 0 ¨ T p 2 y r f ¨ ) 2 T p 3 b 4 = 30 y r 0 30 y r f + 16 T p y r 0 ˙ + 14 T p y r f ˙ + 3 T p 2 y r 0 ¨ 2 T p 2 y r f ¨ 2 T p 4 b 5 = ( 12 y r 0 12 y r f + 6 T p y r 0 ˙ + 6 T p y r f ˙ + T p 2 y r 0 ¨ T p 2 y r f ¨ ) 2 T p 5
The coefficients in (1) and (2) can be obtained with (15) and (16). Since the initial states are known, the prediction of a trajectory depends on the final states, x r f ˙ ,   x r f ¨ and y r f ,   y r f ˙ ,   y r f ¨ . Therefore, the target turns into predicting the five variables as the final states.
Here, we re-form the problem of prediction as the following description:
For time moment t, we assume the data collected from sensors are St. Assuming the current time is t0, all the data collected in the past k time step up to t0 form a matrix, S t 0 = ( s t 0 k + 1 , s t 0 k + 2 , , s t 0 1 , s t 0 ) . The problem changes into finding a model that gives the prediction of the final states in the longitudinal and lateral direction in Tp seconds later.

3. Methodology

With a clear definition of the issue, the key issue turns into building a regression model. The Recurring Neural Network (RNN) is proven to be an efficient tool for such tasks, particularly for a problem that is difficult to describe in a simple mathematical function. Meanwhile, as a popular RNN, the Gated-Recurrent-Unit (GRU) network has received a great deal of attention since its proposal in 2014 [26]. It is a powerful tool in time-series-based classification and regression problems. In addition, GRU has a simplified architecture, but shows similar performance in lots of tasks [26] in comparison with another widely used Recurrent Neural Network, Long-Short-Term-Memory (LSTM) [27].
The RNN-based deep learning method is widely used for solving such regression problems. It requires data sets for training and testing. Therefore, we conducted a Driving Simulator (DS) experiment for collecting data.

3.1. Data Set

Nine participants (including seven men and two women) were invited to participate in the experiment. Volunteers ranged in age between 19 and 26 years (mean age = 22.8, standard deviation = 2.4). All participants held a Japanese driver’s license (mean driving years = 2.6 years, standard deviation = 1.9 years). All participants received payment for their participation. The experiment was endorsed by the University of Tokyo’s Office for Life Science Research Ethic and Safety.
Each participant drove on a highway in a simulator environment (as shown in Figure 2). Participants were free to steer the car and were suggested to drive under a speed limitation (60 km/h). The experiment scenario included several segments, in which a leading vehicle whose speed was limited to slower than the ego vehicle then triggered a lane change event to collect driving trajectories. For each event, the leading vehicle would decelerate at a certain rate which was randomly selected from −1 m/s2 to −3 m/s2 until its speed was slower than the ego vehicle with at least 10 km/h difference.
The details of sensor data S t 0 for prediction are shown in Table 1. A gaze-tracking system was used to measure the gaze and head movements [28]. The gaze-tracking system (in Figure 3) did not create physical burdens on the participants. All the drivers drove in a natural state and were told to follow the basic traffic rules.
Therefore, S t 0 is
S t 0 = [ H e a d x , t 0 k + 1 G a z e y , t 0 k + 1 d t 0 k + 1 ψ t 0 k + 1 H e a d x , t 0 k + 2 G a z e y , t 0 k + 2 d t 0 k + 2 ψ t 0 k + 2 H e a d x , t 0 G a z e y , t 0 d t 0 ψ t 0 ]

3.2. Labeling

The prediction model outputs a vector [ x r f ˙ ,   x r f ¨ , y r f , y r f ˙ ,   y r f ¨ ], as mentioned in Section 2. An example of predicting a trajectory is shown in Figure 4. The model predicted [ x r f ˙ ,   x r f ¨ , y r f , y r f ˙ ,   y r f ¨ ] in the next Tp seconds then obtain the coefficients in polynomials from (15) and (16). By having the coefficients, the trajectory can be calculated.

3.3. Network Architecture

A short time before the present moment may cover important information to predict future behaviors, because drivers are meant to perform a series of behaviors prior to a maneuverer. As a result, this task requires a method that is able to remember the information at an earlier time and uncover the hidden dependence in the time series. RNN can retain a state and has been proven effective in performing such tasks. However, researchers pointed out that the vanilla RNN had a potential risk of gradient exploding or vanishing. These problems sometime made the vanilla RNN hard to train [29]. GRU and LSTM are two popular extended RNNs [26,27]. In comparison with the vanilla RNN, their modified architectures overcome such disadvantages. Both LSTM and GRU achieve memory function through a gating mechanism, but GRU simplifies the architecture and does not have memory nodes. The performance of the two RNNs is similar for most of the tasks, but GRU usually needs less time to train. In this research, we focused on utilizing GRU to build the model [7].
GRU can decide how much the unit should update its content through two gates and remember information through the unit hidden state  h   t j . To update h   t j (j = the j-th unit, t = time step), the process is as follows:
Hidden   state :   h   t j = ( 1 z   t j ) h   t 1 j + z   t j h   ˜ t j
where z   t j is the update gate controlling how much content should be updated. When z   t j = 0 , h   t 1 j will be copied to h   t j without loss; z   t j is described as
Update   gate :   z t j = σ ( W z x t + U z h t 1 ) j
where the new memory  h   ˜ t j summarizes the information from the new entry xt and ht−1 (the hidden state at time step t − 1):
New   memory :   h   ˜ t j = tanh ( W z x t + U ( r t h t 1 ) ) j
where r   t j is the reset gate deciding the contents to be abandoned from the hidden state. The previous hidden state h   t 1 j will not pass to the new memory h ˜ t j if r   t j = 0 , where r   t j is described as
Reset   gate :   r t j = σ ( W r x t + U r h t 1 ) j
the hidden state representing the memory can be updated selectively through the update gate and the reset gate. Wr, Wz, Ur and Uz are learnable parameters during the network training.
In this research, the network begins with a layer of GRU with 180 units and is connected by two dense layers with 128 units. When we validated our method, we found that more recurrent units would cost significantly more time to train but did not improve the accuracy obviously, while reducing the number of network units caused the network to lack the ability to predict. Figure 5 shows the network architecture.

4. Results

A data set covering nine participants was collected during the experiment. Each participant drove for about 45 to 60 minutes on a highway. All the participants were free to steer or accelerate while driving and could change lanes whenever they thought necessary. In each trial, the drivers were required to follow the basic traffic rules in Japan.
The data set covered trajectories during the whole driving process. We collected 23,168 samples in total, including 16,256 samples for training, 4032 samples for validating and 2880 samples for testing. Tp was set as four seconds for our research. Each sample contained an S t 0 and [ x r f ˙ ,   x r f ¨ , y r f , y r f ˙ ,   y r f ¨ ] at corresponding times. The testing samples were from three randomly selected participants. Training samples were shuffled and collected from all participants. We fixed the sampling window as three seconds and the sampling rate = 60 Hz. The time step parameter k in S t 0 was set as 180.

4.1. Predicted Trajectories

Following the steps in the previous section, we established a model for predicting a driver’s target trajectories four seconds (Tp = 4 s) later, as shown in Figure 6. The trajectory was from subject 7 during lane changing.
Since the predicted trajectory and the actual trajectory were almost aligned, our model could give a prediction of the driver’s target trajectory.

4.2. Feature Selection for Prediction

For a driver, he/she intends to check side mirrors to make sure there is a safe environment to perform maneuvers. Observational behaviors can be useful for predicting trajectories. Some researchers pointed out that the gaze and head data performed closely on improving accuracy for driving behavior prediction [3,4]. However, more discussion is still needed to evaluate the roles of gaze and head dynamic data in trajectory prediction performance. Here, we compared the Final Displacement Error (FDE) that was obtained from models trained via different combinations of the features in Table 1. The FDE is given by
F D E = ( x t , i p r e d x t , i a c t u a l ) 2
where x t , i p r e d and x t , i a c t u a l are the predicted position and the actual position of a trajectory at time step t, respectively, i = the i-th sample and t = Tp = 4 s. FDE suggests the general performance of trajectory prediction.
Table 2 shows that combination (d) has the smallest prediction error in the lateral direction among the four groups. The combination (a) without driver states showed the largest error.
In Table 3, the Kruskal–Wallis tests between each pair of combinations indicated that combination (d) did not show a significant difference in comparison with combination (c). Combination (c), which only included head dynamic data, also significantly reduces the prediction error compared with combination (a), while combination (b) showed no difference in improving the prediction accuracy.
Meanwhile, for longitudinal direction, the combinations did not show differences in reducing the prediction error.

4.3. Trajectory Prediction Performance

To further evaluate the prediction performance, we estimated the Root Mean Square Error (RMSE), given by (18) as shown in Figure 7:
R M S E = 1 n i = 1 n ( x t , i p r e d x t , i a c t u a l ) 2
where x t , i p r e d and x t , i a c t u a l are the predicted position and the actual position of a trajectory at time step t, respectively, n = the number of samples and i = the i-th sample.
Figure 7 depicts the lateral and longitudinal RMSE of the prediction model up to four seconds. It can be seen that with the prediction horizon increasing, the prediction error also increases in both directions.
For each sample, we also estimated its Final Displacement Difference (FDD) in both directions. The FDD is given by
F D D = x t , i p r e d x t , i a c t u a l
where x t , i p r e d , x t , i a c t u a l , i and t are the same as in (17) and t = Tp = 4 s.
Figure 8 suggests the prediction displacement distribution. For lateral movement, the FDD followed a normal distribution with μ = 0.037 and σ = 0.68, as shown in Figure 8a. For longitudinal movement, the FDD followed a normal distribution with μ = 0.241 and σ = 0.924, as shown in Figure 8b.
Figure 9 describes the displacement distribution in both lateral and longitudinal directions. It can be seen that the displacement difference distributes around (0, 0).

4.4. Comparison of Different Neural Networks

LSTM and GRU have similar architectures, but GRU has fewer parameters, which makes GRU less prone to over-fitting and easier to train [7]. Tuning the LSTM network will take a longer time and more epochs than GRU by the same optimizer [29,30]. Here, we trained two models with the same architecture as Figure 5 and made a comparison of their time costs under the same environment. All other training parameters were also kept the same. The time costs are shown in Figure 10 and Table 4.
The training was run on an nVIDIA GTX 1070 with Cudnn acceleration and lasted 120 epochs. It can be seen from Figure 10 and Table 4 that GRU spent significantly less time than LSTM for each training epoch. In fact, in order to tune LSTM and GRU to reach almost the same performance, LSTM needs about 10%~15% more time than GRU.
Meanwhile, LSTM did not significantly reduce the prediction error. Therefore, for this task, GRU was proved more suitable than LSTM.

5. Discussion

The purpose of this research is to propose a method of predicting a driver’s target trajectory in the upcoming future. We compared the combinations of features to select the most effective set for reducing the prediction error. Combination (d) (with gaze and head) showed the smallest error among the four combinations. Taking combination (a) as the baseline, it can be seen from Table 2 that combination (b) (only including gaze) did not improve the prediction accuracy. Combinations (d) and (c) (only including head) also indicated no difference. These results suggested that gaze might have no effect in improving the accuracy. It is possible that gaze is not robust enough for monitoring a driver’s behavior, as shown in Figure 11.
It can be seen that when time went to 73.7 to 74.5 s, Gaze suddenly jumped to zero. This occurred when our eye-tracking device was unable to capture the gaze movement and it happened a few times. Instead, head remained consistent. A driver may peek at mirrors and move their head while driving. Since the cameras were fixed, it was likely that the cameras failed to track the driver’s pupils when he/she moved too fast. These results correspond with other research [3,7], indicating that sometimes gaze movements are not stable enough for estimating a driver’s behavior accurately. In addition, in a few situations, the system may give a prediction with relatively big errors. These may also result from the noise of head or gaze movement. Such conclusions suggest that for an accurate trajectory prediction model, head movement data are essential, but gaze movement, due to its unreliability, may not affect the prediction results significantly.
We evaluated the RMSE and FDD of predicting a trajectory across the testing set, and the RMSE estimated based on our method was smaller than the prediction method in [2]. For longitudinal movement, the FDD distribution followed a normal distribution with μ = 0.241 and σ = 0.924, which also suggests less error than the method in [2]. However, for lateral movement, the FDD followed a normal distribution, which suggests that our method can predict a driver’s target trajectory in both directions accurately.
We compared the performance of LSTM and GRU. LSTM and GRU performed closely for our task, but training GRU usually takes less time, which makes it easier to use and tune.

6. Conclusions

This research explores a way to predict a driver’s target trajectory by describing a trajectory via polynomials on lateral and longitudinal directions separately. Through a DS experiment, we collected a data set covering drivers, environment and driver states. A model based on GRU was established, trained and validated. The results suggested that a robust prediction model mainly relied on head movement monitoring. In the Discussion, the reasons for the unreliability of gaze movement were discussed and analyzed. It was because that tracking gaze movement accurately could be hard for the current device. Our method has proved to be effective in predicting a driver’s future trajectory up to four seconds. The GRU showed to be more effective in saving training time without losing prediction accuracy.
This prediction system can be applied to an ADAS to help the system to better cooperate with a driver. The future plan needs to consider recruiting more volunteers to enlarge the data set.

Author Contributions

Conceptualization, Z.Y.; validation, Z.Y.; formal analysis, Z.Y.; data curation, Z.Y.; visualization, Z.Y.; writing—original draft preparation, Z.Y.; writing—review and editing, B.Y., Z.W. and K.N.; supervision, B.Y., Z.W. and K.N.; project administration, K.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by a Grant-in-Aid for Early-Career Scientists (21K17781) from the Japan Society.

Institutional Review Board Statement

The experiment received approval by the Office for Life Science Research Ethic and Safety, the University of Tokyo.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization. Global Status Report on Road Safety; World Health Organization: Geneva, Switzerland, 2015; pp. 11–13. [Google Scholar]
  2. Khakzar, M.; Bond, A.; Rakotonirainy, A.; Oviedo Trespalacios, O.; Dehkordi, S.G. Driver Influence on Vehicle Trajectory Prediction. Accid. Anal. Prev. 2021, 157, 106165. [Google Scholar] [CrossRef] [PubMed]
  3. Doshi, A.; Trivedi, M.M. On the Roles of Eye Gaze and Head Dynamics in Predicting Driver’s Intent to Change Lanes. IEEE Trans. Intell. Transp. Syst. 2009, 10, 453–462. [Google Scholar] [CrossRef] [Green Version]
  4. Lethaus, F.; Baumann, M.R.K.; Köster, F.; Lemmer, K. A Comparison of Selected Simple Supervised Learning Algorithms to Predict Driver Intent Based on Gaze Data. Neurocomputing 2013, 121, 108–130. [Google Scholar] [CrossRef]
  5. Wang, Z.; Zheng, R.; Kaizuka, T.; Shimono, K.; Nakano, K. The Effect of a Haptic Guidance Steering System on Fatigue-Related Driver Behavior. IEEE Trans. Human-Machine Syst. 2017, 47, 741–748. [Google Scholar] [CrossRef]
  6. Raouf, I.; Khan, A.; Khalid, S.; Sohail, M.; Azad, M.M.; Kim, H.S. Sensor-Based Prognostic Health Management of Advanced Driver Assistance System for Autonomous Vehicles: A Recent Survey. Mathematics 2022, 10, 3233. [Google Scholar] [CrossRef]
  7. Yan, Z.; Yang, K.; Wang, Z.; Yang, B.; Kaizuka, T.; Nakano, K. Time to Lane Change and Completion Prediction Based on Gated Recurrent Unit Network. In Proceedings of the IEEE Intelligent Vehicles Symposium, Paris, France, 9–12 June 2019. [Google Scholar]
  8. Yan, Z.; Yang, K.; Wang, Z.; Yang, B.; Kaizuka, T.; Nakano, K. Intention-Based Lane Changing and Lane Keeping Haptic Guidance Steering System. IEEE Trans. Intell. Veh. 2021, 6, 622–633. [Google Scholar] [CrossRef]
  9. Harper, C.D.; Hendrickson, C.T.; Samaras, C. Cost and Benefit Estimates of Partially-Automated Vehicle Collision Avoidance Technologies. Accid. Anal. Prev. 2016, 95, 104–115. [Google Scholar] [CrossRef] [PubMed]
  10. Kumar, P.; Perrollaz, M.; Lefevre, S.; Laugier, C. Learning-Based Approach for Online Lane Change Intention Prediction. In Proceedings of the IEEE Intelligent Vehicles Symposium, London, UK, 15–17 July 2013. [Google Scholar]
  11. Kim, I.H.; Bong, J.H.; Park, J.; Park, S. Prediction of Driver’s Intention of Lane Change by Augmenting Sensor Information Using Machine Learning Techniques. Sensors 2017, 17, 1350. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Dang, H.Q.; Fürnkranz, J.; Biedermann, A.; Hoepfl, M. Time-to-Lane-Change Prediction with Deep Learning. In Proceedings of the IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC, Maui, HI, USA, 4–7 November 2018. [Google Scholar]
  13. Messaoud, K.; Deo, N.; Trivedi, M.M.; Nashashibi, F. Trajectory Prediction for Autonomous Driving Based on Multi-Head Attention with Joint Agent-Map Representation. In Proceedings of the IEEE Intelligent Vehicles Symposium, Nagoya, Japan, 11–17 July 2021. [Google Scholar]
  14. Park, S.H.; Kim, B.; Kang, C.M.; Chung, C.C.; Choi, J.W. Sequence-to-Sequence Prediction of Vehicle Trajectory via LSTM Encoder-Decoder Architecture. In Proceedings of the IEEE Intelligent Vehicles Symposium, Changshu, China, 26–30 June 2018. [Google Scholar]
  15. Sohn, K.; Yan, X.; Lee, H. Learning Structured Output Representation Using Deep Conditional Generative Models. In Proceedings of the Advances in Neural Information Processing Systems, Montréal, QC, Canada, 7–10 December 2015. [Google Scholar]
  16. Khakzar, M.; Rakotonirainy, A.; Bond, A.; Dehkordi, S.G. A Dual Learning Model for Vehicle Trajectory Prediction. IEEE Access 2020, 8, 21897–21908. [Google Scholar] [CrossRef]
  17. Antin, J.F.; Lee, S.; Perez, M.A.; Dingus, T.A.; Hankey, J.M.; Brach, A. Second Strategic Highway Research Program Naturalistic Driving Study Methods. Saf. Sci. 2019, 119, 2–10. [Google Scholar] [CrossRef]
  18. Liu, X.; Wang, Y.; Zhou, Z.; Nam, K.; Wei, C.; Yin, C. Trajectory Prediction of Preceding Target Vehicles Based on Lane Crossing and Final Points Generation Model Considering Driving Styles. IEEE Trans. Veh. Technol. 2021, 70, 8720–8730. [Google Scholar] [CrossRef]
  19. Schubert, R.; Richter, E.; Wanielik, G. Comparison and Evaluation of Advanced Motion Models for Vehicle Tracking. In Proceedings of the 11th International Conference on Information Fusion, Cologne, Germany, 30–31 July 2008. [Google Scholar]
  20. Hu, Y.; Zhan, W.; Tomizuka, M. Probabilistic Prediction of Vehicle Semantic Intention and Motion. In Proceedings of the IEEE Intelligent Vehicles Symposium, Changshu, China, 26–30 June 2018. [Google Scholar]
  21. Lee, N.; Choi, W.; Vernaza, P.; Choy, C.B.; Torr, P.H.S.; Chandraker, M. DESIRE: Distant Future Prediction in Dynamic Scenes with Interacting Agents Namhoon. In Proceedings of the Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  22. Zhao, T.; Xu, Y.; Monfort, M.; Choi, W.; Baker, C.; Zhao, Y.; Wang, Y.; Wu, Y.N. Multi-Agent Tensor Fusion for Contextual Trajectory Prediction. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  23. Benloucif, A.; Nguyen, A.T.; Sentouh, C.; Popieul, J.C. Cooperative Trajectory Planning for Haptic Shared Control between Driver and Automation in Highway Driving. IEEE Trans. Ind. Electron. 2019, 66, 9846–9857. [Google Scholar] [CrossRef]
  24. Houenou, A.; Bonnifait, P.; Cherfaoui, V.; Yao, W. Vehicle Trajectory Prediction Based on Motion Model and Maneuver Recognition. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Tokyo, Japan, 3–8 November 2013. [Google Scholar]
  25. Glaser, S.; Vanholme, B.; Mammar, S.; Gruyer, D.; Nouvelière, L. Maneuver-Based Trajectory Planning for Highly Autonomous Vehicles on Real Road with Traffic and Driver Interaction. IEEE Trans. Intell. Transp. Syst. 2010, 11, 589–606. [Google Scholar] [CrossRef]
  26. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. 2014, pp. 1–9. Available online: https://arxiv.org/abs/:1412.3555 (accessed on 10 November 2022).
  27. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  28. Wang, Z.; Zheng, R.; Kaizuka, T.; Nakano, K. Relationship between Gaze Behavior and Steering Performance for Driver-Automation Shared Control: A Driving Simulator Study. IEEE Trans. Intell. Veh. 2019, 4, 154–166. [Google Scholar] [CrossRef]
  29. Lipton, Z.C.; Berkowitz, J.; Elkan, C. A Critical Review of Recurrent Neural Networks for Sequence Learning. 2015, pp. 1–38. Available online: https://arxiv.org/abs/1506.00019 (accessed on 10 November 2022).
  30. Yang, S.; Yu, X.; Zhou, Y. LSTM and GRU Neural Network Performance Comparison Study: Taking Yelp Review Dataset as an Example. In Proceedings of the Proceedings—2020 International Workshop on Electronic Communication and Artificial Intelligence, IWECAI 2020, Shanghai, China, 12–14 June 2020. [Google Scholar]
Figure 1. Prediction of a target trajectory (Tp seconds later).
Figure 1. Prediction of a target trajectory (Tp seconds later).
Sensors 23 01405 g001
Figure 2. Segment for lane-changing event in DS environment.
Figure 2. Segment for lane-changing event in DS environment.
Sensors 23 01405 g002
Figure 3. Smart Eye Pro eye-tracking system and DS image. The eye-tracking device relies on infrared cameras (indicated in the photo) to track the movement of the head and eyes.
Figure 3. Smart Eye Pro eye-tracking system and DS image. The eye-tracking device relies on infrared cameras (indicated in the photo) to track the movement of the head and eyes.
Sensors 23 01405 g003
Figure 4. An example of a predicted target trajectory (in Tp later).
Figure 4. An example of a predicted target trajectory (in Tp later).
Sensors 23 01405 g004
Figure 5. The Neural Network architecture.
Figure 5. The Neural Network architecture.
Sensors 23 01405 g005
Figure 6. A predicted target trajectory when the driver was changing lanes (from subject 7).
Figure 6. A predicted target trajectory when the driver was changing lanes (from subject 7).
Sensors 23 01405 g006
Figure 7. The RMSE in lateral and longitudinal directions. When predicted horizon = 4 s, the prediction reaches max error with 0.68 m (lateral) and 0.95 m (longitudinal).
Figure 7. The RMSE in lateral and longitudinal directions. When predicted horizon = 4 s, the prediction reaches max error with 0.68 m (lateral) and 0.95 m (longitudinal).
Sensors 23 01405 g007
Figure 8. (a) The histogram of lateral displacement difference with fitted normal distribution. (b) The histogram of longitudinal displacement difference with fitted normal distribution.
Figure 8. (a) The histogram of lateral displacement difference with fitted normal distribution. (b) The histogram of longitudinal displacement difference with fitted normal distribution.
Sensors 23 01405 g008
Figure 9. The histogram of lateral and longitudinal displacement difference.
Figure 9. The histogram of lateral and longitudinal displacement difference.
Sensors 23 01405 g009
Figure 10. The time costs of LSTM and GRU.
Figure 10. The time costs of LSTM and GRU.
Sensors 23 01405 g010
Figure 11. The head movement and gaze movement of subject 7 before changing lanes (occurred at about time = 75 s).
Figure 11. The head movement and gaze movement of subject 7 before changing lanes (occurred at about time = 75 s).
Sensors 23 01405 g011
Table 1. Details of monitoring data.
Table 1. Details of monitoring data.
DataDescription
Driver stateHead movementHead, left/right rotation of the head, the angle in the horizontal direction
Gaze movementGaze, the gaze movement measured in the horizontal direction, normalized to [−1, 1]
Vehicle
state
Accelerationax, ay
Acceleration in the horizontal and vertical direction
Velocityvx, vy
Velocity in the horizontal and vertical direction
Steering wheel angleSWA
Environment stateLateral distanced, the distance to the adjacent lane
Yaw angleψ, the angle between the vehicle’s longitudinal axis and the lane
Table 2. Lateral mean FDE of different combinations of features.
Table 2. Lateral mean FDE of different combinations of features.
CombinationLateral Mean FDE (m)
(a)Environment + vehicle0.5789
(b)Environment + vehicle + gaze0.5674
(c)Environment + vehicle + head0.5228
(d)Environment + vehicle + head + gaze0.4835
The mean FDE = 1 n i = 1 n ( x t , i p r e d x t , i a c t u a l ) 2 , where n = the number of samples.
Table 3. Kruskal–Wallis test of lateral FDE of different combinations of features.
Table 3. Kruskal–Wallis test of lateral FDE of different combinations of features.
(a)(b)(c)(d)
(a)----
(b)0.707---
(c)0.000 ***0.000 ***--
(d)0.000 ***0.000 ***0.888-
***: p < 0.001.
Table 4. The time costs of LSTM and GRU.
Table 4. The time costs of LSTM and GRU.
RNNMean Time Cost of Each Epoch (s)
GRU2.77 ***
LSTM3.16
***: p < 0.001. The Kruskal–Wallis test result between GRU and LSTM.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, Z.; Yang, B.; Wang, Z.; Nakano, K. A Predictive Model of a Driver’s Target Trajectory Based on Estimated Driving Behaviors. Sensors 2023, 23, 1405. https://doi.org/10.3390/s23031405

AMA Style

Yan Z, Yang B, Wang Z, Nakano K. A Predictive Model of a Driver’s Target Trajectory Based on Estimated Driving Behaviors. Sensors. 2023; 23(3):1405. https://doi.org/10.3390/s23031405

Chicago/Turabian Style

Yan, Zhanhong, Bo Yang, Zheng Wang, and Kimihiko Nakano. 2023. "A Predictive Model of a Driver’s Target Trajectory Based on Estimated Driving Behaviors" Sensors 23, no. 3: 1405. https://doi.org/10.3390/s23031405

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop