Next Article in Journal
Point Cloud Resampling by Simulating Electric Charges on Metallic Surfaces
Next Article in Special Issue
Energy Management Strategy Based on a Novel Speed Prediction Method
Previous Article in Journal
A DNN-Based UVI Calculation Method Using Representative Color Information of Sun Object Images
Previous Article in Special Issue
5G and IoT Based Reporting and Accident Detection (RAD) System to Deliver First Aid Box Using Unmanned Aerial Vehicle
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Input and Multi-Channel Convolutional Neural Network Model for Vehicle Speed Prediction

1
State Key Laboratory of Automotive Dynamic Simulation and Control, Jilin University, Changchun 130022, China
2
Research Institute, Jilin University, Yibin 644500, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(22), 7767; https://doi.org/10.3390/s21227767
Submission received: 5 November 2021 / Revised: 20 November 2021 / Accepted: 20 November 2021 / Published: 22 November 2021

Abstract

:
With the development of technology, speed prediction has become an important part of intelligent vehicle control strategies. However, the time-varying and nonlinear nature of vehicle speed increases the complexity and difficulty of prediction. Therefore, a CNN-based neural network architecture with two channel input (DICNN) is proposed in this paper. With two inputs and four channels, DICNN can predict the speed changes in the next 5 s by extracting the temporal information of 10 vehicle signals and the driver’s intention. The prediction performances of DICNN are firstly examined. The best RMSE, MAE, ME and R2 are obtained compared with a Markov chain combined with Monte Carlo (MCMC) simulation, a support vector machine (SVM) and a single input CNN (SICNN). Secondly, equivalent fuel consumption minimization strategies (ECMS) combining different vehicle speed prediction methods are constructed. After verification by simulation, the equivalent fuel consumption of the simulation increases by only 4.89% compared with dynamic-programming-based energy management strategy and decreased by 5.40% compared with the speed prediction method with low accuracy.

1. Introduction

Vehicle speed prediction refers to the estimation of future vehicle speed with available information. By utilizing the predicted vehicle speed, the vehicle controller can adopt appropriate control strategies and the driver can adjust the manipulation in advance [1,2]. At present, vehicle speed prediction has become one of the research focuses in the field of intelligent vehicles as well as intelligent transportation, and is widely used in vehicle path planning, collision warning, advanced driver assistance systems, etc. [3,4,5]. However, vehicle speed prediction is affected by various factors such as human, vehicle, and road and traffic conditions [6]. The characteristics of time-varying and nonlinearity increase the complexity and difficulty of vehicle speed prediction. Improving the accuracy of vehicle speed prediction has important theoretical value and wide application prospects.
The current research methods for vehicle speed prediction can be generally divided into a model-based method and a data-based method. Model-based methods include the microscopic model, macroscopic model, etc. The microscopic model focuses more on the speed variation of each vehicle. Differential equations are established to simulate vehicle traveling. The speed can be updated by continuously solving the differential equations. Ghandour et al. proposed a new method for predicting the future vehicle motion state, including longitudinal speed, which was applied to predict potential danger and achieved satisfactory results [7]. Ma, Z et al. proposed a novel method to identify the normal load of each tire in a heavy truck in order to estimate the longitudinal velocity. The simulation showed that the predicted velocity could follow the trend of the real velocity and no significant lag was found [8]. The purpose of the microscopic model is not to predict vehicle speed with high accuracy, but to analyze vehicle dynamics in conjunction with speed prediction trends. The macroscopic model takes the vehicle group as the object of study and focuses more on the statistical characteristics of traffic flow. Magliolo et al. proposed a multiclass freeway traffic model that was applied to estimate the average speed of traffic for different classes of vehicles. Good prediction results were obtained in a test on a section of the Italian freeway [9]. Bie et al. proposed a macroscopic traffic prediction model that took weather into account. It was verified that the accuracy of prediction was improved compared with the traditional model [10]. The macroscopic model usually focuses on considering the average speed of a road or area. Therefore, it is difficult to provide accurate information about the speed variation of individual vehicles.
The model-based speed prediction method has fast parameter learning speed and does not require large data sets. However, its weak fitting ability leads to difficulties in constructing accurate mathematical models for highly stochastic vehicle speed.
The data-based speed prediction method can improve the ability to fit functional relationships between variables by utilizing large amounts of data. It does not make excessive assumptions about the prediction model and is free to learn arbitrary forms of mapping relationships from the training data. The most commonly applied data-based methods include stochastic prediction [11], shallow learning [12] and deep learning [13]. The most commonly applied method in stochastic prediction is the Markov chain constructed on the basis of the Markov process. The Markov process is a class of stochastic processes with a non-aftereffect property. The term “non-aftereffect” means that changes in the system state after the current phase are related only to the current phase and not to all previous phases. Shin et al. proposed a speed prediction algorithm based on fuzzy Markov chains to solve the problem of increasing computational power, which reduces the computational time by 85.5% while maintaining 99.1% prediction accuracy [14]. Zhou et al. proposed a vehicle speed prediction method based on a self-learning multi-step Markov model. The simulation results showed that the average prediction error was reduced by 25.70% compared with the traditional Markov predictor [15]. The Markov model assumes that the probability of a vehicle state change is kept fixed. If the driving condition differs significantly from the historical data, the prediction effectiveness of Markov chain decreases markedly.
The main purpose of shallow learning is to construct a data-driven mathematical model that finds the mapping between the input set and the output set. Common shallow learning methods are neural networks [1,6,16], support vector machines [4,17], extreme learning machines [18,19], etc. Zhang et al. developed a time-based neural network model and a distance-based neural network model for vehicle speed prediction from two perspectives. The results showed that the predicted speed profiles of the two models matched the variation of the actual speed profiles [6]. Li et al. combined the niche immune genetic algorithm-support vector machine (NIGA-SVM) prediction algorithm and genetic algorithm-support vector machine (GA-SVM) prediction algorithm, which can greatly improve the accuracy and timeliness of vehicle speed prediction [4]. These models have had great success in both theoretical analysis and application. The models of shallow learning are limited by finite samples and computational units. When it comes to representing complex functions, they are not competent.
The purpose of deep learning is the same as shallow learning. The difference is that deep learning enables complex function approximation by building a deep nonlinear network structure. Commonly applied deep learning methods include the convolutional neural network (CNN) and the long short-term memory network (LSTM). Song et al. constructed a CNN-based traffic prediction model utilizing time data and speed data collected on the main roads in Seoul, Korea, and verified that the prediction performance of the model was higher than that of the other two multi-layer perceptron models [20]. Han et al. combined a one-dimensional convolutional neural network with a bidirectional long short-term memory network to predict vehicle speed utilizing the information provided by V2V and V2I communication, and the model proposed significantly improved the prediction accuracy of short-term speed prediction compared with traditional shallow learning [21]. Both CNN and LSTM have achieved excellent prediction results in multiple fields. In terms of extracting data features, CNN is a master of extracting regional features and LSTM does well in extracting temporal features.
The data-based speed prediction method can theoretically fit a highly nonlinear mapping. However, low-value data can cause the model to learn wrong mappings. Therefore, it is important to collect diverse experimental data and explore the deep value of the data.
In this context, a dual-input CNN model (DICNN) is proposed in this paper. DICNN can extract features from vehicle signal sequences and driver intent to predict future speed changes. Firstly, road experiments are designed to collect rich vehicle data for a parallel plug-in hybrid electrical vehicle (PHEV). Secondly, the performance of DICNN is discussed compared with the other three prediction models. Finally, the prediction models are combined with the energy management strategy to further verify the relationship between the accuracy of prediction and the effect of energy saving.
The main innovations and contributions of this work are concluded as follows:
(1)
Deep learning models are used to build the speed prediction model. The deep architecture has more tunable parameters, which can lead to better prediction results.
(2)
The training set is constructed based on rich vehicle data obtained from road experiments. The various vehicle signal sequences provide a more detailed picture of the current operating status of the vehicle.
(3)
The second type of input for the model characterizes the pedal signal. As a direct representation of driver intent in vehicle signals, the second type of input enables the prediction model to more sensitively capture trends in vehicle speed changes.
The paper is organized as follows. Section 2 describes the road experiments. Section 3 introduces the structure of CNN and DICNN. Section 4 presents the energy management strategy applied in this paper for vehicle simulation validation and introduces the method combined with vehicle speed prediction. Section 5 presents the model performance and simulation results. Finally, Section 6 concludes the paper.

2. Road Experiments

Road experiments are designed to collect vehicle traveling data. The CAN bus signals of the vehicle are obtained from the devices of data collection. From all signals, 10 vehicle signals are selected and converted into a format that meets the input requirements of DICNN. The predicted vehicle speed is output based on the model. The whole flow is shown in Figure 1.

2.1. Vehicle Model

The research object of this paper is a parallel PHEV. According to the placement of the electrification component, the vehicle belongs to the P2 type, i.e., the motor is placed behind the clutch 0 and before the transmission [22]. The vehicle configuration is shown by Figure 2. When the clutch 0 is disengaged, the motor drives the vehicle alone, avoiding the engine operating in the inefficiency area and allowing for brake energy recovery. When the clutch 0 is engaged, the engine and motor work together [23]. The switching conditions between modes are mainly determined by the overall vehicle demand power and the state of charge (SOC) [24]. Vehicle and component parameters are shown by Table 1.

2.2. Experiment Method

We apply the on-board recorder to collect vehicle signals. In order to evaluate the data quality of the on-board recorder and provide an accurate basis for processing, VBOX is applied to evaluate and calibrate the key information such as vehicle speed and acceleration. The hardware scheme is shown in Figure 3. The VBOX with a GPS antenna and an acceleration sensor sends the collected key signals such as latitude, longitude, speed and acceleration to the CANalyzer and the CANDTU as CAN signals. CANalyzer is a versatile analysis tool for network and distributed system development [25]. It enables easy observation and analysis of data transmission [26]. CANDTU records packet information in the CAN bus for an extended period of time, for subsequent data processing. Body CAN and powertrain CAN (PT CAN) are displayed and recorded in real time by the upper computer.
In order to reduce the stress on data storage, computation and signal filtering, the signal list was improved as shown in Table 2. The VBOX-related signals were added. The signals related to the energy consumption calculation of each assembly and component were retained. ABS vehicle speed, VCU vehicle speed, VBOX vehicle speed and VBOX acceleration were adopted for redundant calculations to improve speed and acceleration acquisition accuracy.

2.3. Data Processing

Due to the different signal frequencies, noise, abnormal data, etc., the collected data need to be processed first. The basic steps include data synchronization, filtering, etc. Due to the different sampling frequencies, the different signals are not sampled at strictly the same time. The time stamp of the VCU vehicle speed is used as the reference to normalize the sampled time of other signals.
The vehicle system is a continuous system. The derivatives of vehicle signals such as vehicle speed, for example, are still physically significant and should not have abrupt changes. The Hermite interpolation polynomial ensures that the polynomial has equal function values and derivative values at the interpolation nodes. For long time sequences, the degree of the polynomial obtained by Hermite interpolation is high, and there is also the Runge phenomenon. Therefore, the segmented cubic Hermite interpolation is adopted in this paper, and its formula is shown as follows.
I h ( x ) = x x k + 1 x k x k + 1 2 1 + 2 x x k x k + 1 x k f k + x x k x k + 1 x k 2 1 + 2 x x k 1 x k x k 1 +                             x x k + 1 x k x k + 1 2 x x k f k + x x k x k + 1 x k 2 x x k + 1 f k + 1
where I h ( x ) is the cubic polynomial on the interval of [ x k , x k + 1 ] ; f k is the function value at the node k ; and f k is the derivative value of the node k .
In the process of data processing, some data are beyond a reasonable range, such as the acceleration calculated from the vehicle speed. They generally behave as non-operational achievable high frequency jitter. Therefore, a recursive mean filter is adopted for the data process. Its filtering rule is that low frequency signals can pass normally, while high frequency signals exceeding the set frequency threshold are blocked and attenuated. The pseudo code of the recursive mean filter is shown in the Algorithm 1.
Algorithm 1. Recursive mean filter
Data: origin signal vector S , vector length L , window length N
Result: filtered signal vector S
for  i     1 to L  do
2  if  i + N 1 > L  then break
3  Else then S [ i ]   mean( S [ i , ... , N + i 1 ] )
4  end
5 end

3. Speed Prediction Model

3.1. CNN

In 1962, Hubel and Wiesel discovered that the unique network structure of neurons used for local sensitivity and orientation selection in the cat cerebral cortex could effectively reduce the complexity of feedback neural networks, and then proposed CNN. In recent times, CNN has become a key focus for research in many scientific fields. Its layout is closer to that of an actual biological neural network, and weight sharing reduces the complexity of the network. In particular, images with multidimensional inputs can be fed directly to the network, which avoids the complex data reconstruction during feature extraction. Its structure consists of an input layer, an intermediate layer and an output layer, where the intermediate layer includes a convolutional layer, pooling layer, activation function and batch normalization layer.

3.1.1. Convolution Layer

The role of the convolution layer is compression and purification in order to enhance input signals and reduce noise. The equation for the convolution layer is described by
z ( l ) = w ( l ) x ( l 1 ) + b ( l )
where l is the number of the layer; w is a convolution kernel; x is the activity value; b is a learnable bias; and z is the output of convolution layer. The neurons between the convolutional layers are locally connected and share the weights [27]. The characteristics greatly speed up the training and decrease the number of weights and biases. Local connectivity means that the neurons in the convolution layer are connected to the neurons in the next layer only in a local area, which form a locally connected network.

3.1.2. Batch Normalization Layer

To speed up the training of convolutional neural networks and reduce the sensitivity to network initialization, a batch normalization layer is employed, which normalizes each input channel in a small batch [28]. The strategy of batch normalization is to first compute the sample mean and sample variance, and then perform normalization, translation and scaling on the sample data. The equations are described by Equations (3)–(6).
μ B = 1 m i = 1 m x i
σ B 2 = 1 m i = 1 m ( x i μ B ) 2
x ^ i = x i μ B σ B 2 + ε
B N γ , β ( x i ) = γ x ^ i + β
where x is the sample; m is the number of elements in the sample; μ B is the sample mean; σ B 2 is the sample variance; ε is an infinitesimal positive number; and γ and β are training parameters.

3.1.3. Activation Layer

The activation layer is responsible for activating the features extracted from the convolution layer. Since the convolution operation is a linear transformation, it is necessary to introduce an activation layer belonging to a nonlinear function for nonlinear mapping. The ReLU and the Softsign functions are applied. The ReLU function performs a threshold operation to set any input less than zero to zero [29]. The Softsign function has the flatr curve and slow decreasing derivatives for more efficient learning [30]. Their calculation equations are shown by Equations (7) and (8), respectively.
R e L U ( x ) = x x 0 0 x < 0
S o f t s i g n ( x ) = x 1 + x

3.1.4. Pooling Layer

The pooling layer, also known as the subsampling layer, serves to perform feature selection and reduces the number of features. The pooling layer itself does not require any learning, but it reduces the number of parameters to be learned in the subsequent layers and also helps to reduce overfitting. The pooling function adopted in this paper is average pooling. For a region, the average pooling calculates the mean of all neurons in this region as the representation of this region, as shown in Equation (9).
y m , n d = 1 R m , n d i R m , n d x i
where R m , n d is the selected region; x i is the value in R m , n d ; and y m , n d is the representation of R m , n d .

3.1.5. Fully Connected Layer

Each node in the fully connected layer is connected to all nodes in the previous layer, which combines the features extracted from the previous layer. For the vehicle speed prediction model belonging to the regression model, the output size of the fully connected layer is set to 1 which is the future speed of the vehicle. In this paper, five models with different parameters are trained to predict the future speed from 1 to 5 s in turn.

3.2. The Architecture of DICNN

The architecture of DICNN is shown in Figure 4. DICNN has two types of input and a total of four channels.
The first type of input consists of two channels. Each channel contains the 10*10 fake images. The first type of input reflects the characteristics of multiple vehicle signals at short-term adjacent moments. Its first channel adopts a fake image composed of 10 s timing vectors of 10 vehicle signals. The signals selected in this paper are VCU vehicle speed, VBOX vehicle speed, longitudinal acceleration, motor speed, motor torque, DC bus voltage, DC bus current, state of charge, BMS voltage and BMS current. The second channel adopts the horizontal gradient map of the fake image. The gradient map can describe the local intensity gradient or image edge well, and is a common method in image feature extraction. It is obtained by performing gradient operation on the image. For the fake image, the gradient in the horizontal direction has practical physical meaning. Extracting the characteristics of the signal gradient changes helps to provide a reference for future vehicle speed changes. Therefore, the horizontal gradient map of the first channel is used as the input of the second channel.
The second type of input also consists of two channels, each containing a 1 × 10 vector to characterize the driver’s driving intentions. The driver’s intention is directly represented in the vehicle signal by the drive pedal opening and brake pedal opening. The drive pedal on is scaled from 0 to 1 and the brake pedal on is scaled from −1 to 0. The drive pedal opening and brake pedal opening are combined into one vehicle signal named pedal opening. The temporal pedal openings and their temporal accumulation values are used as the two channels of the second type of input to extract driving intentions and predict speed changes. The format of the input to DICNN is shown by
P 1 i , j , 1 = S i j         P 1 i , j , 2 = S i j S i 1 j Δ t P 2 i , 1 , 1 = γ i P 2 i , 1 , 2 = m = 1 i γ m                 ( i I s , j [ 1 , 2 , ... , 10 ] )
where P 1 and P 2 are the first type and the second type of the input of DICNN, respectively; i is the number of sampling points; S j represents the type of signal, from 1 to 10, in order of VCU vehicle speed, VBOX vehicle speed, longitudinal acceleration, motor speed, motor torque, DC bus voltage, DC bus current, state of charge, BMS voltage and BMS current, respectively; Δ t is the time interval between two adjacent sampling points; and γ is the pedal opening. Data set allocations are established: 70% of the data sets are allocated for training, 15% for validation, and the remaining 15% for testing.

4. Energy Management Strategy

PHEV improves the fuel economy by applying both fuel and electricity. In order to meet the vehicle’s power requirements while optimizing the vehicle economy, the control strategy needs to adjust the power distribution between the engine and electrical energy at any given moment according to the real-time road conditions. The development of a reasonable energy management strategy is the basis for achieving better economy for PHEV. This section analyzes the calculation principle of the equivalent fuel consumption minimization strategy (ECMS) and combines it with vehicle speed prediction for optimization.

4.1. Fundamentals of ECMS

ECMS is derived from Pontryagin’s minimum principle (PMP) [31]. Combined with empirical engineering knowledge, it has become a semi-analytic approach to optimal energy management [32]. This principle allows to determine the optimal control decision based on the boundary conditions of the chosen mathematical model in the case of a continuously running system. Further, the optimal trajectory of the state variables of the system and the optimal covariates are solved. When PMP is applied to the energy management of PHEV, the Hamiltonian equation is expressed by
H ( t ) = P f u e l ( t ) + s · P e l e c ( t )
where H is the power cost; s is the co-state variable which is equivalent to a weight; P e l e c is the electrochemical power of the battery; and P f u e l is the fuel consumption power, which has an energy conversion relationship with the engine power, as shown in the Equation (12).
P f u e l ( t ) = P e ( t ) η e ( t )
where P e is the engine power; and η e is the engine efficiency. Similarly, there is an energy conversion between the electrochemical power and the electrical power of the battery. When the battery is discharged, the process of converting the chemical energy into electrical energy generates energy loss as shown in Equation (13).
P e l e c ( t ) = P b ( t ) η b ( t )
where P b is the battery power; and η b is the battery efficiency. Substituting Equation (12) and Equation (13) into Equation (11) yields Equation (14).
H ( t ) = P e ( t ) η e ( t ) + s · P b ( t ) η b ( t )
The control objective is to minimize the equivalent fuel consumption at each moment. Therefore, the dependent variable of the Hamiltonian function should be transformed into the instantaneous equivalent fuel consumption. By dividing both sides of Equation (14) by the calorific value of the fuel at the same time, Equation (15) is obtained.
d m e q d t = P e ( t ) η e ( t ) · Q f h + s · P b ( t ) η b ( t ) · Q f h = d m f u e l d t + s · d m e l e c d t
where d m e q d t is the instantaneous total equivalent fuel consumption rate; d m f u e l d t and s · d m e l e c d t are the fuel consumption rate of the engine and the equivalent fuel consumption rate of the battery, respectively. The different types of energy are unified into the same form for calculation. The ratio of engine and battery power that minimizes the instantaneous total equivalent fuel consumption rate is chosen as the current optimal solution.

4.2. ECMS with Speed Prediction

The optimization objective of the conventional ECMS is to minimize the equivalent fuel consumption at the current moment. The optimization objective of the ECMS with speed prediction is to minimize the sum of the fuel consumption of the engine and the equivalent fuel consumption of the motor at each moment in the predicted time domain. At each sampling moment, an open-loop optimization problem in a finite time domain is solved based on the obtained vehicle information. The first element of the obtained control sequence is applied to the controlled object. At the next sampling moment, the above process is repeated to complete the control of the energy management strategy.
In this paper, a dynamic programming (DP) algorithm is adopted to solve the optimization problem. The basic idea of the dynamic programming algorithm is to transform a multi-stage problem into multiple sub-problems of the same type, and gradually recursively search for optimization by solving the optimal solution of the sub-problems [33]. The algorithm is based on the Behrman optimization principle. In any state of a multi-stage globally optimal control strategy, regardless of its past states and decisions, the remaining many decisions must constitute an optimal sub-strategy. In short, any part of the sub-strategies in the optimal strategy must be optimal.
In this paper, the optimization task is described in terms of travel time. The sampling interval is set to 1 s, i.e., T s = 1 . The ratio of the total duration of the driving cycle to the sampling interval is taken as the phase of the optimization problem, i.e., N = T / T s . The state variable, control variable and parameters are set as described in Equation (16).
x = S O C u = λ p = [ v , a , g e a r ]   T
where x is the state variable; u is the control variable; p is the parameters; S O C is the state of charge of the battery; λ is the ratio of motor power to vehicle demand power; and v , a and g e a r are the speed, acceleration and gear number, respectively, of the vehicle. The state variable is updated as shown in Equation (17).
S O C ( k + 1 ) = S O C ( k ) I b ( u , p ) · T s 3600 C b
where k is the current stage; I b is the battery current; and C b is the battery capacity. The cost function is shown by Equation (14). At this point, a complete energy management strategy is constructed.

5. Validation

In this paper, the prediction effectiveness of DICNN is examined from both performance evaluation and vehicle simulation in turn. Three other speed prediction methods are selected for comparison and validation, which are Markov chain combined with Monte Carlo (MCMC), support vector machine (SVM) and single input CNN (SICNN). Both MCMC and SVM only use the vehicle speed with a history of 10 s as the model input. SICNN only usea the first type of input of DICNN as model input. The driving cycle adopted for validation is the worldwide harmonized light vehicles test cycle (WLTC).

5.1. Preformance Evaluation

Figure 5 shows the speed prediction results of four methods for WLTC. Figure 6, Figure 7, Figure 8 and Figure 9 show the prediction error of MCMC, SVM, SICNN and DICNN, respectively. Figure 10 and Figure 11 show the prediction error for the next 1 s and the next 5 s. Figure 12 shows the box plot of prediction error. Figure 13, Table 3, Table 4, Table 5 and Table 6 show the performances of four methods.
The black line in Figure 5 is the vehicle speed of WLTC. Each colored line is the prediction result of 5 s after the moment corresponding to the beginning of the line. Figure 5 shows that yellow lines have the widest distribution, which represents that MCMC has the largest prediction error. The prediction results of SVM are better than MCMC, but there are larger prediction errors at high speed. The coverage of SICNN and DICNN is similar. However, more deviations in the lines of SICNN can be observed in some moments when the acceleration and deceleration processes switch. As can be seen from the partial enlargement of Figure 5, MCMC has a strong randomness in the predicted speed change due to the randomness of the process of moving from one state to another in the state space. The detail view of SVM has more regular prediction results compared with MCMC because SVM belongs to supervised learning. SICNN can grasp the acceleration variation of the vehicle more precisely after utilizing multiple vehicle signals. In combination with pedal opening, DICNN is more sensitive to speed changes and obtains smoother prediction curves than SICNN. The detail view of DICNN shows that the future state of the vehicle can be predicted before the vehicle state changes, which indicates that DICNN can find a more exact mapping relationship between input and output.
Figure 6, Figure 7, Figure 8 and Figure 9 show the prediction error and predicted vehicle speed of MCMC, SVM, SICNN and DICNN in turn. Figure 10 and Figure 11 compare all methods for the same prediction horizon. Firstly, the prediction errors of different time horizons are compared. It can be seen that the coverage of the blue curve increases for all methods as the prediction horizon increases. This represents a gradual increase in prediction error and a significant decrease in prediction accuracy. Comparing the high-speed prediction with the low-speed prediction, the former predicts better results than the latter for all methods and all horizons. The results arise from the fact that low-speed conditions are mostly caused by environmental constraints which lead to highly complex calculations and difficulties for prediction. Comparing the different methods, the prediction of MCMC is the worst among all methods, followed by SVM. There is no significant difference between SICNN and DICNN observed only from the figures.
Box plots are used to reflect the characteristics of the data distribution. Figure 12 shows the box plots for different prediction methods and horizons. On each box, the central mark indicates the median, and the bottom and top edges of the box indicate the 25th and 75th percentiles, respectively. The whiskers extend to the most extreme data points not considered outliers, and the outliers are plotted individually using the ‘+’ symbol. As the prediction horizon increases, both the bottom and top edges of the box gradually move away from 0 km/h, and the outliers gradually decrease with greater coverage between the bottom and top edges. The prediction errors of SICNN have more negative values, which means that SICNN is inclined to predict a high probability of braking or decelerating actions of the vehicle in the future.
Table 3, Table 4, Table 5 and Table 6 show the performances of the different methods and horizons. Figure 13 shows the results graphically. Root mean square error (RMSE), mean absolute error (MAE), maximum absolute error (ME) and R-square (R2) are popular evaluation indicators for the prediction results. When the predicted speed is close to the actual speed, RMSE, MAE, ME and R2 are close to zero. Conversely, R2 close to 1 indicates an excellent match. It can be precisely analyzed that DICNN outperforms the other three speed prediction methods in terms of MAE, RMSE and R2 at all horizons except the third second ME.

5.2. Simulation

Simulation validation is performed under the Matlab/Simulink platform. This paper establishes a rule-based energy management strategy (RB), a DP-based energy management strategy and ECMS-based energy management strategies adopting the vehicle speed prediction methods of MCMC, SVM, SICNN and DICNN. Figure 14 and Table 7 show the equivalent fuel consumption. Figure 15 shows the changes of fuel consumption. Figure 16 shows the changes of SOC. Figure 17 shows the distribution area of the engine operating points. Figure 18 shows the distribution area of motor operating points. Figure 19 shows the number of engine operating points in each efficiency range. Figure 20 shows the number of motor operating points in each efficiency range.
As can be seen in Figure 14 and Table 7, as a criterion of optimization effectiveness, DP has the best economy with an equivalent fuel consumption of only 3.725 L/100 km. DICNN-based ECMS, which has the best speed prediction capability, achieved the best ECMS-based optimization results. The equivalent fuel consumption of the DICNN-based ECMS is 3.907 L/100 km, which is 4.89% higher than DP. As the accuracy of speed prediction decreases, the number of control decisions that do not fit future conditions increases, resulting in an increase in equivalent fuel consumption. Since the rule-based energy management strategy does not have the ability to dynamically adjust the control strategy, the worst economy with an equivalent fuel consumption of 4.435 L/100 km is obtained.
The changes of fuel consumption in Figure 15 show that the DP starts the engine frequently to keep the fuel consumption slowly rising. DICNN-based ECMS, SICNN-based ECMS and SVM-based ECMS approximately maintain the same operating period and fuel consumption. The rule-based energy management strategy and MCMC-based ECMS adopt a high engine power and consume more fuel than other methods. Combined with Figure 16, results show that DP often starts the engine to assist the motor in sharing the vehicle demand power. Therefore, its SOC changes are also relatively flat and nearly linearly decreasing. Other methods use the charge depleting mode in the first half of the simulation and the charge sustaining mode in the second half because there is no global information about the driving cycle. DICNN-based ECMS, SICNN-based ECMS and SVM-based ECMS have a gentle curve and keep the motor running to take the main demand power at high speed conditions. The rule-based energy management strategy and MCMC-based ECMS has obvious upper and lower limits of the SOC fluctuation range. Therefore, at high speed conditions, fuel consumption increases sharply to ensure that SOC does not drop.
In order to further analyze the working condition of components, distributions of engine operating points and motor operating points are analyzed. From Figure 17 and Figure 18, points represent the operating points of components and colors represent the efficiency of components. DP provides the theoretically optimal control strategy and does not take into account the start/stop process of the engine. Most of engine operating points of DP are distributed in the high efficiency area of the engine. The rest of the methods contain the engine starting process and the pedal-pressing process, so the operating points are distributed in both high efficiency and low efficiency areas. ECMS, combining different speed prediction models, have similar distributions of operating points. The operating points of the rule-based energy management strategy are less distributed in both the high efficiency and low efficiency areas. With regard to the motor, DP has the most concentrated distribution of operating points and RB has the most dispersed distribution of operating points. The distributions of ECMS-based energy management strategies are all relatively similar.
Figure 19 and Figure 20 further show the number of operating points of the engine and motor distributed in different efficiency ranges. For the engine, the operating points of DP are almost entirely distributed in the area above 30% efficiency. In the ECMS methods, DICNN has the largest number of operating points in the high efficiency range and the lowest number of operating points in the low efficiency range, which is exactly the opposite of RB. DICNN-based ECMS, SICNN-based ECMS, SVM-based ECMS and MCMC-based ECMS are distributed in similar numbers in each range, where DICNN-based ECMS is closest to DP. For the motor, the distribution of operating points of DP is poor because DP focuses on adjusting the position of operating points of the engine in the high efficiency area. For the other methods, the distributions do not differ significantly. However, DICNN-based ECMS obtains the largest number of distributions in the 95–100% efficiency range by a narrow margin.

6. Conclusions

In conclusion, we propose a novel CNN-based neural network architecture named DICNN. With two inputs and four channels, DICNN can predict the speed change in the next 5 s by extracting the temporal information of 10 vehicle signals and the driver’s intention. It is compared with representative methods in stochastic prediction, machine learning and deep learning. The results show that DICNN outperforms other methods in terms of performance evaluation metrics including RMSE, MAE, ME and R2. We also construct ECMS energy management strategies with speed prediction for simulation and analysis. With the increase of prediction accuracy of vehicle speed, the difference between ECMS and DP can be reduced from 10.87% to 4.89%. More operating points of the power components are distributed in the high efficiency area to achieve lower equivalent fuel consumption.
In the future, we will explore the impact of the type and number of vehicle signals on the prediction effect. In addition, more deep-learning-based models will be combined to further extract features from the training data and improve the accuracy of prediction.

Author Contributions

Conceptualization, J.X. and L.C.; methodology, J.X. and C.G.; software, J.X. and S.P.; validation, J.X. and L.C.; writing—original draft preparation, J.X.; writing—review and editing, C.G. and Z.H.; supervision, L.C. and Z.H.; funding acquisition, C.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Planning Project in Yibin (2020GY001) and the Science and Technology Planning Project in Tianjin (20YFZCGX00770).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nie, H.; Cui, N.X.; Du, Y.; Wang, M. An Improved Velocity Forecasts Method Considering the Speed and Following Distance of Two Vehicles. In Proceedings of the 39th Chinese Control Conference (CCC), Shenyang, China, 27–29 July 2020; pp. 5517–5522. [Google Scholar]
  2. Jiang, B.N.; Fei, Y.S. Vehicle Speed Prediction by Two-Level Data Driven Models in Vehicular Networks. IEEE Trans. Intell. Transp. Syst. 2017, 18, 1793–1801. [Google Scholar] [CrossRef]
  3. Li, Y.F.; Chen, M.N.; Zhao, W.Z. Investigating long-term vehicle speed prediction based on BP-LSTM algorithms. IET Intell. Transp. Syst. 2019, 13, 1281–1290. [Google Scholar] [CrossRef]
  4. Li, Y.F.; Chen, M.N.; Lu, X.D.; Zhao, W.Z. Research on optimized GA-SVM vehicle speed prediction model based on driver-vehicle-road-traffic system. Sci. China-Technol. Sci. 2018, 61, 782–790. [Google Scholar] [CrossRef]
  5. Yeon, K.; Min, K.; Shin, J.; Sunwoo, M.; Han, M. Ego-Vehicle Speed Prediction Using a Long Short-Term Memory Based Recurrent Neural Network. Int. J. Automot. Technol. 2019, 20, 713–722. [Google Scholar] [CrossRef]
  6. Zhang, L.S.; Zhang, J.Y.; Liu, D.Q.; IEEE. Neural Network based Vehicle Speed Prediction for Specific Urban Driving. In Proceedings of the Chinese Automation Congress (CAC), Xi’an, China, 30 November–2 December 2018; pp. 1798–1803. [Google Scholar]
  7. Ghandour, R.; da Cunha, F.H.R.; Victorino, A.; Charara, A.; Lechner, D.; IEEE. A method of vehicle dynamics prediction to anticipate the risk of future accidents: Experimental validation. In 2011 14th International IEEE Conference on Intelligent Transportation Systems; IEEE International Conference on Intelligent Transportation Systems-ITSC; IEEE: New York, NY, USA, 2011; pp. 1692–1697. [Google Scholar]
  8. Ma, Z.Y.; Zhang, Y.Q.; Yang, J. Velocity and normal tyre force estimation for heavy trucks based on vehicle dynamic simulation considering the road slope angle. Veh. Syst. Dyn. 2016, 54, 137–167. [Google Scholar] [CrossRef]
  9. Magliolo, M.; Bacchi, M.; Sacone, S.; Siri, S.; IEEE. Traffic Mean Speed Evaluation in Freeway Systems; IEEE: New York, NY, USA, 2004; pp. 233–238. [Google Scholar]
  10. Bie, Y.W.; Qiu, T.Z.; Zhang, C.; Zhang, C.B. Introducing Weather Factor Modelling into Macro Traffic State Prediction. J. Adv. Transp. 2017, 2017, 4879170. [Google Scholar] [CrossRef] [Green Version]
  11. Zhang, S.N.; Wu, D.A.; Wu, L.; Lu, Y.B.; Peng, J.Y.; Chen, X.Y.; Ye, A.D.; IEEE. A Markov Chain Model with High-Order Hidden Process and Mixture Transition Distribution. In Proceedings of the International Conference on Cloud Computing and Big Data (CLOUDCOM-ASIA), Fuzhou, China, 16–18 December 2013; pp. 509–514. [Google Scholar]
  12. Liu, H.; Chen, C.; Lv, X.W.; Wu, X.; Liu, M. Deterministic wind energy forecasting: A review of intelligent predictors and auxiliary methods. Energy Convers. Manag. 2019, 195, 328–345. [Google Scholar] [CrossRef]
  13. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Shin, J.; Kim, S.; Sunwoo, M.; Han, M.; IEEE. Ego-Vehicle Speed Prediction Using Fuzzy Markov Chain With Speed Constraints. In 2019 30th IEEE Intelligent Vehicles Symposium; IEEE Intelligent Vehicles Symposium; IEEE: New York, NY, USA, 2019; pp. 2106–2112. [Google Scholar]
  15. Zhou, Y.; Ravey, A.; Pera, M.C.; IEEE. A Velocity Prediction Method based on Self-Learning Multi-Step Markov Chain. In 45th Annual Conference of the IEEE Industrial Electronics Society; IEEE Industrial Electronics Society; IEEE: New York, NY, USA, 2019; pp. 2598–2603. [Google Scholar]
  16. Zhang, L.P.; Liu, W.; Qi, B.N. Combined Prediction for Vehicle Speed with Fixed Route. Chin. J. Mech. Eng. 2020, 33, 60. [Google Scholar] [CrossRef]
  17. Li, Y.F.; Zhang, J.; Ren, C.; Lu, X.D. Prediction of vehicle energy consumption on a planned route based on speed features forecasting. IET Intell Transp Syst. 2020, 14, 511–522. [Google Scholar]
  18. Mozaffari, L.; Mozaffari, A.; Azad, N.L. Vehicle speed prediction via a sliding-window time series analysis and an evolutionary least learning machine: A case study on San Francisco urban roads. Eng. Sci. and Technol.—Int. J.-Jestech 2015, 18, 150–162. [Google Scholar] [CrossRef] [Green Version]
  19. He, H.W.; Wang, Y.L.; Han, R.Y.; Han, M.; Bai, Y.F.; Liu, Q.W. An improved MPC-based energy management strategy for hybrid vehicles using V2V and V2I communications. Energy 2021, 225, 120273. [Google Scholar] [CrossRef]
  20. Song, C.; Lee, H.; Kang, C.; Lee, W.; Kim, Y.B.; Cha, S.W.; IEEE. Traffic Speed Prediction under Weekday Using Convolutional Neural Networks Concepts. In 2017 28th IEEE Intelligent Vehicles Symposium; IEEE Intelligent Vehicles Symposium; IEEE: New York, NY, USA, 2017; pp. 1293–1298. [Google Scholar]
  21. Han, S.J.; Zhang, F.Q.; Xi, J.Q.; Ren, Y.F.; Xu, S.H.; IEEE. Short-term Vehicle Speed Prediction Based on Convolutional Bidirectional LSTM Networks. In 2019 IEEE Intelligent Transportation Systems Conference; IEEE International Conference on Intelligent Transportation Systems-ITSC; IEEE: New York, NY, USA, 2019; pp. 4055–4060. [Google Scholar]
  22. Zhang, B.D.; Yang, F.Y.; Teng, L.; Ouyang, M.G.; Guo, K.F.; Li, W.F.; Du, J.Y. Comparative Analysis of Technical Route and Market Development for Light-Duty PHEV in China and the US. Energies 2019, 12, 23. [Google Scholar] [CrossRef]
  23. Zhou, X.Y.; Qin, D.; Hu, J.J. Multi-objective optimization design and performance evaluation for plug-in hybrid electric vehicle powertrains. Appl. Energy 2017, 208, 1608–1625. [Google Scholar] [CrossRef]
  24. Li, X.M.; Liu, H.; Wang, W.D. Mode Integration Algorithm Based Plug-In Hybrid Electric Vehicle Energy Management Strategy Research. In Proceedings of the 8th International Conference on Applied Energy (ICAE), Beijing Inst Technol, Beijing, China, 8–11 October 2016; pp. 2647–2652. [Google Scholar]
  25. Memon, R.A.; Khaskheli, G.B.; Dahani, M.A. Estimation of operating speed on two lane two way roads along N-65 (SIBI -Quetta). Int. J. Civ. Eng. 2012, 10, 25–31. [Google Scholar]
  26. Zhang, X.B.; Xu, L.F.; Li, J.Q.; Ouyang, M.G.; IEEE. Real-time estimation of vehicle mass and road grade based on multi-sensor data fusion. In Proceedings of the 9th IEEE Vehicle Power and Propulsion Conference (VPPC), Beijing, China, 15–18 October 2013; pp. 486–492. [Google Scholar]
  27. Zhang, Q.R.; Zhang, M.; Chen, T.H.; Sun, Z.F.; Ma, Y.Z.; Yu, B. Recent advances in convolutional neural network acceleration. Neurocomputing 2019, 323, 37–51. [Google Scholar] [CrossRef] [Green Version]
  28. Kolarik, M.; Burget, R.; Riha, K. Comparing Normalization Methods for Limited Batch Size Segmentation Neural Networks. In Proceedings of the 43rd International Conference on Telecommunications and Signal Processing (TSP), Electr Network, Milan, Italy, 7–9 July 2020; pp. 677–680. [Google Scholar]
  29. Hu, Z.; Li, Y.P.; Yang, Z.Y.; IEEE. Improving Convolutional Neural Network Using Pseudo Derivative ReLU. In Proceedings of the 5th International Conference on Systems and Informatics (ICSAI), Nanjing, China, 10–12 November 2018; pp. 283–287. [Google Scholar]
  30. Chang, C.H.; Zhang, E.H.; Huang, S.H.; IEEE. Softsign Function Hardware Implementation Using Piecewise Linear Approximation. In Proceedings of the International Symposium on Intelligent Signal Processing and Communication Systems (lSPACS), Taipei, Taiwan, 3–6 December 2019. [Google Scholar]
  31. Xie, Y.; Savvaris, A.; Tsourdos, A. Fuzzy logic based equivalent consumption optimization of a hybrid electric propulsion system for unmanned aerial vehicles. Aerosp. Sci. Technol. 2019, 85, 13–23. [Google Scholar] [CrossRef] [Green Version]
  32. Rezaei, A.; Burl, J.B.; Zhou, B. Estimation of the ECMS Equivalent Factor Bounds for Hybrid Electric Vehicles. IEEE Trans. Control Syst. Technol. 2018, 26, 2198–2205. [Google Scholar] [CrossRef]
  33. Liu, L.Z.; Zhang, B.Z.; Liang, H.Y. Global Optimal Control Strategy of PHEV Based on Dynamic Programming. In Proceedings of the 6th International Conference on Information Science and Control Engineering (ICISCE), Shanghai, China, 20–22 December 2019; pp. 758–762. [Google Scholar]
Figure 1. Flow chart of vehicle speed prediction.
Figure 1. Flow chart of vehicle speed prediction.
Sensors 21 07767 g001
Figure 2. The schematic of the PHEV configuration.
Figure 2. The schematic of the PHEV configuration.
Sensors 21 07767 g002
Figure 3. The schematic of data collection.
Figure 3. The schematic of data collection.
Sensors 21 07767 g003
Figure 4. The architecture of DICNN.
Figure 4. The architecture of DICNN.
Sensors 21 07767 g004
Figure 5. The speed prediction results of four methods for WLTC. The dots (lines) with color are predicted values and the black dots (lines) are actual values.
Figure 5. The speed prediction results of four methods for WLTC. The dots (lines) with color are predicted values and the black dots (lines) are actual values.
Sensors 21 07767 g005
Figure 6. The prediction error (blue) and prediction speed (orange) of MCMC.
Figure 6. The prediction error (blue) and prediction speed (orange) of MCMC.
Sensors 21 07767 g006
Figure 7. The prediction error (blue) and prediction speed (orange) of SVM.
Figure 7. The prediction error (blue) and prediction speed (orange) of SVM.
Sensors 21 07767 g007
Figure 8. The prediction error (blue) and prediction speed (orange) of SICNN.
Figure 8. The prediction error (blue) and prediction speed (orange) of SICNN.
Sensors 21 07767 g008
Figure 9. The prediction error (blue) and prediction speed (orange) of DICNN.
Figure 9. The prediction error (blue) and prediction speed (orange) of DICNN.
Sensors 21 07767 g009
Figure 10. The prediction error (blue) and prediction speed (orange) for the next 1 s.
Figure 10. The prediction error (blue) and prediction speed (orange) for the next 1 s.
Sensors 21 07767 g010
Figure 11. The prediction error (blue) and prediction speed (orange) for the next 5 s.
Figure 11. The prediction error (blue) and prediction speed (orange) for the next 5 s.
Sensors 21 07767 g011
Figure 12. The box plots of the prediction errors.
Figure 12. The box plots of the prediction errors.
Sensors 21 07767 g012
Figure 13. The performances of the four methods.
Figure 13. The performances of the four methods.
Sensors 21 07767 g013
Figure 14. Equivalent fuel consumption per 100 km.
Figure 14. Equivalent fuel consumption per 100 km.
Sensors 21 07767 g014
Figure 15. The changes of fuel consumption under various strategies.
Figure 15. The changes of fuel consumption under various strategies.
Sensors 21 07767 g015
Figure 16. The changes of SOC under various strategies.
Figure 16. The changes of SOC under various strategies.
Sensors 21 07767 g016
Figure 17. The distribution area of engine operating points under various strategies.
Figure 17. The distribution area of engine operating points under various strategies.
Sensors 21 07767 g017
Figure 18. The distribution area of motor operating points under various strategies.
Figure 18. The distribution area of motor operating points under various strategies.
Sensors 21 07767 g018
Figure 19. The number of engine operating points in each efficiency range.
Figure 19. The number of engine operating points in each efficiency range.
Sensors 21 07767 g019
Figure 20. The number of motor operating points in each efficiency range.
Figure 20. The number of motor operating points in each efficiency range.
Sensors 21 07767 g020
Table 1. Vehicle and component parameters in the studied vehicle.
Table 1. Vehicle and component parameters in the studied vehicle.
ComponentParameterValueUnit
VehicleVehicle weight1200kg
Tire radius0.323m
Final drive ratio4.021/
EngineMaximum torque165Nm
Maximum power105@6500kW@rpm
MotorMaximum torque307Nm
Maximum power126@12584kW@rpm
BatteryRated capacity20.8Ah
Rated voltage366V
GearFirst3.527/
Second2.025/
Third1.382/
Fourth1.058/
Fifth0.958/
Table 2. The signal list.
Table 2. The signal list.
KnotSignalSourceFrequency
VCUVCU vehicle speedPT CAN100 Hz
Drive pedal openingPT CAN100 Hz
Brake pedal openingPT CAN100 Hz
ABSABS vehicle speedBody CAN100 Hz
VBOXLatitudeBody CAN100 Hz
LongitudeBody CAN100 Hz
VBOX vehicle speedBody CAN100 Hz
longitudinal accelerationBody CAN100 Hz
MCUMotor speedPT CAN100 Hz
Motor torquePT CAN100 Hz
DC bus voltagePT CAN100 Hz
DC bus currentPT CAN100 Hz
ECUEngine speedPT CAN100 Hz
Engine torquePT CAN100 Hz
BMSState of chargePT CAN10 Hz
BMS voltagePT CAN10 Hz
BMS currentPT CAN10 Hz
Table 3. RMSE performance.
Table 3. RMSE performance.
MethodPrediction Time (s)
1s2s3s4s5s
MCMC4.126.338.259.9611.63
SVM1.812.413.664.717.08
SICNN1.362.744.285.917.49
DICNN0.811.833.074.415.75
Table 4. MAE performance.
Table 4. MAE performance.
MethodPrediction Time (s)
1s2s3s4s5s
MCMC3.004.706.267.598.88
SVM1.481.762.803.475.70
SICNN0.971.963.064.355.55
DICNN0.481.152.022.993.99
Table 5. ME performance.
Table 5. ME performance.
MethodPrediction Time (s)
1s2s3s4s5s
MCMC17.9727.0437.1236.0238.72
SVM6.1512.8614.2721.6729.67
SICNN6.2211.4315.6621.6327.68
DICNN5.129.0314.8520.8725.95
Table 6. R2 performance.
Table 6. R2 performance.
MethodPrediction Time (s)
1s2s3s4s5s
MCMC0.98710.96960.94830.92460.8973
SVM0.99750.99560.98990.98320.9619
SICNN0.99860.99430.98610.97350.9574
DICNN0.99950.99750.99290.98520.9749
Table 7. Simulation results.
Table 7. Simulation results.
MethodFuel
Consumption
(mL)
Final SOC
(%)
Equivalent Fuel Consumption
(L/100 km)
Increased
Equivalent Fuel Consumption
Compared with DP (%)
DP366.7624.053.725/
MICNN558.1925.133.9074.89
SICNN540.2124.474.0027.44
SVM582.1324.964.0378.38
MCMC809.3031.164.13010.87
RB975.5430.384.43519.06
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xing, J.; Chu, L.; Guo, C.; Pu, S.; Hou, Z. Dual-Input and Multi-Channel Convolutional Neural Network Model for Vehicle Speed Prediction. Sensors 2021, 21, 7767. https://doi.org/10.3390/s21227767

AMA Style

Xing J, Chu L, Guo C, Pu S, Hou Z. Dual-Input and Multi-Channel Convolutional Neural Network Model for Vehicle Speed Prediction. Sensors. 2021; 21(22):7767. https://doi.org/10.3390/s21227767

Chicago/Turabian Style

Xing, Jiaming, Liang Chu, Chong Guo, Shilin Pu, and Zhuoran Hou. 2021. "Dual-Input and Multi-Channel Convolutional Neural Network Model for Vehicle Speed Prediction" Sensors 21, no. 22: 7767. https://doi.org/10.3390/s21227767

APA Style

Xing, J., Chu, L., Guo, C., Pu, S., & Hou, Z. (2021). Dual-Input and Multi-Channel Convolutional Neural Network Model for Vehicle Speed Prediction. Sensors, 21(22), 7767. https://doi.org/10.3390/s21227767

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop