Next Article in Journal
Failure Mode and Effects Analysis Considering Consensus and Preferences Interdependence
Next Article in Special Issue
Thermal Environment Prediction for Metro Stations Based on an RVFL Neural Network
Previous Article in Journal
Inverse Properties in Neutrosophic Triplet Loop and Their Application to Cryptography
Previous Article in Special Issue
Modified Convolutional Neural Network Based on Dropout and the Stochastic Gradient Descent Optimizer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Online Energy Management Control for Hybrid Electric Vehicles Based on Neuro-Dynamic Programming

1
Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
2
Shenzhen College of Advanced Technology, University of Chinese Academy of Sciences, Shenzhen 518055, China
3
Jining Institutes of Advanced Technology, Chinese Academy of Sciences, Jining 272000, China
4
School of Mechatronic Engineering and Automation, Shanghai University, Shanghai 200072, China
*
Author to whom correspondence should be addressed.
Algorithms 2018, 11(3), 33; https://doi.org/10.3390/a11030033
Submission received: 9 February 2018 / Revised: 11 March 2018 / Accepted: 13 March 2018 / Published: 19 March 2018
(This article belongs to the Special Issue Advanced Artificial Neural Networks)

Abstract

:
Hybrid electric vehicles are a compromise between traditional vehicles and pure electric vehicles and can be part of the solution to the energy shortage problem. Energy management strategies (EMSs) are highly related to energy utilization in HEVs’ fuel economy. In this research, we have employed a neuro-dynamic programming (NDP) method to simultaneously optimize fuel economy and battery state of charge (SOC). In this NDP method, the critic network is a multi-resolution wavelet neural network based on the Meyer wavelet function, and the action network is a conventional wavelet neural network based on the Morlet function. The weights and parameters of both networks are obtained by an algorithm of backpropagation type. The NDP-based EMS has been applied to a parallel HEV and compared with a previously reported NDP EMS and a stochastic dynamic programing-based method. Simulation results under ADVISOR2002 have shown that the proposed NDP approach achieves better performance than both the methods. These indicate that the proposed NDP EMS, and the CWNN and MRWNN, are effective in approximating a nonlinear system.

1. Introduction

Hybrid electric vehicles (HEVs) are regarded as an energy-saving solution in the vehicle industry. An energy management strategy (EMS) that coordinates the output power from an internal combustion engine (ICE) and an electric motor simultaneously is important for a HEV because EMSs affect fuel economy and battery state of charge (SOC) directly.
Many approaches have been taken to EMS, from effective but vehicle design sensitive rule-based strategies [1] to optimal control choices [2]. The global optimal solutions can be obtained based on a Pontryagin’s minimum principle (PMP) under reasonable assumptions [3]. However, the global optimal solution for a realistic vehicle problem suffers from two disadvantages: requiring complete knowledge of a drive cycle in advance, and the heavy computational burden induced by too large state space and control space. The most famous global optimal control law is based on a Bellman optimality principle [4]. A typical EMS of this kind is the dynamic programming (DP) approach, which is often utilized in two aspects. The first is to evaluate other algorithms; the second is to analyze vehicle control laws, extract implementable rules in controls (e.g., transitioning between different operating modes and gear shift sequences), and improve rule-based methods [5]. We can infer that the DP-based global optimal solution is driving cycle sensitive. To enhance the robustness of an EMS method to the driving cycle sensitive problem, a stochastic dynamic programming (SDP) approach in an infinite-horizon form is proposed to solve a future driving cycle with a statistical Markov chain model [6]. The stochastic approach optimizes the control policy over a broader set of driving patterns: the best policy achieves a minimum of the expected cost, which is an average over all sample paths of the stochastic model. However, the method still has a high computational burden, and relies on the Markov chain model.
With the development of artificial intelligence, some neural network-based approaches have been proposed to relax the driving cycle sensitive problem [7]. In [8], a supervised competitive learning vector quantization neural network was used to predict real-time driving patterns. Chaining neural network, multi-layer perceptron model, and convolutional neural networks were all employed to predict velocity with the help of information provided by vehicle-to-vehicle communication technology, vehicle-to-infrastructure communication technology, and other intelligent transportation technologies [9,10]. A forecasting velocity was used in an equivalent consumption minimization approach to improve the performance of EMS. In [11], a PMP-based method was used to obtain optimal control of different driving cycles. A neural network was trained to learn the optimal SOC curves. Then, with the partial trip information provided by the intelligent transportation system, a reference optimal SOC curve was produced by the neural network and used for a fuzzy logic controller. We can conclude that the studies mentioned above usually train one kind of neural network and then use it online to predict some variables (e.g., vehicle velocity, optimal SOC curve, optimal value function, etc.). These methods are still not a real online EMS. In [12,13], fully data-driven EMSs based on deep reinforcement learning (DRL) approach were proposed. The DRL-based method needs huge learning samples and significant computational resources. Although an online learning architecture was proposed in [13], utilizing a real-time ITS system, it was still not applied in a real vehicle. In [14], a neural dynamic programming (NDP)-based EMS was first proposed. In this NDP, a radial basis function neural network (RBFNN) is adopted both in critic network and action network. However, the RBFNN requires too long a training time for the large network architecture.
This study focuses on developing a near-optimal torque split control strategy for online application potential that will improve the fuel economy and prevent battery degradation without previewing future traffic information. Control using torque instead of power can handle a velocity of zero. This research also investigates a NDP EMS. However, different to [14], the critic network in this EMS utilizes a multi-resolution wavelet neural network (MRWNN), and the action network adopts a conventional wavelet neural network (CWNN). The CWNN action network does not need clustering preprocessing on the input data, which is essential in the RBFNN action network. Meanwhile, MRWNN benefits from less training time for the dilation and the translation in every activation function being determined in advance, and it is less redundant than RBFNN for orthonormal characteristics—that is, each activation function in the hidden layer is different from each other, and there is no overlap. In the critic network, the MRWNN is based on multi-resolution theory, Meyer scaling, and wavelet functions. The proposed NDP EMS is then applied to the simplest torque-coupling parallel HEV to simulate its performance.
In this paper, the preparation of knowledge on the proposed NDP EMS is presented in Section 2. This knowledge includes the structure of the studied parallel hybrid powertrain, the EMS optimal control problem to be solved, the topology and training method of the traditional wavelet neural network, the multi-resolution analysis theory for a function, and the topology and training approach for a MRWNN. After that, neuro-dynamic programming based on a MRWNN and a CWNN, and its implementation procedure, is described in Section 3. In addition, the NDP-based EMS performance of the parallel HEV is shown and compared with that of two other EMS in Section 4. Finally, conclusions about the proposed NDP EMS are given in Section 5.

2. Preparation of Knowledge about the NDP EMS

In this section, we first give a short description of the studied parallel HEV powertrain. Secondly, the optimal EMS problem is demonstrated in Section 2.2, which introduces the HEV modeling, key concepts of the EMS control problem, and instantaneous cost in optimization, which have been described in [6]. We present the preparation of knowledge for the proposed NDP method, which contains a wavelet transformation theory, in Section 2.3; a CWNN in Section 2.4; a multi-resolution analysis theory of a function in Section 2.5; and a multi-resolution wavelet neural network in Section 2.6. A CWNN is based on a wavelet transformation theory and traditional background neural network. The MRWNN in this research is based on a CWNN and multi-resolution analysis for a function.

2.1. Hybrid Powertrain

This research selects a torque-coupling parallel HEV as the objective HEV. Vehicle parameters for simulation are listed in Table 1. Figure 1 shows the parallel hybrid powertrain that has a spark ignition engine, an electric motor, a nickel metal hydride battery pack, an automatic clutch, a gearbox, and a final drive. The electric motor is located between the engine and the transmission. In this configuration, the motor rotor works as the torque coupler. When the clutch is engaged, the engine and motor have the same speed. The advantage of this architecture is the elimination of the mechanical torque coupler, which results in a simple and compact drivetrain. This configuration also benefits from abundant operational modes, where the motor functions as an engine starter, along with an electrical generator for regenerative braking and an engine power assistant. Here, a vehicle control unit directly commands the torque required from both power sources of the vehicle. The powertrain information has been shown in [6].

2.2. Optimal Control Problem

In this section, we demonstrate key concepts for the EMS control problem. For example, state vector, control vector, control constraints, and instant cost function.
In this research, we model the parallel HEV as a discrete-time dynamic system as follows:
x ( k + 1 ) = f ( x ( k ) , u ( k ) ) k = 0 , 1 , , N 1 ,
where k represents the current discrete time step; N is the total sampling time of the whole driving cycle; x ( k ) is the state variables which summarize past information that is relevant for future optimization; u ( k ) is the control variables, which is the decision to be selected at time step k from a given set; f ( . ) is a function according to which the HEV updates. We choose torque demand at the driven wheels from the driver T d e m ( k ) , wheel speed ω w , and battery state of charge S O C ( k ) as state variables (e.g., x ( k ) = ( T d e m ( k ) , ω w ( k ) , S O C ( k ) ) T ) and engine torque T e ( k ) and gear ratio g r ( k ) as control variables (e.g., u ( k ) = ( T e ( k ) , g r ( k ) ) T ).
The optimal control problem for this EMS is to find an optimal control sequence to minimize the fuel economy, and constrain the battery SOC within a reasonable scope during a driving cycle. The optimal control action u * ( k ) can be obtained by solving the following Bellman equation:
{ J * ( x ( k ) ) = min u ( k ) [ L ( x ( k ) , u ( k ) ) + J * ( x ( k + 1 ) ) ] , 0 k < N 1 J * ( x ( N 1 ) ) = min u ( N 1 ) [ L ( x ( N 1 ) , u ( N 1 ) ) ] , k = N 1 ,
where J * ( k ) is the optimal cumulated cost function from time step k to the end of the driving cycle. As for the infinite horizon, Equation (2) will change to the following equation:
J * ( x ( k ) ) = min u ( k ) E { i = k δ i k L ( x ( i ) , u ( i ) ) } = min u ( k ) E { L ( x ( k ) , u ( k ) ) + δ J * ( x ( k + 1 ) ) } ,
where δ ( 0 , 1 ) is a discount factor, which ensures the total cost function convergence and indicates the importance of future cost to current cost; L ( x ( k ) , u ( k ) , w ( k ) ) denotes the cost incurred at time step k .
L ( x ( k ) , u ( k ) ) = L f u e l ( k ) + β L S O C ( k ) ,
where L f u e l ( k ) , and L S O C ( k ) are the fuel consumption and SOC derivation from a desired value from time step k to time step k + 1 , respectively; the weight β provides tunable parameters that meet the SOC constraints.
When solving the above optimal control problem, the following boundary conditions should be satisfied to guarantee that the parallel HEV components will work properly:
{ ω e _ min ω e ( k ) ω e _ max T e _ min ( ω e ( k ) ) T e ( k ) T e _ max ( ω e ( k ) ) T m _ min ( ω m ( k ) , S O C ( k ) ) T m ( k ) T m _ max ( ω m ( k ) , S O C ( k ) ) S O C min S O C ( k ) S O C max .

2.3. Wavelet Transformation Theory

In this section, we will give a brief introduction to a wavelet transformation theory, which is the basis of a CWNN.
If a function ψ ( t ) L 2 ( R ) satisfies the following item:
C ψ = R | ψ ^ ( ω ) | 2 | ω | d ω < ,
where ψ ^ ( ω ) is the Fourier transformation of ψ ( t ) , which is called a wavelet or mother wavelet.
Given a wavelet ψ ( t ) , a family of wavelets { ψ a , b ( t ) } can be derived with different a and b from the following equation:
ψ a , b ( t ) = 1 a ψ ( t b a ) ,
where a > 0 is a dilation factor and b is a translation factor. ψ a , b ( t ) is called the wavelet basis function.
For a function f ( t ) L 2 ( R ) , we can get its wavelet transform W T f ( a , b ) from the following equation:
W T f ( a , b ) = 1 a R f ( t ) ψ * ( t b a ) d t = f ( t ) , ψ a , b ( t ) ,
where ψ * ( t ) is the conjugate function of ψ ( t ) . W T f ( a , b ) is the inner product of f ( t ) and a family of wavelets { ψ a , b ( t ) } .
For a common input x defined in discrete state, and the dilation set at 2, we can get the following wavelets:
ψ m , k ( x ) = 1 2 m ψ ( x k 2 m 2 m ) = 2 m ψ ( 2 m x k ) ,
where m Z and k Z .

2.4. Common Wavelet Neural Network

The wavelet neural network is based on a wavelet transformation theory and neural network. Different from broadly used back propagation neural network (BPNN) and RBFNN, a wavelet neural network, which we call a common wavelet neural network (CWNN), utilizes a wavelet function as the activation function in the hidden layer. Compared with a RBFNN, a CWNN does not need a clustering preprocessing on the input data, and has same approximation ability. A commonly used topology of a CWNN is shown in Figure 2.
In the CWNN, the input of hidden layer is as follows:
h h = i = 1 I ω i h x i , h = 1 , 2 , , H ,
where X = { x 1 , , x i , , x I } is the input vector; I and H are the total neuron number of the input layer, and hidden layer, respectively; and ω i h is the weight factor between the i th input neuron and the h th hidden layer.
The output of the hidden layer is as follows:
h h = f ( h h b h a h ) = f ( i = 1 I ω i h x i b h a h ) ,   h = 1 , 2 , , H ,
where b h is the shift factor of the wavelet basis function; a h is the scaling factor; ω i h is the weight factor between input layer to hidden layer; and f ( . ) = f ( x ) is the activation function in the hidden layer, which is a wavelet function, such as a Morlet function.
The input of the output layer is as follows:
y j = h = 1 H ω h j h h , j = 1 , 2 , , O ,
where ω h j is the weight factor between the h th hidden neuron and the j th output neuron; O is the total neuron number of the output layer; Y = { y 1 , , y j , , y O } is the output vector of the network.
In a CWNN, the weight factors are trained according to a gradient descent algorithm, while the prediction error propagates backwards. The network parameters ( m + 1 ) th training is as follows:
{ ω h j ( m + 1 ) = ω h j ( m ) η ( m + 1 ) E ω h j ( m ) ω i h ( m + 1 ) = ω i h ( m ) η ( m + 1 ) E ω i h ( m ) a h ( m + 1 ) = a h ( m ) η ( m + 1 ) E a h ( m ) b h ( m + 1 ) = b h ( m ) η ( m + 1 ) E b h ( m ) ,
where E = 1 2 j = 1 O e j 2 = 1 2 j = 1 O ( y j y j ) 2 is the prediction error indication of the network; η is the learning rate of the network. Meanwhile, E a ω a h j ( m ) , E a ω a i h ( m ) , E a a h ( m ) and E a b h ( m ) are as follows:
{ E ω h j ( m ) = E e j e j y j y j ω h j = e j h h E ω i h ( m ) = j = 1 O ( E e j e j y j y j h h ) h h h h h h ω i h = j = 1 O ( e j ω h j ) f ( h h ) a h x i E b h ( m ) = j = 1 O ( E e j e j y j y j h h ) h h b h = j = 1 O ( e j ω h j ) ( f ( h h ) a h ) E a h ( m ) = j = 1 O ( E e j e j y j y j h h ) h h a h = j = 1 O ( e j ω h j ) ( f ( h h ) a h 2 ) ( h h b h ) .

2.5. Multi-Resolution Analysis of Functions

The multi-resolution analysis of functions is based on the theory of multi-resolution analysis proposed by Mallat in [15,16].
Let V m ( m Z ) is a closed subspace of a square-integrable space L 2 ( R ) at the resolution 2 m , which is also the sampling interval. The subspace family { V m } has the following features:
(1)
Causality and consistency.
V m + 1 V m V m 1 ,   m Z
(2)
Scaling regularity. When we approximate a function f ( x ) L 2 ( R ) with f m ( x ) = A m f ( x ) at the resolution 2 m , where A m is a projection of f ( x ) in subspace V m . Then the following equation will be satisfied:
f m ( x ) = A m f ( x ) V m f m 1 ( x ) = A m 1 f ( 2 x ) V m 1 ,   m Z .
(3)
Gradual completeness.
m = m = + V m = { 0 } ,   m = m = + V m = L 2 ( R )
(4)
Orthogonal basis existence.
V m = l i n e a r   s p a n { ϕ m k , k Z } ¯ ,
where ϕ ( t ) V 0 is a unique function, called a scaling function. Meanwhile, ϕ m k ( x ) = 2 m ϕ ( 2 m x k ) , m Z and k Z .
When we approximate a function f ( x ) with f m ( x ) = A m f ( x ) at the resolution 2 m , we can conclude the following equation based on Feature (4):
f m ( x ) = A m f ( x ) = k = + a m k ϕ m k ( x ) .
We define W m be the orthogonal space of V m in V m 1 . In other words, V m 1 = V m W m and V m W m . Then,
f m 1 ( x ) = A m 1 f ( x ) = [ A m D m ] f ( x ) = A m f ( x ) D m f ( x ) ,
where D m is a projection of f ( x ) on the space W m . We can say D m f ( x ) is a detail of function f ( x ) at resolution 2 m . Mallat has pointed out that a unique wavelet ψ ( x ) exists, whose dilations and translations constitute a family of functions { ψ m k } , m Z , k Z . These { ψ m k } are the orthonormal basis of space W m . As a result, we get the following equation:
D m f ( x ) = k = + d m k ψ m k ( x ) ,
where d m k is the projection of f ( x ) on the orthonormal basis ψ m k ( x ) . We obtain the following equation:
f m 1 ( x ) = A m f ( x ) D m f ( x ) = k = + a m k ϕ m k ( x ) + k = + d m k ψ m k ( x ) .
In other words, at the resolution 2 L 1 ,
f ( x ) f L 1 ( x ) = k = + a m k ϕ m k ( x ) + k = + d m k ψ m k ( x ) = k = 1 n L a L k ϕ L k ( x ) + k = 1 n L d L k ψ L k ( x ) = f L ( x ) + k = 1 n L d L k ψ L k ( x ) ,
where L is the lowest resolution and n L is the total number of scaling functions. In a real application, we can obtain L according to the following equation:
L = int ( lg ( x max x min ) / lg 2 )   or   L = int ( lg ( x max x min ) / lg 2 ) ± 1 ,
where x max , and x min are the maximum and minimum value of x . Using Equation (19), we will get the following equation at resolution 2 0 :
f ( x ) f 0 ( x ) = f 1 ( x ) + k = 1 2 L 1 n L d 1 k ψ 1 k ( x ) = k = 1 n L a L k ϕ L k ( x ) + m = 1 L k = 1 2 L m n L d m k ψ m k ( x ) .
Equations (19) and (21) are the multi-resolution analysis of functions proposed by Moody in [17].

2.6. Multi-Resolution Wavelet Neural Network

In this research, we proposed a multi-resolution wavelet neural network (MRWNN) based on the theories mentioned in Section 2.4 and Section 2.5. The structure of the proposed MRWNN is shown in Figure 3. Different to RBFNN, the hidden layer in MRWNN contains two types of nodes: wavelet nodes ( ψ -nodes, which are in the green ellipse) and scaling nodes ( ϕ -nodes, which are in the blue ellipse, where one red circle represents the scaling nodes at a same resolution). Compared with RBFNN, the MRWNN does not need a clustering preprocessing for the input data, and enjoys faster convergence and higher approximate accuracy for the orthogonal characteristic of activation function in hidden nodes.
The output of the proposed MRWNN is as follows:
y { k = 1 n L a L k ϕ L k ( H 1 ( k ) ) ,   at   resolution   2 L k = 1 n L a L k ϕ L k ( H 1 ( k ) ) + k = 1 n L d L k ψ L k ( H 2 ( m , k ) ) ,   at   resolution   2 L 1 , m = L k = 1 n L a L k ϕ L k ( H 1 ( k ) ) + m = 1 L k = 1 2 L m n L d m k ψ m k ( H 2 ( m , k ) ) ,   at   resolution   2 0 ,
where H 1 ( k ) = i = 1 N ω i k x i is the input of the k th hidden layer or scaling node; H 2 ( m , k ) = i = 1 N ω i p x i ( p = 2 L m n L + k ) is the input of the p th hidden layer or the ( p n L ) th wavelet node.
In real applications, we usually train the network at resolution 2 L . If the output does not meet our precision requirements, we keep the scaling nodes and add the detail of resolution 2 L at resolution 2 L 1 . We proceed like this until the training precision meets our requirements.
In this MRWNN, the weight factors are trained according to the same gradient descent algorithm as in Equations (13) and (14). The weight factor between the h th hidden node and the output node is as follows:
ω h = { a L h , h = 1 , 2 , , n L d m k , h = 2 L m n L + k .
In this research, we use the Meyer wavelet, which is infinitely differentiable with effective support and defined in the frequency domain. In [18], Valenzuela and de Oliveira give the closed analytical expression for the Meyer scaling function and Meyer wavelet function shown as follows:
ϕ ( t ) = { 2 3 + 4 3 π , t = 0 sin ( 2 π 3 t ) + 4 3 t cos ( 4 π 3 t ) π t 16 π 9 t 3 , t 0 ,
where ϕ ( t ) is the scaling function. The wavelet function is as follows:
ψ ( t ) = ψ 1 ( t ) + ψ 2 ( t ) ,
where
{ ψ 1 ( t ) = 4 3 π ( t 1 2 ) cos [ 2 π 3 ( t 1 2 ) ] 1 π sin [ 4 π 3 ( t 1 2 ) ] ( t 1 2 ) 16 9 ( t 1 2 ) 3 ψ 2 ( t ) = 8 3 π ( t 1 2 ) cos [ 8 π 3 ( t 1 2 ) ] + 1 π sin [ 4 π 3 ( t 1 2 ) ] ( t 1 2 ) 64 9 ( t 1 2 ) 3 .
The Meyer wavelet and scaling functions in time domain are shown in Figure 4.

3. Proposed NDP for Parallel HEVs

Figure 5 is the structure of the proposed NDP method, which contains a critic network, an action network, a correction model for the output of action network, and a time delay from time step k to time step k + 1 . The critic network utilizes a MRWNN presented in Section 2.5. The action network utilizes a CWNN, in which the activation function is a Morlet function. The correction model modifies the output variable T e of action network to fit for Limitation (5). Because the output gear ratio of action network is a continuous variable, we also use the correction model to discretize the continuous gear ratio to the real vehicle gear ratio of the vehicle. The critic network, action network, and implementation procedure of the proposed NDP EMS will be described in this section.

3.1. Critic Network Description

In the NDP method, the critic network is used to get the approximate value J ¯ ( k ) of the optimal value function J * ( k ) , which is shown in Equation (2), and used in dynamic programming. As a result, we get the following equation:
J ¯ ( x ( k ) ) = min u ( k ) [ L ( x ( k ) , u ( k ) ) + J ¯ ( x ( k + 1 ) ) ] ,   0 k < N 1 .
J ¯ ( k ) will be obtained by training the weights of the network according to current state variables x ( k ) and control variables u ( k ) .
In this paper, the critic network is a MRWNN, described in Section 2.5. In this critic network, X c ( k ) = ( T d e m ( k ) , ω w ( k ) , S O C ( k ) ) T is the input vector, and Y c = y c = J is the output vector. The prediction error is defined according to a T D ( λ ) learning approach as follows:
e c ( k ) = L ( x ( k ) , u ( k ) ) + γ J ( k + 1 ) J ( k ) ,
where γ ( 0 , 1 ) is a discount factor; L ( x ( k ) , u ( k ) ) used as a reinforcement signal, is the instantaneous cost, and can be calculated according to Equation (4).
Meanwhile, the prediction error indication of the critic network is defined as follows:
E c ( k ) = 1 2 e c 2 ( k ) .
In the critic network, the weight factors (e.g., ω c h and ω c i h ) are adapted according to Equations (13) and (14).

3.2. Action Network Description

The objective of the action network is to find an action u ¯ ( k ) to approximate the optimal control u * ( k ) shown in Equation (2), which is done by adapting the weights and parameters of the network according to the variation of the output of critic network J ( k ) . In this network, the state vector of the vehicle is the input vector ( X a ( k ) = ( T d e m ( k ) , ω w ( k ) , S O C ( k ) ) T ), and the control vector of the vehicle is the output vector ( Y a ( k ) = ( T e ( k ) , g r ( k ) ) T ).
In this paper, the action network is a common wavelet neural network (CWNN) consisting of three layers of neurons—input, hidden, and output layer—as shown in Section 2.4. The adaptation of the action network is achieved by back-propagating the following prediction error e a :
e a ( k ) = J ( k ) u ( k ) = [ e a 1 e a 2 ] = [ J ( k ) T e ( k ) J ( k ) g r ( k ) ] .
Meanwhile, the prediction error indication of the action network is defined as follows:
E a ( k ) = 1 2 e a T ( k ) e a ( k ) .
In the action network, the weight factors (e.g., ω a i h and ω a h j ) and wavelet function parameters (e.g., a h and b h ) are adapted according to the same gradient descent algorithm in Equations (13) and (14).

3.3. NDP Implementation Procedure

The NDP implementation procedure is presented as follows:
  • Select an appropriate structure for the critic network and action network separately, which means determining the number of hidden layers. In this design, the number of hidden layers of the critic network and action network is 30 and 11, respectively.
  • Initialize the weights and parameters of both networks according to a rule-based EMS. Initialize the output of the critic network, the maximum training error of critic network E c , the maximum of the norm for weights of action network variation ε , and maximum times for training both networks, N c and N a . In this research, the maximum training times for the action network and critic network are all 50.
  • Set the whole iteration time i = 0 and L ( x ( 0 ) , u ( 0 ) ) = 0 .
  • Set i = i + 1 . Calculate the current control vector u ( i ) .
  • Set the iteration time of critic network to m = 0 .
  • Set m = m + 1 . Calculate the error of critic network e c m ( i ) and E c m ( i ) according to Equations (32) and (33). Update the weights and parameters of the critic network according to Equations (13) and (14);
  • When m N c or E c m E c , save the weights and parameters of the critic network as the weights and parameters at time step k , and proceed to step 8. Otherwise, return to step 6.
  • Set the iteration time of action network to m = 0 .
  • Set m = m + 1 . Calculate the error of action network e a m ( i ) and E a m ( i ) according to Equations (34) and (35). Update the weights and parameters of the action network according to Equations (13) and (14);
  • When m N a or E a m E a , save the weights and parameters of the critic network as the weights and parameters at time step k , and proceed to step 11. Otherwise, return to step 9.
  • When the norm of weights for action network variation is less than ε , save the weights of both networks as the weights at the current time step; otherwise, return to step 4.

4. Results and Discussion

In the simulation test of the proposed NDP-based EMS, the detailed model of the parallel HEV is built in ADVISOR2002 (Advanced Vehicles Simulator) to accurately reflect the vehicle performance, which is based on Simulink. The simulation was carried out on a computer with Intel Core 2 Duo 3.10 GHZ CPU and 32G memory. We have compared the proposed NDP method with the NDP method in [14] to test the proposed NDP based on CWNN and MRWNN. We have also compared the NDP EMS with a SDP method proposed in [6], which is an optimal method, in driving cycles of CYC_UDDS, CYC_1015 and a composite driving cycle, which is composed of a CYC_1015, a CYC_UDDS, and a CYC_WVUCITY driving cycle, to test the optimality of the proposed NDP method.
According to Section 2.2, the orientation of the EMS is to maximize the fuel economy and to satisfy SOC constraint. In simulations, we evaluate fuel economy with miles per gallon (mpg) gasoline equivalent. The SOC constraint is to ensure battery SOC [ 0.3 , 0.8 ] . Meanwhile, we evaluate the component efficiency of the parallel HEV with engine average efficiency, average motoring efficiency, and average generating efficiency.
Table 2 shows the fuel economy and HEV component efficiency of the proposed NDP and NDP EMS in [14] in driving cycles of CYC_UDDS. From Table 2, we can see that the proposed NDP improves fuel economy by about 5.4%, average generating efficiency by about 7.6%, and average motoring efficiency by about 1.2% compared with the NDP in [14]. The two NDP methods keep SOC in nearly the same range of 0.55–0.7. These indicate that the proposed NDP with CWNN and MRWNN has better control results in the UDDS driving cycle than the NDP with RBFNN in [14]. In the NDP in [14], they are 25 and 20. This may be due to the CWNN and MRWNN enjoying better approximation and learning ability in handling nonlinear problems than a RBFNN. To further improve the performance of the proposed NDP EMS, we can increase the maximum training time of the action and critic network, or decrease the training error of the two networks.
Table 3, Table 4 and Table 5 show the fuel economy and component efficiency of SDP, and the proposed NDP-based EMS in driving cycles of CYC_UDDS, CYC_1015, and the composite driving cycle. We can conclude that, compared with the SDP in [6], the proposed NDP EMS improves fuel economy by about 25.9%, 6.2%, and 25.6%, respectively, under CYC_UDDS, CYC_1025, and the composite driving cycle. In the three driving cycles, the proposed NDP and SDP all keep SOC in nearly the same range. The proposed NDP EMS also improves the average generating efficiency by about 6.7%, 5.2%, and 4%, respectively, under CYC_UDDS, CYC_1025, and the composite driving cycle. This shows that the proposed NDP is better than the SDP method. There are three reasons for the simulation results. The first is that the state space and action space are all discretized in SDP and successive in NDP, which reduces the error induced by low discretization resolution. The second reason is that the NDP-based method has generalization ability and improves the generating efficiency in braking. Last but not least, the probability transfer matrix of SDP method is built with the UDDS driving cycle, which induces that the SDP is only suitable for driving cycles with the same probability transfer matrix.
Figure 6 shows the torque demand at the input of gear box, engine torque, motor torque, SOC, and gear ratio variation of proposed NDP EMS results in CYC_UDDS driving cycle. We can see that the engine is in a higher torque output mode most of the time, and the motor provides energy in peak and engine starting times. Battery SOC is constrained in the region of 0.55–0.7, which ensures safe battery usage.
Figure 7 shows the engine operation efficiency in NDP EMS. We can see that, in most situations, the engine is in an efficiency region larger than 0.35.
Figure 8 shows the motor operation efficiency in NDP EMS. We can see that, in most situations, the motor is in an efficiency region larger than 0.8. The operation points in the lower efficiency region are distributed uniformly, which coincides with the actual operational points of the motor in real applications.

5. Conclusions

For an HEV, EMS design is critically related to the fuel economy and battery SOC degradation. In this paper, a neuron-dynamic program (NDP) has been designed for the EMS problem of a parallel HEV. In this NDP, a multi-resolution wavelet neural network is used as the critic network to approximate the cumulated value function; a common wavelet neural network is employed to train the optimal action according to current state vector and the output of the action network. A gradient descent algorithm is utilized to train both the neural networks.
ADVISOR2002 simulation results show that the proposed NDP EMS achieves better performance in fuel economy, generating better efficiency and average engine efficiency than the NDP reported before with RBFNN as a critic network and an action network under CYC_UDDS driving cycle. The simulation results also indicate that, compared with the SDP approach in three different driving cycles, the proposed NDP EMS significantly improves fuel economy and average generating efficiency. The NDP method also guarantees battery SOC at a safe range of 0.55–0.7 from the start of one cycle to the end. The performance of the NDP EMS demonstrates the usefulness of the NDP in online applications. The comparison with RBFNN-based NDP EMS confirms the efficiency of the CWNN and MRWNN.

Acknowledgments

This work is supported by the National Key Research and Development Program of China (Grant No. 2016YFD0700602), the Technical Research Foundation of Shenzhen (Grant No. JSGG20141020103523742), the National Natural Science Foundation of China (61573337), and the National Natural Science Foundation of China (61603377).

Author Contributions

Feiyan Qin and Weimin Li conceived the NDP control strategy. Feiyan Qin designed and wrote the whole manuscript. Weimin Li checked the whole manuscript. Guoqing Xu and Yue Hu checked the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, J.; Xu, C.; Wu, C.; Xu, W. Adaptive fuzzy logic control of fuel-cell-battery hybrid systems for electric vehicles. IEEE Trans. Ind. Inform. 2018, 14, 292–300. [Google Scholar] [CrossRef]
  2. Delprat, S.; Lauber, J.; Guerra, T.M.; Rimaux, J. Control of a parallel hybrid powertrain: Optimal control. IEEE Trans. Veh. Technol. 2004, 53, 872–881. [Google Scholar] [CrossRef]
  3. Kim, N.; Cha, S.; Peng, H. Optimal control of hybrid electric vehicles based on Pontryagin’s minimum principle. IEEE Trans. Control Syst. Technol. 2011, 19, 1279–1286. [Google Scholar]
  4. Perez, L.V.; Bossio, G.R.; Moitre, D.; Garcia, G.O. Optimization of power management in an hybrid electric vehicle using dynamic programming. Math. Comput. Simul. 2006, 73, 244–254. [Google Scholar] [CrossRef]
  5. Stephan, U.; Nikolce, M.; Conny, T.; Bernard, B. Optimal energy management and velocity control of hybrid electric vehicles. IEEE Trans. Veh. Technol. 2018, 67, 327–337. [Google Scholar]
  6. Qin, F.; Xu, G.; Hu, Y.; Xu, K.; Li, W. Stochastic optimal control of parallel hybrid electric vehicles. Energies 2017, 10, 214. [Google Scholar] [CrossRef]
  7. Munoz, P.M.; Correa, G.; Gaudiano, M.E.; Fernandez, D. Energy management control design for fuel cell hybrid electric vehicles using neural networks. Int. J. Hydrog. Energy 2017, 42, 28932–28944. [Google Scholar] [CrossRef]
  8. Zhang, Q.; Deng, W.; Li, G. Stochastic control of predictive power management for battery/supercapacitor hybrid energy storage systems of electric vehicles. IEEE Trans. Ind. Inform. 2018. [Google Scholar] [CrossRef]
  9. Zhang, F.; Xi, J.; Langari, R. Real-time energy management strategy based on velocity forecasts using V2V and V2I communications. IEEE Trans. Control Syst. 2017, 18, 416–430. [Google Scholar] [CrossRef]
  10. Song, C.; Lee, H.; Kang, C.; Lee, W.; Kim, Y.B.; Cha, S.W. Traffic speed prediction under weekday using convolutional neural networks concepts. In Proceedings of the IEEE Intelligent Vehicles Symposium, Redondo Beach, CA, USA, 11–14 June 2017; IEEE: New York, NY, USA, 2017; pp. 1293–1298. [Google Scholar]
  11. Tian, H.; Lu, Z.; Huang, Y.; Tian, G. Adaptive fuzzy logic energy management strategy based on reasonable SOC reference curve for online control of Plug-in Hybrid electric city bus. IEEE Trans. Intell. Transp. Syst. 2018. [Google Scholar] [CrossRef]
  12. Qi, X.; Luo, Y.; Wu, G.; Boriboonsomsin, K.; Brath, M.J. Deep reinforcement learning-based vehicle energy efficiency autonomous learning system. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Redondo Beach, CA, USA, 11–14 June 2017. [Google Scholar]
  13. Hu, Y.; Li, W.; Xu, K.; Zahid, T.; Qin, F.; Li, C. Energy management strategy for a hybrid electric vehicle based on deep reinforcement learning. Appl. Sci. 2018, 8, 187. [Google Scholar] [CrossRef]
  14. Li, W.; Xu, G.; Xu, Y. Online learning control for hybrid electric vehicle. Chin. J. Mech. Eng. 2012, 25, 98–106. [Google Scholar] [CrossRef]
  15. Mallat, S.G. Multiresolution approximations and wavelet orthogonal bases of L2(R). Trans. Am. Math. Soc. 1989, 315, 67–87. [Google Scholar]
  16. Mallat, S.G. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 674–693. [Google Scholar] [CrossRef]
  17. Moody, J. Fast Learning in Multi-Resolution Hierarchies; research report; Yale University: New Haven, CT, USA, 1989. [Google Scholar]
  18. Valenzuela, V.V.; Oliveira, H.M. Close expressions for Meyer wavelet and scale function. arXiv, 2015; arXiv:1502.00161. [Google Scholar]
Figure 1. Configuration of the parallel HEV powertrain.
Figure 1. Configuration of the parallel HEV powertrain.
Algorithms 11 00033 g001
Figure 2. Topology of CWNN.
Figure 2. Topology of CWNN.
Algorithms 11 00033 g002
Figure 3. Structure of multi-resolution wavelet neural network (MRWNN).
Figure 3. Structure of multi-resolution wavelet neural network (MRWNN).
Algorithms 11 00033 g003
Figure 4. Meyer wavelet function and scaling function in the time domain.
Figure 4. Meyer wavelet function and scaling function in the time domain.
Algorithms 11 00033 g004
Figure 5. Structure of the proposed NDP method.
Figure 5. Structure of the proposed NDP method.
Algorithms 11 00033 g005
Figure 6. NDP EMS simulation results under CYC_UDDS driving cycle.
Figure 6. NDP EMS simulation results under CYC_UDDS driving cycle.
Algorithms 11 00033 g006
Figure 7. Engine efficiency in NDP EMS under the CYC_UDDS driving cycle.
Figure 7. Engine efficiency in NDP EMS under the CYC_UDDS driving cycle.
Algorithms 11 00033 g007
Figure 8. Motor efficiency in NDP EMS under the CYC_UDDS driving cycle.
Figure 8. Motor efficiency in NDP EMS under the CYC_UDDS driving cycle.
Algorithms 11 00033 g008
Table 1. Parameters of the parallel HEV.
Table 1. Parameters of the parallel HEV.
ItemValue
Final drive gear efficiency (%)0.9296
wheel radius (m)0.275
Coefficient of aerodynamic drag0.25
Vehicle frontal area (m2)1.9
Air density (kg/m3)1.2
Rolling resistance coefficient0.0054
Vehicle mass (kg)1000
Table 2. Control orientation improvement with other NDP EMS.
Table 2. Control orientation improvement with other NDP EMS.
EMSNDP in [14]Proposed NDP
Fuel Economy (mpg)81.686.0
SOC range0.55–0.70.54–0.7
Engine Eff. (%)34.635.0
Motoring Eff. (%)91.686.1
Generating Eff. (%)92.799.7
Table 3. Control orientation improvement with another optimal method under CYC_UDDS.
Table 3. Control orientation improvement with another optimal method under CYC_UDDS.
EMSSDPProposed NDP
Fuel Economy (mpg)68.386.0
SOC range0.54–0.70.54–0.7
Engine Eff. (%)35.735.0
Motoring Eff. (%)91.186.1
Generating Eff. (%)93.499.7
Table 4. Control orientation improvement with another optimal method under CYC_1015.
Table 4. Control orientation improvement with another optimal method under CYC_1015.
EMSSDPProposed NDP
Fuel Economy (mpg)58.261.8
SOC range0.54–0.70.55–0.7
Engine Eff. (%)36.133.5
Motoring Eff. (%)87.088.6
Generating Eff. (%)90.595.2
Table 5. Control orientation improvement with another optimal method under CYC_1015 + CYC_UDDS + CYC_WVUCITY.
Table 5. Control orientation improvement with another optimal method under CYC_1015 + CYC_UDDS + CYC_WVUCITY.
EMSSDPProposed NDP
Fuel Economy (mpg)62.178.0
SOC range0.54–0.70.53–0.7
Engine Eff. (%)36.336.1
Motoring Eff. (%)86.183.7
Generating Eff. (%)91.895.5

Share and Cite

MDPI and ACS Style

Qin, F.; Li, W.; Hu, Y.; Xu, G. An Online Energy Management Control for Hybrid Electric Vehicles Based on Neuro-Dynamic Programming. Algorithms 2018, 11, 33. https://doi.org/10.3390/a11030033

AMA Style

Qin F, Li W, Hu Y, Xu G. An Online Energy Management Control for Hybrid Electric Vehicles Based on Neuro-Dynamic Programming. Algorithms. 2018; 11(3):33. https://doi.org/10.3390/a11030033

Chicago/Turabian Style

Qin, Feiyan, Weimin Li, Yue Hu, and Guoqing Xu. 2018. "An Online Energy Management Control for Hybrid Electric Vehicles Based on Neuro-Dynamic Programming" Algorithms 11, no. 3: 33. https://doi.org/10.3390/a11030033

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop