Next Article in Journal
Impact of Optimum Allocation of Renewable Distributed Generations on Distribution Networks Based on Different Optimization Algorithms
Next Article in Special Issue
A Hybrid Seasonal Mechanism with a Chaotic Cuckoo Search Algorithm with a Support Vector Regression Model for Electric Load Forecasting
Previous Article in Journal
Adaptively Constrained Stochastic Model Predictive Control for the Optimal Dispatch of Microgrid
Previous Article in Special Issue
A High Precision Artificial Neural Networks Model for Short-Term Energy Load Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Belief Network Based Hybrid Model for Building Energy Consumption Prediction

1
School of Information and Electrical Engineering, Shandong Jianzhu University, Jinan 250101, China
2
Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
*
Author to whom correspondence should be addressed.
Energies 2018, 11(1), 242; https://doi.org/10.3390/en11010242
Submission received: 15 December 2017 / Revised: 16 January 2018 / Accepted: 16 January 2018 / Published: 19 January 2018
(This article belongs to the Special Issue Short-Term Load Forecasting by Artificial Intelligent Technologies)

Abstract

:
To enhance the prediction performance for building energy consumption, this paper presents a modified deep belief network (DBN) based hybrid model. The proposed hybrid model combines the outputs from the DBN model with the energy-consuming pattern to yield the final prediction results. The energy-consuming pattern in this study represents the periodicity property of building energy consumption and can be extracted from the observed historical energy consumption data. The residual data generated by removing the energy-consuming pattern from the original data are utilized to train the modified DBN model. The training of the modified DBN includes two steps, the first one of which adopts the contrastive divergence (CD) algorithm to optimize the hidden parameters in a pre-train way, while the second one determines the output weighting vector by the least squares method. The proposed hybrid model is applied to two kinds of building energy consumption data sets that have different energy-consuming patterns (daily-periodicity and weekly-periodicity). In order to examine the advantages of the proposed model, four popular artificial intelligence methods—the backward propagation neural network (BPNN), the generalized radial basis function neural network (GRBFNN), the extreme learning machine (ELM), and the support vector regressor (SVR) are chosen as the comparative approaches. Experimental results demonstrate that the proposed DBN based hybrid model has the best performance compared with the comparative techniques. Another thing to be mentioned is that all the predictors constructed by utilizing the energy-consuming patterns perform better than those designed only by the original data. This verifies the usefulness of the incorporation of the energy-consuming patterns. The proposed approach can also be extended and applied to some other similar prediction problems that have periodicity patterns, e.g., the traffic flow forecasting and the electricity consumption prediction.

1. Introduction

With the growth of population and the development of economy, more and more energy is consumed in the residential and office buildings. Building energy conservation plays an important role in the sustainable development of economy. However, some ubiquitous issues, e.g., the poor building management and the unreasonable task scheduling, are impeding the efficiency of the energy conservation policies. To improve the building management and the task scheduling of building equipment, one way is to provide accurate prediction of the building energy consumption.
Nowadays, numerous data-driven artificial intelligence approaches have been proposed for building energy consumption prediction. In [1], the random forest and the artificial neural network (ANN) were applied to the high-resolution prediction of building energy consumption, and their experimental results demonstrated that both models have comparable predictive power. In [2], a hybrid model combining different machine learning algorithms was presented for optimizing energy consumption of residential buildings under the consideration of both continuous and discrete parameters of energy. In [3], the extreme learning machine (ELM) was used to estimate the building energy consumption, and simulation results indicated that the ELM performed better than the genetic programming (GP) and the ANN. In [4], the clusterwise regression method, also known as the latent class regression, which integrates clustering and regression, was utilized to the accurate and stable prediction of building energy consumption data. In [5], the feasibility and applicability of support vector machine (SVM) for building energy consumption prediction were examined in a tropical region. Moreover, in [6,7,8,9], a variation of SVM, the support vector regressor (SVR) was proposed for forecasting the building energy consumption and the electric load. Furthermore, in [10], a novel machine learning model was constructed for estimating the commercial building energy consumption.
The historical building energy consumption data have high levels of uncertainties and randomness due to the influence of the human distribution, the thermal environment, the weather conditions and the working hours in buildings. Thus, there still exists the need to improve the prediction precision for this application. To realize this objective, we can take two strategies into account. The first strategy is to adopt the more powerful modeling methods to learn the information hidden in the historical data, while the other one is to incorporate the knowledge or patterns from our experience or data into the prediction models.
On the one hand, the deep learning technique provides us one very powerful tool for constructing the prediction model. In the deep learning models, more representative features can be extracted from the lowest layer to the highest layer [11,12]. Until today, this miraculous technique has been widely used in various fields. In [13], a novel predictor, the stacked autoencoder Levenberg–Marquardt model was constructed for the prediction of traffic flow. In [14], an extreme deep learning approach that integrates the stacked autoencoder (SAE) with the ELM was proposed for building energy consumption prediction. In [15], the deep learning was employed as an ensemble technique for cancer detection. In [16], the deep convolutional neural network (CNN) was utilized for face photo-sketch recognition. In [17], a deep learning approach, the Gaussian–Bernoulli restricted Boltzmann machine (RBM) was applied to 3D shape classification through using spectral graph wavelets and the bag-of-features paradigm. In [18], the deep belief network (DBN) was applied to solve the natural language understanding problem. Furthermore, in [19], the DBN was utilized to fuse the virtues of multiple acoustic features for improving the robustness of voice activity detection. As one popular deep learning method, the DBN has shown its superiority in machine learning and artificial intelligence. This study will adopt and modify the DBN to make it be suitable for the prediction of building energy consumption.
On the other hand, knowledge or patterns from our experience can provide additional information for the design of the prediction models. In [20,21,22], different kinds of prior knowledge were incorporated into the SVM models. In [23], the knowledge of symmetry was encoded into the type-2 fuzzy logic model to enhance its performance. In [24,25], the knowledge of monotonicity was incorporated into the fuzzy inference systems to assure the models’ monotonic input–output mappings. In [26,27,28,29], how to encode the knowledge into neural networks was discussed. As shown in these studies, through incorporating the knowledge or pattern, the constructed machine learning models will yield better performance and have significantly improved generalization ability.
From the above discussion, both the deep learning method and the domain knowledge are helpful for the prediction models’ performance improvement. Following this idea, this study tries to present a hybrid model that combines the DBN model with the periodicity knowledge of the building energy consumption to further improve the prediction accuracy. The final prediction results of the proposed hybrid model are obtained by combining the outputs from the modified DBN model and the energy-consuming pattern model. Here, the energy-consuming pattern represents the periodicity property of building energy consumption and can be extracted from the observed historical energy consumption data. In this study, firstly, the structure of the proposed hybrid model will be presented, and how to extract the energy-consuming pattern will be demonstrated. Then, the training algorithm for the modified DBN model will be provided. The learning of the DBN model mainly includes two steps, which firstly optimizes the hidden parameters by the contrastive divergence (CD) algorithm in a pre-train way, and then determines the output weighting vector by the least squares method. Furthermore, the proposed hybrid model will be applied to the prediction of the energy consumption in two kinds of buildings that have different energy-consuming patterns (daily-periodicity and weekly-periodicity). Additionally, to show the superiority of the proposed hybrid model, comparisons with four popular artificial intelligence methods—the backward propagation neural network (BPNN), the generalized radial basis function neural network (GRBFNN), the extreme learning machine (ELM), and the support vector regressor (SVR) will be made. From the comparison results, we can observe that all the predictors (DBN, BPNN, GRBFNN, ELM and SVR) designed using both the periodicity knowledge and residual data perform much better than those designed only by the original data. Hence, we can judge that the periodicity knowledge is quite useful for improving the prediction performance in this application. The experiments also show that, among all the prediction models, the proposed DBN based hybrid model has the best performance.
The rest of this paper is as follows. In Section 2, the deep belief network will be reviewed. In Section 3, the proposed hybrid model will be presented firstly, and then the modified DBN will be provided. In Section 4, two energy consumption prediction experiments for buildings that have different energy-consuming patterns will be done. In addition, the experimental and comparison results will be given. Finally, in Section 5, the conclusions of this paper will be drawn.

2. Introduction of DBN

The DBN is a stack of restricted Boltzmann machine (RBM) [11,30]. Therefore, for better understanding, we will introduce the RBM before the introduction of the DBN in this section.

2.1. Restricted Boltzmann Machine

The structure of a typical RBM model is shown in Figure 1. The RBM is an undirected, bipartite graphical model, which consists of the visible (input) layer and the hidden (output) layer. The visible layer and the hidden layer are respectively made up of n visible units and m hidden units, and there is a bias in each unit. Moreover, there are no interconnection within the visible layer or the hidden layer [31].
The activation probability of the jth hidden unit can be computed as follows when a visible vector v ( v 1 , , v i , , v n ) is given [32]
p ( h j = 1 | v ) = σ ( b j + i = 1 n v i w i j ) ,
where σ ( · ) is the sigmoid function, w i j is the connection weight between the ith visible unit and jth hidden unit, and b j is the bias of the jth hidden unit.
Similarly, when a hidden vector h ( h 1 , , h j , , h m ) is known, the activation probability of the ith visible unit can be computed as follows:
p ( v i = 1 | h ) = σ ( a i + j = 1 m h j w i j ) ,
where i = 1 , 2 , , n , and a i is the bias of the ith visible unit.
Hinton et al. [33] have proposed the contrastive divergence (CD) algorithm to optimize the RBM. The CD algorithm based RBM’s iterative learning procedures for binomial units are listed as follows [32].
Step 1:
Initialize the number of visible units n, the number of hidden units m, the number of training data N, the weighting matrix W , the visible bias vector a , the hidden bias vector b and the learning rate ϵ .
Step 2:
Assign a sample x from the training data to be the initial state v 0 of the visible layer.
Step 3:
Calculate p ( h 0 j = 1 | v 0 ) according to Equation (1), and extract h 0 j { 0 , 1 } from the conditional distribution p ( h 0 j = 1 | v 0 ) , where j = 1 , 2 , , m .
Step 4:
Calculate p ( v 1 i = 1 | h 0 ) according to Equation (2), and extract v 1 i { 0 , 1 } from the conditional distribution p ( v 1 i = 1 | h 0 ) , where i = 1 , 2 , , n .
Step 5:
Calculate p ( h 1 j = 1 | v 1 ) according to Equation (1).
Step 6:
Update the parameters according to the following equations:
W = W + ϵ ( p ( h 0 = 1 | v 0 ) v 0 T p ( h 1 = 1 | v 1 ) v 1 T ) ,
a = a + ϵ ( v 0 v 1 ) ,
b = b + ϵ ( p ( h 0 = 1 | v 0 ) p ( h 1 = 1 | v 1 ) ) .
Step 7:
Assign another sample from the training data to be the initial state v 0 of the visible layer, and iterate Steps 3 to 7 until all the N training data have been used.

2.2. Deep Belief Network

As aforementioned, the DBN as a miraculous deep model is a stack of RBMs [11,30,34,35]. Figure 2 illustrates the architecture of the DBN with k hidden layers and its layer-wise pre-training process.
The activation of the kth hidden layer with respect to input sample x can be computed as
A k ( x ) = σ b k + W k σ + W 2 σ b 1 + W 1 x ,
where W u and b u ( u = 1 , 2 , , k ) are, respectively, the weighting matrices and hidden bias vectors of the uth RBM. Furthermore, σ is the logistic sigmoid function σ ( x ) = 1 / ( 1 + e x ) .
In order to obtain better feature representation, the DBN utilizes deep architecture and adopts the layer-wise pre-training to optimize the inter-layer weighting matrix [11]. The training algorithm of the DBN will be given in the next section in detail.

3. The Proposed Hybrid Model

In this section, the structure of the hybrid model will be proposed first. Then, the extraction of the energy-consuming pattern and the generation of the residual data will be given. Finally, the modified DBN (MDBN) and its training algorithm will be presented.
To begin, we assume that we have collected the sampling data for M consecutive days, and, in each day, we collected T data points. Then, sampled time series of energy consumption data can be written as a series of 1D vectors as
Y = Y 1 , Y 2 , , Y M ,
where
Y 1 = [ y 1 ( 1 ) , y 1 ( 2 ) , , y 1 ( T ) ] , Y M = [ y M ( 1 ) , y M ( 2 ) , , y M ( T ) ] ,
and T is the sampling number per day.

3.1. Structure of the Hybrid Model

The hybrid model combines the modified DBN (MDBN) model with the periodicity knowledge of the building energy consumption to obtain better prediction accuracy. The design procedure of the proposed model is depicted in Figure 3 and is also given as follows:
Step 1:
Extract the energy-consuming pattern as the periodicity knowledge from the training data.
Step 2:
Remove the energy-consuming pattern from the training data to generate the residual data.
Step 3:
Utilize the residual data to train the MDBN model.
Step 4:
Combine the outputs from the MDBN model with the periodicity knowledge to obtain the final prediction results of the hybrid model.
It is obvious that the extraction of the energy-consuming pattern, the generation of the residual data and the construction of the MDBN model are crucial in order to build the proposed hybrid model. Consequently, we will introduce them in detail in the following subsections.

3.2. Extraction of the Energy-Consuming Patterns and Generation of the Residual Data

Obviously, various regular patterns of energy consumption (e.g., daily-periodicity, weekly-periodicity, monthly-periodicity and even yearly-periodicity) exist in different kinds of buildings. In this study, we will take the daily-periodic and the weekly-periodic energy-consuming patterns as examples to introduce the method for extracting them from the original data.

3.2.1. The Daily-Periodic Pattern

For daily-periodic energy-consuming pattern, it can be extracted from the original time series by the following equation:
Y ¯ A v e = 1 M z = 1 M y z ( 1 ) , 1 M z = 1 M y z ( 2 ) , , 1 M z = 1 M y z ( T ) .
Then, the residual time series Y R e s of the data set after removing the daily-periodic pattern can be generated as
Y R e s = Y 1 Y ¯ A v e , Y 2 Y ¯ A v e , , Y M Y ¯ A v e .

3.2.2. The Weekly-Periodic Pattern

Being different from the daily-periodic energy-consuming pattern, the weekly-periodic energy-consuming pattern includes two parts, which are the patterns of weekdays and weekends. The weekday pattern and the weekend pattern can be respectively computed as
Y ¯ W e e k d a y = 1 M 1 z = 1 M 1 p z ( 1 ) , 1 M 1 z = 1 M 1 p z ( 2 ) , , 1 M 1 z = 1 M 1 p z ( T ) ,
Y ¯ W e e k e n d = 1 M 2 z = 1 M 2 q z ( 1 ) , 1 M 2 z = 1 M 2 q z ( 2 ) , , 1 M 2 z = 1 M 2 q z ( T ) ,
where
P = P 1 = [ p 1 ( 1 ) , , p 1 ( T ) ] , , P M 1 = [ p M 1 ( 1 ) , , p M 1 ( T ) ] ,
Q = Q 1 = [ q 1 ( 1 ) , , q 1 ( T ) ] , , Q M 2 = [ q M 2 ( 1 ) , , q M 2 ( T ) ] ,
are, respectively, the data sets of weekdays and weekends, and M 1 + M 2 = M .
Then, to generate the residual time series Y R e s for the building energy consumption data set, we use the following rules:
I f Y z P , t h e n Y z , R e s = Y z Y ¯ W e e k d a y ,
I f Y z Q , t h e n Y z , R e s = Y z Y ¯ W e e k e n d ,
where z = 1 , 2 , , M .
Subsequently, Y R e s can be written as
Y R e s = Y 1 , R e s , Y 2 , R e s , , Y M , R e s .

3.3. Modified DBN and Its Training Algorithm

In this subsection, the structure of the MDBN will be shown firstly. Then, the pre-training process of the DBN part will be described in detail. At last, the least squares method will be employed to determine the weighting vector of the regression part.

3.3.1. Structure of the MDBN

In the parameter optimization of the traditional DBNs, the CD algorithm is adopted to pre-train the parameters of multiple RBMs, and the BP algorithm is used to finely tune the parameters of the whole network. In this paper, we add an extra layer as the regression part to the DBN to realize the prediction function. Thus, we call it the modified DBN (MDBN). The structure of the MDBN is demonstrated in Figure 4. In addition, we propose a training algorithm that combines the CD algorithm with the least squares method for the learning of the MDBN model.
We divide the training process of the MDBN into two steps. The first step adopts the contrastive divergence algorithm to optimize the hidden parameters in a pre-train way, while the second one determines the output weighting vector by the least squares method. The detailed description will be given as below.

3.3.2. Pre-Training of the DBN Part

Generally speaking, with the number of hidden layers increasing, the effectiveness of the BP algorithm for optimizing the parameters of the deep neural network is getting lower and lower because of the gradient divergence. Fortunately, Hinton et al. [11] proposed a fast learning algorithm for the DBN. This novel approach realizes layer-wise pre-train of the multiple RBMs in the DBN in a bottom-up way as described below:
Step 1:
Initialize the number of hidden layers k, the number of the training data N and the initial sequence number of hidden layer u = 2 .
Step 2:
Assign a sample x from the training data to be the input data of the DBN.
Step 3:
Regard the input layer and the first hidden layer of the DBN as an RBM, and compute the activation A 1 ( x ) by Equation (3) when the training process of this RBM is finished.
Step 4:
Regard the uth and the ( u + 1 ) th hidden layer as an RBM with the input A u 1 ( x ) , and compute the activation A u ( x ) by Equation (3) when the training process of this RBM is completed.
Step 5:
Let u = u + 1 , and iterate Step 4 until u > k .
Step 6:
Use the A k ( x ) as the input of the regression part.
Step 7:
Assign another sample from the training data as the input data of the DBN, and iterate Step 3 to 7 until all the N training data have been assigned.

3.3.3. Least Squares Learning of the Regression Part

Suppose that the training set is = { ( x ( l ) , y ( l ) ) | x ( l ) R n , y ( l ) R , l = 1 , , N } . As aforementioned, once the pre-training of the DBN part is completed, the activation of the final hidden layer of the MDBN with respect to the input x ( l ) can be obtained to be A k ( x ( l ) ) , where l = 1 , 2 , , N . Furthermore, the activation of the final hidden layer of the MDBN with respect to all the N training data can be written in the matrix form as
A k ( X ) = [ A k ( x ( 1 ) ) , A k ( x ( 2 ) ) , , A k ( x ( N ) ) ] T = σ b k + w k σ + w 2 σ b 1 + w 1 x ( 1 ) σ b k + w k σ + w 2 σ b 1 + w 1 x ( 2 ) σ b k + w k σ + w 2 σ b 1 + w 1 x ( N ) N × n k ,
where n k is the number of neurons of the kth hidden layer.
We always expect that each actual value y ( l ) with respect to x ( l ) can be approximated by the output y ^ ( l ) of the predictor with no error. This expectation can be mathematically expressed as
l = 1 N y ^ ( l ) y ( l ) = 0 ,
where y ^ ( l ) is the output of the MDBN and can be computed as
y ^ ( l ) = A k ( x ( l ) ) β
in which β is the output weighting vector and can be expressed as
β = [ β 1 , β 2 , , β n k ] n k × 1 T .
Then, Equation (16) can be rewritten in the matrix form as
A k ( X ) β = Y ,
where
Y = [ y ( 1 ) , y ( 2 ) , , y ( N ) ] N × 1 T .
From Equation (19), the output weighting vector β can be derived by the least squares method as [36,37,38,39]
β = A k ( X ) Y ,
where A k ( X ) is the Moore–Penrose generalized inverse of A k ( X ) .

4. Experiments

In this section, first of all, four comparative artificial intelligence approaches will be introduced briefly. Next, the applied data sets and experimental setting will be discussed. Then, the proposed hybrid model will be applied to the prediction of the energy consumption in a retail store and an office building that respectively have daily-periodic and weekly-periodic energy-consuming patterns. Finally, we will give the comparisons and discussions of the experiments.

4.1. Introduction of the Comparative Approaches

To make a quantitative assessment of the proposed MDBN based hybrid model, four popular artificial intelligence approaches, the BPNN, GRBFNN, ELM, and SVR, are chosen as the comparative approaches and introduced briefly below.

4.1.1. Backward Propagation Neural Network

The structure of BPNN with L hidden layers is demonstrated in Figure 5. The BPNN as one popular kind of ANN adopts back propagation algorithm to obtain the optimal weighting parameters of the whole network [40,41,42].
As shown in Figure 5, the final output of the network can be expressed as [40,41,42]
y ^ = f ( s = 1 n L w s 1 L + 1 f ( j = 1 n 1 w j k 2 f ( i = 1 n w i j 1 x i ) ) ) ,
where w i j k is the connection weight between the ith unit of kth layer and the jth unit of ( k + 1 ) th layer, and f ( · ) is the logistic sigmoid function.
In order to obtain the optimal parameters of the BPNN, the Backward Propagation (BP) algorithm is adopted to minimize the following cost function for each training data point
E ( t , w ) = ( y ^ ( t ) y ( t ) ) 2 ,
where y ^ ( t ) and y ( t ) are the predicted and actual values with respect to the input x ( t ) .
The update rule for the weight w i j k can be expressed as
w i j k ( t + 1 ) = w i j k ( t ) η E ( t , w ) w i j k ,
where η is the learning rate, and E ( t , w ) w i j k is the gradient of the parameter w i j k , and can be calculated by the backward propagation of the errors.
The BP algorithm has two phases—forward propagation and weight update. In the forward propagation stage, when an input vector is input to the NN, it is propagated forward through the whole network until it reaches the output layer. Then, the error between the output of the network and the desired output is computed. In the weight update phase, the error is propagated from the output layer back through the whole network, until each neuron has an associated error value that can reflect its contribution to the original output. These error values are then used to calculate the gradients of the loss function that are fed to the update rules to renew the weights [40,41,42].

4.1.2. Generalized Radial Basis Function Neural Network

The radial basis function (RBF) NN is a feed-forward NN with only one hidden layer whose structure is demonstrated in Figure 6. The RBFNN has Gaussian functions as its hidden neurons. The GRBFNN is a modified RBFNN and adopts the generalized Gaussian functions as its hidden neurons [43,44].
The output of the GRBFNN can be expressed as [43,44]
y ^ = j = 1 n 1 w j G j ( x ) = j = 1 n 1 w j e x p x c j τ j d j τ j ,
where n 1 is the number of hidden neurons, τ j is the shape parameter of the jth radial basis function in the hidden layer, and c j and d j are, respectively, the center and width of the jth radial basis function.
In order to determine the parameters τ , c and d in the hidden layer and the connection weight w j , the aforementioned BP algorithm can also be employed.

4.1.3. Extreme Learning Machine

The ELM is also a feed-forward neural network with only one hidden layer as demonstrated in Figure 6. However, the ELM and GRBFNN have different parameter learning algorithms and different activation functions in the hidden neurons.
In the ELM, the activation functions in the hidden neurons can be the hard-limiting activation function, the Gaussian activation function, the Sigmoidal function, the Sine function, etc. [36,37].
In addition, the learning algorithm for the ELM is listed below:
  • Randomly assign input weights or the parameters in the hidden neurons.
  • Calculate the hidden layer output matrix H , where
    H = G 1 ( x ( 1 ) ) G n 1 ( x ( 1 ) ) G 1 ( x ( N ) ) G n 1 ( x ( N ) ) N × n 1 .
  • Calculate the output weights w = [ w 1 , w 2 , , w n 1 ] T = H + Y , where Y = [ y ( 1 ) , y ( 2 ) , , y ( N ) ] T and H + is the Moore–Penrose generalized inverse of the matrix H .
This learning process is very fast and can lead to excellent modeling performance. Hence, the ELM has found lots of applications in different research fields.

4.1.4. Support Vector Regression

The SVR is a variant of SVM. It can yield improved generalization performance through minimizing the generalization error bound [45]. In addition, the kernel trick is adopted to realize the nonlinear transformation of input features.
The model of the SVR can be defined by the following function
y ^ = f ( x , w ) = w T φ ( x ) + b ,
where w = [ w 1 , , w n ] , φ ( x ) is the nonlinear mapping function.
Using the training set = { ( x ( l ) , y ( l ) ) } l = 1 N , we can determine the parameters w and b, and then obtain the SVR model as
y ^ = f ( x ) = l = 1 N w * T φ ( x ) + b * ,
where
w * = l = 1 N ( α l * α l ) φ ( x ( l ) ) , b * = 1 y l w * T φ ( x ( l ) ) ,
in which α l and α l * are the Langrange multipliers and can be determined by solving the following dual optimization problem [46]:
max α , α * ε l = 1 N ( α l * + α l ) + l = 1 N ( α l * α l ) y ( l ) 1 2 l , t = 1 N ( α l * α l ) ( α t * α t ) φ T ( x ( l ) ) φ ( x ( t ) ) , l = 1 N α l * = l = 1 N α l , 0 < α l , α l * < C ,
where C is the regularization parameter and ε is the error tolerance parameter.

4.2. Applied Data Sets and Experimental Setting

In this subsection, first of all, the building energy consumption data sets will be described. Next, three design factors that are utilized to determine the optimal structure of the MDBN will be shown. Finally, five indices will be given to evaluate the performances of the predictive models.

4.2.1. Applied Data Sets

Two kinds of building energy consumption data sets were downloaded from [47]. The first data set includes 34,848 samples from 2 January 2010 to 30 December 2010. The data in this data set were collected every 15 min in one retail store in Fremont, CA, USA. We then aggregated them to generate the hourly energy consumption data. The second data set contains 22,344 samples from 4 April 2009 to 21 October 2011. The data in this set were collected every 60 min in one office building in Fremont, CA, USA. Parts of the samples of the two data sets are depicted in Figure 7.

4.2.2. Design Factors for MDBN

To determine the optimal structure of the MDBN for building energy consumption prediction, we will take three design factors, the number of hidden layers, hidden neurons and input variables, with their corresponding levels into account. The three design factors and their corresponding levels are presented in Table 1 and discussed in detail below.
  • Design Factor i: the number of hidden layers k
    The number of hidden layers determines how many RBMs are stacked. In this study, we consider the number of hidden layers 2, 3 and 4 as Levels 1, 2 and 3, respectively.
  • Design Factor ii: the number of uth hidden units n u
    The number of hidden units is an important factor that greatly influences the performance of the MDBN model. Here, we assume that the numbers of neurons in all hidden layers are equal, i.e., n 1 = n 2 = = n k . In this paper, we set the number of neurons 50, 100 and 150 as Levels 1, 2 and 3, respectively.
  • Design Factor iii: the number of input variables r
    In this paper, we utilize r energy consumption data in the building energy consumption time series before time t to predict the value at time t. In other words, we utilize x = [ y ( t 1 ) , y ( t 2 ) , , y ( t r ) ] to predict the value of y = y ( t ) . Here, we consider the number of input variables 4, 5 and 6 as Levels 1, 2 and 3, respectively.

4.2.3. Comparison Setting

In this study, the performances of all the predictors constructed by utilizing the energy-consuming patterns are compared with those designed by the original data. To evaluate the performances of the models, we utilize the following two kinds of indices.
We first consider the mean absolute error (MAE), the root mean square error (RMSE), and the mean relative error (MRE), and calculate them as
M A E = 1 K l = 1 K y ^ ( l ) y ( l ) ,
R M S E = l = 1 K ( y ^ ( l ) y ( l ) ) 2 K ,
M R E = 1 K l = 1 K y ^ ( l ) y ( l ) y ( l ) × 100 % ,
where K is the number of training or testing data pairs, and y ^ ( l ) , y ( l ) are, respectively, the predicted value and actual value with respect to the input x ( l ) .
The MAE, RMSE and MRE are common measures of forecasting errors in time series analysis. They serve to aggregate the magnitudes of the prediction errors into a single measure. The MAE is an average of the absolute errors between the predicted values and actual observed values. In addition, the RMSE represents the sample standard deviation of the differences between the predicted values and the actual observed values. As larger errors have a disproportionately large effect on MAE and RMSE, they are sensitive to outliers. The MRE, also known as the mean absolute percentage deviation, can remedy this drawback, and it expresses the prediction accuracy as a percentage through dividing the absolute errors by their corresponding actual values. For prediction applications, the smaller the values of MAE, RMSE and MRE are, the better the forecasting performance will be.
To better show the validity of the models, we also consider another two statistical indices, which are, respectively, the Pearson correlation coefficient, denoted as r, and the coefficient of determination, denoted as R 2 . These two indices can be calculated as
r = K ( l = 1 K y ^ ( l ) · y ( l ) ) ( l = 1 K y ^ ( l ) ) · ( l = 1 K y ( l ) ) ( K l = 1 K ( y ^ ( l ) ) 2 ( l = 1 K y ^ ( l ) ) 2 ) · ( K l = 1 K ( y ( l ) ) 2 ( l = 1 K y ( l ) ) 2 ) ,
R 2 = l = 1 K ( y ^ ( l ) y ^ A v e ) · ( y ( l ) y A v e ) 2 l = 1 K ( y ^ ( l ) y ^ A v e ) · l = 1 K ( y ( l ) y A v e ) ,
where K is also the number of training or testing data pairs, and y ^ A v e , y A v e are, respectively, the averages of the predicted and actual values.
The statistic r is a measure of the linear correlation between the actual values and the predicted values. It ranges from 1 to 1, where 1 means the total negative linear correlation, while 1 is total positive linear correlation. The statistic R 2 provides a measure of how well actual observed values are replicated by the predicted values. In other words, it is a measure of how good a predictor might be constructed from the observed training data [48]. The value of R 2 ranges from 0 to 1. In regression applications, the larger the values of r and R 2 are, the better the prediction performances will be.

4.3. Energy Consumption Prediction for the Retail Store

In this subsection, the energy-consuming pattern of the retail store will be extracted from the retail store data set firstly. Then, the configurations of the five prediction models for predicting the retail store energy consumption will be shown in detail. At last, the experimental results will be given.

4.3.1. Energy-Consuming Pattern of the Retail Store

We utilize Equations (6) and (7) to obtain the daily-periodic energy-consuming pattern and the residual time series of the retail store.
Figure 8a shows the daily-periodic energy-consuming pattern. In addition, the residual time series of the retail store, which is used to optimize the MDBN is demonstrated in Figure 8b.

4.3.2. Configurations of the Prediction Models

As aforementioned, we will take three design factors, the number of hidden layers, hidden neurons and input variables, with their corresponding levels into account to determine the optimal structure of the MDBN model for building energy consumption prediction. Consequently, 3 3 = 27 trials are ran. In addition, the experimental results are shown in Table 2. It is obvious that trail 19 can obtain the best performance. In other words, the optimal structure of the MDBN for retail store energy consumption prediction has four hidden layers, 150 hidden units and four input variables.
Furthermore, the parameter configurations of the other four comparative predictors for retail store energy consumption prediction are listed in detail as follows.
  • For the BPNN, there were 110 neurons in the hidden layer that can realize the nonlinear transformation of features by the sigmoid function. Additionally, the algorithm was ran for 7000 iterations to achieve the learning objective.
  • For the GRBFNN, the 6-fold cross-validation was adopted to determine the optimized spread of the radial basis function. Furthermore, the spread was chosen from 0.01 to 2 with the 0.1 step length.
  • For the ELM, there were 100 neurons in the hidden layer, and the hardlim function was chosen as the activation function for converting the original features into another space.
  • For the SVR, the penalty coefficient was set to be 80, and the radial basis function was chosen as the kernel function to realize the nonlinear transformation of input features.

4.3.3. Experimental Results

For the testing data of the retail store, parts of the prediction results of the five predictors constructed by utilizing the energy-consuming pattern are illustrated in Figure 9. Furthermore, for better visualization, the prediction error histograms of the five predictors are shown in Figure 10. It is obvious that the more the prediction errors float around zero, the better the forecasting performance of the predictor will be.
Then, to examine the superiority of the hybrid model for the retail store energy consumption prediction, the five prediction models are compared considering different data types (the original and residual data). The original data means that the predictors are learned using the original data series, while the residual data means that the predictors are constructed by both the energy-consuming pattern and the residual data series. Experimental results are demonstrated in detail in Table 3.

4.4. Energy Consumption Prediction for the Office Building

In this subsection, first of all, the energy-consuming pattern of the office building will be extracted from the office building data set. Then, the configurations of the five prediction models for predicting the office building energy consumption will be shown in detail. Finally, the experimental results will be given.

4.4.1. Energy-Consuming Pattern of the Office Building

Being similar to the retail store experiment, we utilize Equations (8)–(14) to obtain the weekly-periodic energy-consuming pattern and the residual time series of the office building.
As mentioned previously, the weekly-periodic energy-consuming pattern should include two parts, which are the weekday pattern and the weekend pattern. The obtained weekday pattern is depicted in Figure 11a, while the weekend pattern is shown in Figure 11b. We can observe that the energy consumption in weekends is quite different from that in weekdays. After removing the energy-consuming pattern, the residual time series of the office building is demonstrated in Figure 11c. This residual time series is utilized to train the MDBN in the hybrid model.

4.4.2. Configurations of the Prediction Models

Similarly, we run 3 3 = 27 trials to determine the optimal structure of the MDBN model for the office building energy consumption prediction. The experimental results are listed in Table 4. As shown in Table 4, the trail 13 obtains the best performance. Consequently, the optimal structure of the MDBN in the hybrid model for office building has three hidden layers, 100 hidden units in each layer and four input variables.
For the other four comparative predictors, their parameter configurations for the office building energy consumption prediction are listed as follows:
  • For the BPNN, there were 200 neurons in the hidden layer. Furthermore, the sigmoid function was chosen to realize the nonlinear transformation of features. Additionally, we ran the BP algorithm 1000 times to obtain the final outputs.
  • For the GRBFNN, the 5-fold cross-validation was utilized to determine the optimized spread of the radial basis function. Furthermore, the spread was chosen from 0.01 to 2 with a 0.1 step length.
  • For the ELM, there were 150 neurons in the hidden layer, and the hardlim function was chosen as the activation function for converting the original features into another space.
  • For the SVR, the penalty coefficient was set to be 10 and the sigmoid function was chosen as the kernel function to realize the nonlinear transformation of input features.

4.4.3. Experimental Results

For the testing data of the office building, parts of the prediction results of the five predictors are illustrated in Figure 12. Again, for better visualization, the prediction error histograms of the five predictors are shown in Figure 13.
Then, in order to examine the superiority of the hybrid model for the office building energy consumption prediction, the five prediction models are compared under the consideration of different data types (the original and residual data). Experimental results are demonstrated in Table 5.

4.5. Comparisons and Discussions

As discussed previously, smaller values of the MAE, RMSE and MRE represent better prediction results while lager values of r and R 2 correspond to better performance. Considering all the values of such indices as shown in Table 3 and Table 5 (It is worth noting that the values of the indices in Table 3 are about the retail energy consumption while the values in Table 5 are about the office energy consumption. The retail building consumed much more energies than the office building. As a result, some values of the MAE, RMSE and MRE in Table 3 are larger than those in Table 5), the predictors constructed by utilizing the energy-consuming patterns perform better than those designed only by the original data. Taking the RMSE index for example, in the first experiment, the accuracies of the MDBN, BPNN, GRBFNN, ELM and SVR based hybrid models are promoted by 11.1%, 7.0%, 4.2%, 21.6% and 9.6%, respectively, while, in the second experiment, the accuracy improvements of such models are 15.6%, 14.8%, 26.5%, 16.9% and 34.0%, respectively. As a result, we can draw a conclusion that the periodicity knowledge is helpful to improve the accuracy for building energy consumption prediction.
From Figure 9 and Figure 12, we can see that the hybrid DBN model can not only predict the regular testing data well for both the retail store and the office building energy consumption from the global perspective, but also give the best prediction results for the noisy irregular data, e.g., the sampling points from 25 to 50 in Figure 9 in the retail store experiment. These irregular testing data can reflect the uncertainties in the energy consumption time series. In other words, the proposed hybrid DBN model has the most powerful ability to deal with the uncertain and/or the randomness in the historical building energy consumption data.
Figure 10 and Figure 13 demonstrated the prediction error histograms of the five models designed through using the periodicity knowledge in the two experiments. In the histograms, the horizontal direction depicts the exact values of the prediction errors, while the vertical direction indicates the number of the prediction errors in different partitioned intervals. The more the prediction errors float around zero, the better performance the predictors will achieve. From both figures, we can clearly observe that the proposed hybrid DBN model has more prediction errors floating near zero compared with the other four artificial intelligence techniques—that is to say, the approximation capability of the proposed hybrid DBN model is promising for the two experimented buildings. Furthermore, to further validate the accuracy of the MDBN based hybrid model, scatter plots of the actual and predicted values in the two experiments are demonstrated in Figure 14a,b, respectively. From Figure 14, we can observe that the predicted values from the hybrid DBN model can duplicate the actual values well.
Among all the predictors constructed by both the original and residual data, the proposed MDBN based hybrid model has the best prediction accuracy in the two experiments as shown in Table 3 and Table 5. This phenomenon indicates that the proposed deep learning method has the miraculous learning and prediction abilities in time series forecasting applications. This also verifies the powerful feature extraction ability of the deep learning algorithm and the effectiveness of the modified learning strategies.
One thing to be mentioned is that the numbers of the data used in this paper are not very big (about the ten thousand scale). Even though the hybrid MDBN model is not learned by big data in both experiments, it still shows us excellent performances. This is also consistent with some other application results where the DBNs were trained without a mass of data. For example, in [49,50], the DBNs were applied to the time series prediction and the wind power prediction, which also do not have a large quantity of data. In both applications, the experimental results demonstrated that the DBN approach performs best compared with the traditional techniques. All these applications verified the learning ability of the DBN models for not very large data applications.

5. Conclusions

As is well known, many kinds of time series data, e.g., the traffic flow time series and the electricity consumption time series, have the periodicity property. The hybrid model can be expected to yield better performance in the predictions of such time series. In the future, we will extend our approach to these applications. On the other aspect, our study only focuses on the data science that tries to utilize the data to realize the energy-consumption prediction without considering any scientific or practical information of energy related principles. Theoretically, the energy related principles are very helpful to improve the prediction performance. We are now exploring the strategies to construct the novel hybrid prediction models through combining the energy related principles and observed data to further improve the prediction accuracy.
In this paper, a hybrid model is presented to further improve the prediction accuracy for building energy consumption prediction. The proposed model combines the MDBN model with the periodicity knowledge to obtain the final prediction results. The theoretical contributions of this study consist of two aspects: (1) the periodicity knowledge was extracted and encoded into the prediction model. In addition, the prediction accuracy can be greatly improved through utilizing this kind of prior knowledge; (2) a novel learning algorithm that combines the contrastive divergence algorithm and the least squares method was proposed to optimize the parameters of the MDBN. This is the first time that the DBN is applied to the building energy consumption prediction. On the other hand, this study applied the proposed approach to the energy consumption prediction of two kinds of buildings. Experimental and comparison results verified the effectiveness and superiorities of the proposed hybrid model.
References

Acknowledgments

This work is supported by the National Natural Science Foundation of China (61473176, 61105077, 61573225), and the Natural Science Foundation of Shandong Province for Young Talents in Province Universities (ZR2015JL021).

Author Contributions

Chengdong Li, Jianqiang Yi and Yisheng Lv have contributed to developing ideas about energy consumption prediction and collecting the data. Zixiang Ding and Guiqing Zhang programmed the algorithm and tested it. All of the authors were involved in preparing the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ahmad, M.W.; Mourshed, M.; Rezgui, Y. Trees vs Neurons: Comparison between random forest and ANN for high-resolution prediction of building energy consumption. Energy Build. 2017, 147, 77–89. [Google Scholar] [CrossRef]
  2. Banihashemi, S.; Ding, G.; Wang, J. Developing a hybrid model of prediction and classification algorithms for building energy consumption. Energy Procedia 2017, 110, 371–376. [Google Scholar] [CrossRef]
  3. Naji, S.; Keivani, A.; Shamshirband, S.; Alengaram, U.J.; Jumaat, M.Z.; Mansor, Z.; Lee, M. Estimating building energy consumption using extreme learning machine method. Energy 2016, 97, 506–516. [Google Scholar] [CrossRef]
  4. Hsu, D. Comparison of integrated clustering methods for accurate and stable prediction of building energy consumption data. Appl. Energy 2015, 160, 153–163. [Google Scholar] [CrossRef]
  5. Dong, B.; Cao, C.; Lee, S.E. Applying support vector machines to predict building energy consumption in tropical region. Energy Build. 2005, 37, 545–553. [Google Scholar] [CrossRef]
  6. Jung, H.C.; Kim, J.S.; Heo, H. Prediction of building energy consumption using an improved real coded genetic algorithm based least squares support vector machine approach. Energy Build. 2015, 90, 76–84. [Google Scholar] [CrossRef]
  7. Hong, W.C.; Dong, Y.; Zhang, W.Y.; Chen, L.Y.; Panigrahi, B.K. Cyclic electric load forecasting by seasonal SVR with chaotic genetic algorithm. Int. J. Electr. Power Energy Syst. 2013, 44, 604–614. [Google Scholar] [CrossRef]
  8. Fan, G.F.; Peng, L.L.; Hong, W.C.; Sun, F. Electric load forecasting by the SVR model with differential empirical mode decomposition and auto regression. Neurocomputing 2016, 173, 958–970. [Google Scholar] [CrossRef]
  9. Hong, W.C. Chaotic particle swarm optimization algorithm in a support vector regression electric load forecasting model. Energy Convers. Manag. 2009, 50, 105–117. [Google Scholar] [CrossRef]
  10. Robinson, C.; Dilkina, B.; Hubbs, J.; Zhang, W.; Guhathakurta, S.; Brown, M.A.; Pendyala, R.M. Machine learning approaches for estimating commercial building energy consumption. Appl. Energy 2017, 208, 889–904. [Google Scholar] [CrossRef]
  11. Hinton, G.E.; Osindero, S.; Teh, Y.W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef] [PubMed]
  12. Lv, Y.; Duan, Y.; Kang, W.; Li, Z.; Wang, F.Y. Traffic flow prediction with big data: A deep learning approach. IEEE Trans. Intell. Transp. Syst. 2015, 16, 865–873. [Google Scholar] [CrossRef]
  13. Yang, H.F.; Dillon, T.S.; Chen, Y.P.P. Optimized structure of the traffic flow forecasting model with a deep learning approach. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2371–2381. [Google Scholar] [CrossRef] [PubMed]
  14. Li, C.; Ding, Z.; Zhao, D.; Yi, J.; Zhang, G. Building energy consumption prediction: An extreme deep learning approach. Energies 2017, 10, 1525. [Google Scholar] [CrossRef]
  15. Xiao, Y.; Wu, J.; Lin, Z.; Zhao, X. A deep learning-based multi-model ensemble method for cancer prediction. Comput. Methods Programs Biomed. 2018, 153, 1–9. [Google Scholar] [CrossRef] [PubMed]
  16. Galea, C.; Farrugia, R.A. Forensic face photo-sketch recognition using a deep learning-based architecture. IEEE Signal Process. Lett. 2017, 24, 1586–1590. [Google Scholar] [CrossRef]
  17. Masoumi, M.; Hamza, A.B. Spectral shape classification: A deep learning approach. J. Vis. Commun. Image Represent. 2017, 43, 198–211. [Google Scholar] [CrossRef]
  18. Sarikaya, R.; Hinton, G.E.; Deoras, A. Application of deep belief networks for natural language understanding. IEEE/ACM Trans. Audio Speech Lang. Process. (TASLP) 2014, 22, 778–784. [Google Scholar] [CrossRef]
  19. Zhang, X.L.; Wu, J. Deep belief networks based voice activity detection. IEEE Trans. Audio Speech Lang. Process. 2013, 21, 697–710. [Google Scholar] [CrossRef]
  20. Chen, C.C.; Li, S.T. Credit rating with a monotonicity-constrained support vector machine model. Expert Syst. Appl. 2014, 41, 7235–7247. [Google Scholar] [CrossRef]
  21. Wang, L.; Xue, P.; Chan, K.L. Incorporating prior knowledge into SVM for image retrieval. In Proceedings of the International Conference on Pattern Recognition, Cambridge, UK, 23–26 August 2004; pp. 981–984. [Google Scholar]
  22. Wu, X.; Srihari, R. Incorporating prior knowledge with weighted margin support vector machines. In Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Seattle, WA, USA, 22–25 August 2004; ACM: New York, NY, USA, 2004; pp. 326–333. [Google Scholar]
  23. Li, C.; Zhang, G.; Yi, J.; Wang, M. Uncertainty degree and modeling of interval type-2 fuzzy sets: definition, method and application Comput. Math. Appl. 2013, 66, 1822–1835. [Google Scholar] [CrossRef]
  24. Abonyi, J.; Babuska, R.; Verbruggen, H.B.; Szeifert, F. Incorporating prior knowledge in fuzzy model identification. Int. J. Syst. Sci. 2000, 31, 657–667. [Google Scholar] [CrossRef]
  25. Li, C.; Yi, J.; Zhang, G. On the monotonicity of interval type-2 fuzzy logic systems. IEEE Trans. Fuzzy Syst. 2014, 22, 1197–1212. [Google Scholar] [CrossRef]
  26. Chakraborty, S.; Chattopadhyay, P.P.; Ghosh, S.K.; Datta, S. Incorporation of prior knowledge in neural network model for continuous cooling of steel using genetic algorithm. Appl. Soft Comput. 2017, 58, 297–306. [Google Scholar] [CrossRef]
  27. Kohara, K.; Ishikawa, T.; Fukuhara, Y.; Nakamura, Y. Stock price prediction using prior knowledge and neural networks. Intell. Syst. Account. Financ. Manag. 1997, 6, 11–22. [Google Scholar] [CrossRef]
  28. Li, C.; Gao, J.; Yi, J.; Zhang, G. Analysis and design of functionally weighted single-input-rule-modules connected fuzzy inference systems. IEEE Trans. Fuzzy Syst. 2016. [Google Scholar] [CrossRef]
  29. Lin, H.; Lin, Y.; Yu, J.; Teng, Z.; Wang, L. Weighing fusion method for truck scales based on prior knowledge and neural network ensembles. IEEE Trans. Instrum. Meas. 2014, 63, 250–259. [Google Scholar] [CrossRef]
  30. Bengio, Y.; Lamblin, P.; Popovici, D.; Larochelle, H. Greedy layer-wise training of deep networks. In Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; pp. 153–160. [Google Scholar]
  31. Fischer, A.; Igel, C. An introduction to restricted Boltzmann machines. Prog. Pattern Recognit. Image Anal. Comput. Vis. Appl. 2012, 7441, 14–36. [Google Scholar]
  32. Bengio, Y. Learning Deep Architectures for AI; Now Publishers: Boston, MA, USA, 2009; pp. 1–127. [Google Scholar]
  33. Hinton, G. Training products of experts by minimizing contrastive divergence. Neural Comput. 2002, 14, 1771–1800. [Google Scholar] [CrossRef] [PubMed]
  34. Roux, N.L.; Bengio, Y. Representational power of restricted boltzmann machines and deep belief networks. Neural Comput. 2008, 20, 1631–1649. [Google Scholar] [CrossRef] [PubMed]
  35. Bu, S.; Liu, Z.; Han, J.; Wu, J.; Ji, R. Learning high-level feature by deep belief networks for 3-D model retrieval and recognition. IEEE Trans. Multimed. 2014, 16, 2154–2167. [Google Scholar] [CrossRef]
  36. Huang, G.B.; Wang, D.H.; Lan, Y. Extreme learning machines: A survey. Int. J. Mach. Learn. Cybern. 2011, 2, 107–122. [Google Scholar] [CrossRef]
  37. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  38. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks, Budapest, Hungary, 25–29 July 2004; Volume 2, pp. 985–990. [Google Scholar]
  39. Huang, G.B.; Chen, L.; Siew, C.K. Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans. Neural Netw. 2006, 17, 879–892. [Google Scholar] [CrossRef] [PubMed]
  40. Erb, R.J. Introduction to backpropagation neural network computation. Pharm. Res. 1993, 10, 165–170. [Google Scholar] [CrossRef] [PubMed]
  41. Uzlu, E.; Kankal, M.; Akpınar, A.; Dede, T. Estimates of energy consumption in Turkey using neural networks with the teaching-learning-based optimization algorithm. Energy 2014, 75, 295–303. [Google Scholar] [CrossRef]
  42. Yedra, R.M.; Diaz, F.R.; Nieto, M.D.M.C.; Arahal, M.R. A neural network model for energy consumption prediction of CIESOL bioclimatic building. Adv. Intell. Syst. Comput. 2014, 239, 51–60. [Google Scholar]
  43. Lu, J.; Hu, H.; Bai, Y. Generalized radial basis function neural network based on an improved dynamic particle swarm optimization and AdaBoost algorithm. Neurocomputing 2015, 152, 305–315. [Google Scholar] [CrossRef]
  44. Friedrichs, F.; Schmitt, M. On the power of Boolean computations in generalized RBF neural networks. Neurocomputing 2005, 63, 483–498. [Google Scholar] [CrossRef]
  45. Awad, M.; Khanna, R. Support vector regression. Neural Inf. Process. Lett. Rev. 2007, 11, 203–224. [Google Scholar]
  46. Wu, C.H.; Ho, J.M.; Lee, D.T. Travel-time prediction with support vector regression. IEEE Trans. Intell. Transp. Syst. 2004, 5, 276–281. [Google Scholar] [CrossRef]
  47. Buildings Datasets. Available online: https://trynthink.github.io/buildingsdatasets/ (accessed on 13 May 2017).
  48. Glantz, S.A.; Slinker, B.K. Primer of Applied Regression and Analysis of Variance; Health Professions Division, McGraw-Hill: New York, NY, USA, 1990. [Google Scholar]
  49. Hirata, T.; Kuremoto, T.; Obayashi, M.; Mabu, S.; Kobayashi, K. Time series prediction using DBN and ARIMA. In Proceedings of the International Conference on Computer Application Technologies, Atlanta, GA, USA, 10–14 June 2016; pp. 24–29. [Google Scholar]
  50. Tao, Y.; Chen, H. A hybrid wind power prediction method. In Proceedings of the Power and Energy Society General Meeting, Boston, MA, USA, 17–21 July 2016; pp. 1–5. [Google Scholar]
Figure 1. The structure of a typical RBM model.
Figure 1. The structure of a typical RBM model.
Energies 11 00242 g001
Figure 2. The architecture of the DBN with k hidden layers.
Figure 2. The architecture of the DBN with k hidden layers.
Energies 11 00242 g002
Figure 3. The structure of the hybrid model.
Figure 3. The structure of the hybrid model.
Energies 11 00242 g003
Figure 4. The structure of the modified DBN.
Figure 4. The structure of the modified DBN.
Energies 11 00242 g004
Figure 5. The structure of BPNN with L hidden layers.
Figure 5. The structure of BPNN with L hidden layers.
Energies 11 00242 g005
Figure 6. The topological structure of the feed-forward single-hidden-layer NN.
Figure 6. The topological structure of the feed-forward single-hidden-layer NN.
Energies 11 00242 g006
Figure 7. Parts of the samples of two data sets: (a) the first 500 data points of the retail store; (b) the first 500 data points of the office building.
Figure 7. Parts of the samples of two data sets: (a) the first 500 data points of the retail store; (b) the first 500 data points of the office building.
Energies 11 00242 g007
Figure 8. Periodicity knowledge and the residual time series of the retail store data set: (a) the daily-periodic energy-consuming pattern; (b) the residual time series.
Figure 8. Periodicity knowledge and the residual time series of the retail store data set: (a) the daily-periodic energy-consuming pattern; (b) the residual time series.
Energies 11 00242 g008
Figure 9. Parts of prediction results of the five predictors constructed by utilizing the energy-consuming pattern: (a) hybrid DBN model; (b) BPNN; (c) GRBFNN; (d) ELM; and (e) SVR.
Figure 9. Parts of prediction results of the five predictors constructed by utilizing the energy-consuming pattern: (a) hybrid DBN model; (b) BPNN; (c) GRBFNN; (d) ELM; and (e) SVR.
Energies 11 00242 g009
Figure 10. Prediction error histograms of the five predictors constructed by utilizing the energy-consuming pattern: (a) hybrid DBN model; (b) BPNN; (c) GRBFNN; (d) ELM; and (e) SVR.
Figure 10. Prediction error histograms of the five predictors constructed by utilizing the energy-consuming pattern: (a) hybrid DBN model; (b) BPNN; (c) GRBFNN; (d) ELM; and (e) SVR.
Energies 11 00242 g010
Figure 11. Periodicity knowledge and the residual time series of the office building data set: (a) the energy-consuming pattern of weekdays; (b) the energy-consuming pattern of weekends; (c) the residual time series.
Figure 11. Periodicity knowledge and the residual time series of the office building data set: (a) the energy-consuming pattern of weekdays; (b) the energy-consuming pattern of weekends; (c) the residual time series.
Energies 11 00242 g011
Figure 12. Parts of the prediction results of the five predictors constructed by utilizing the energy-consuming pattern: (a) hybrid DBN model; (b) BPNN; (c) GRBFNN; (d) ELM; and (e) SVR.
Figure 12. Parts of the prediction results of the five predictors constructed by utilizing the energy-consuming pattern: (a) hybrid DBN model; (b) BPNN; (c) GRBFNN; (d) ELM; and (e) SVR.
Energies 11 00242 g012
Figure 13. Prediction error histograms of the five predictors constructed by utilizing the energy-consuming pattern: (a) hybrid DBN model; (b) BPNN; (c) GRBFNN; (d) ELM; and (e) SVR.
Figure 13. Prediction error histograms of the five predictors constructed by utilizing the energy-consuming pattern: (a) hybrid DBN model; (b) BPNN; (c) GRBFNN; (d) ELM; and (e) SVR.
Energies 11 00242 g013
Figure 14. Scatter plots of the actual and predicted values of the energy consumptions in the retail building (a) and the office building (b).
Figure 14. Scatter plots of the actual and predicted values of the energy consumptions in the retail building (a) and the office building (b).
Energies 11 00242 g014
Table 1. Design factors and their corresponding levels.
Table 1. Design factors and their corresponding levels.
Design FactorsLevel
123
i2 hidden layers3 hidden layers4 hidden layers
ii50 hidden units100 hidden units150 hidden units
iii4 input variables5 input variables6 input variables
Table 2. Experimental results of the MDBN in 27 trails under the consideration of three design factors and their corresponding levels.
Table 2. Experimental results of the MDBN in 27 trails under the consideration of three design factors and their corresponding levels.
TrailsFactorResidual DataTrailsFactorResidual Data
iiiiiiMAE (kwh)MRE (%)RMSE (kwh)iiiiiiMAE (kwh)MRE (%)RMSE (kwh)
111149.215.2680.791522349.655.3180.03
211248.745.1880.031623150.185.3682.13
311348.735.1978.061723248.435.1178.24
412149.125.2481.201823348.335.1277.96
512248.255.1679.391931147.715.0376.83
612349.165.2479.362031248.375.1177.63
713149.425.2881.852131348.135.1177.60
813249.335.2581.402232148.725.1879.16
913348.655.1878.692332249.665.2879.84
1021148.735.2079.652432349.085.2278.03
1121249.615.2981.242533151.075.5083.35
1221347.955.0877.962633248.815.1879.22
1322148.835.1779.932733348.335.0977.50
1422249.975.3381.33
Table 3. The performances of the five models for the retail store energy consumption prediction.
Table 3. The performances of the five models for the retail store energy consumption prediction.
MethodsData TypeMAE (kwh)MRE (%)RMSE (kwh)r R 2
MDBNResidual data47.715.0376.830.940.89
Original data54.385.5986.430.930.86
BPNNResidual data65.697.2493.380.920.85
Original data75.458.20100.400.940.87
GRBFNNResidual data54.605.7583.870.930.87
Original data52.515.6287.540.930.86
ELMResidual data58.546.2988.620.930.86
Original data78.868.34113.020.890.79
SVRResidual data48.285.1981.310.930.87
Original data52.195.4289.930.920.85
Table 4. Experimental results of the MDBN in 27 trails under the consideration of three design factors and their corresponding levels.
Table 4. Experimental results of the MDBN in 27 trails under the consideration of three design factors and their corresponding levels.
TrailsFactorResidual DataTrailsFactorResidual Data
iiiiiiMAE (kwh)MRE (%)RMSE (kwh)iiiiiiMAE (kwh)MRE (%)RMSE (kwh)
11112.3012.673.69152232.3512.993.69
21122.2212.293.61162312.2512.493.65
31132.3212.743.67172322.3012.783.68
41212.2311.973.63182332.3613.103.71
51222.3512.813.71193112.2112.193.65
61232.4013.103.71203122.2312.293.66
71312.1711.933.58213132.2712.543.67
81322.2912.703.67223212.1712.063.60
91332.2712.553.63233222.2612.513.65
102112.2612.533.65243232.2312.253.67
112122.3112.833.68253312.1411.913.60
122132.3613.103.70263322.3212.643.73
132212.0911.623.54273332.2112.303.64
142222.3112.843.68
Table 5. The performances of the five models with different data types for the office building energy consumption prediction.
Table 5. The performances of the five models with different data types for the office building energy consumption prediction.
MethodsData TypeMAE (kwh)MRE (%)RMSE (kwh)r R 2
MDBNResidual data2.0911.623.540.970.93
Original data2.3211.504.190.950.90
BPNNResidual data2.5712.644.040.960.93
Original data3.8523.214.750.950.91
GRBFNNResidual data2.5412.624.390.950.91
Original data4.3521.945.980.930.87
ELMResidual data3.5017.184.920.960.92
Original data4.6125.525.920.900.82
SVRResidual data3.2314.894.980.940.88
Original data6.1334.427.550.920.85

Share and Cite

MDPI and ACS Style

Li, C.; Ding, Z.; Yi, J.; Lv, Y.; Zhang, G. Deep Belief Network Based Hybrid Model for Building Energy Consumption Prediction. Energies 2018, 11, 242. https://doi.org/10.3390/en11010242

AMA Style

Li C, Ding Z, Yi J, Lv Y, Zhang G. Deep Belief Network Based Hybrid Model for Building Energy Consumption Prediction. Energies. 2018; 11(1):242. https://doi.org/10.3390/en11010242

Chicago/Turabian Style

Li, Chengdong, Zixiang Ding, Jianqiang Yi, Yisheng Lv, and Guiqing Zhang. 2018. "Deep Belief Network Based Hybrid Model for Building Energy Consumption Prediction" Energies 11, no. 1: 242. https://doi.org/10.3390/en11010242

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop