Next Article in Journal
Prediction of Permeability Using Group Method of Data Handling (GMDH) Neural Network from Well Log Data
Next Article in Special Issue
Quantifying Electricity Supply Resilience of Countries with Robust Efficiency Analysis
Previous Article in Journal
Metal-Free Modified Boron Nitride for Enhanced CO2 Capture
Previous Article in Special Issue
The Impact of International Oil Prices on the Stock Price Fluctuations of China’s Renewable Energy Enterprises
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Ensemble Approach for the Forecasting of Energy Demand Based on the Artificial Bee Colony Algorithm

1
Institutes of Science and Development, Chinese Academy of Sciences, Beijing 100190, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Energies 2020, 13(3), 550; https://doi.org/10.3390/en13030550
Submission received: 12 December 2019 / Revised: 15 January 2020 / Accepted: 20 January 2020 / Published: 23 January 2020
(This article belongs to the Special Issue Energy Demand and Prices)

Abstract

:
Accurate forecasting of the energy demand is crucial for the rational formulation of energy policies for energy management. In this paper, a novel ensemble forecasting model based on the artificial bee colony (ABC) algorithm for the energy demand was proposed and adopted. The ensemble model forecasts were based on multiple time variables, such as the gross domestic product (GDP), industrial structure, energy structure, technological innovation, urbanization rate, population, consumer price index, and past energy demand. The model was trained and tested using the primary energy demand data collected in China. Seven base models, including the regression-based model and machine learning models, were utilized and compared to verify the superior performance of the ensemble forecasting model proposed herein. The results revealed that (1) the proposed ensemble model is significantly superior to the benchmark prediction models and the simple average ensemble prediction model just in terms of the forecasting accuracy and hypothesis test, (2) the proposed ensemble approach with the ABC algorithm can be employed as a promising framework for energy demand forecasting in terms of the forecasting accuracy and hypothesis test, and (3) the forecasting results obtained for the future energy demand by the ensemble model revealed that the future energy demand of China will maintain a steady growth trend.

1. Introduction

As a strategic supply, energy is an important foundation for the development of an economy and a society [1]. With the growth of socioeconomic and technological advancements in the past few decades, energy demand has increased considerably [2]. Particularly, as the world’s largest energy consumer, China has witnessed a three-fold increase in the energy demand throughout a 20-year period. In terms of this increased demand, a sufficient amount of energy is required to satisfy the nationwide demands, and the abundance of energy is of importance to guaranteeing the national energy security and sustainable development [3,4].
The scientific accurate prediction of energy demand is crucial for the rational formulation of energy policies and the basis for ensuring the security of energy supply [5,6] while the forecasting of the energy demand is conducive to prevent energy supply risks, reduce the gap between the energy supply and demand, slow down economic cycle fluctuations, and promote sustainable economic development and social stability. Therefore, forecasting its demand is crucial to the energy management and transportation sectors [7].
However, the forecasting of energy demand faces severe challenges due to the characteristics and properties of energy demand data. Energy demand data constitute a time series with strong nonlinear and chaotic characteristics. It contains a series of dynamic random factors, and it is affected by various factors, e.g., economic development level, industrial structure, and technology innovation [8]. Besides, the annual energy demand data are classified as small sample data, presenting higher requirements on the applicability of the prediction model [9]. In this case, the manner in which an effective forecast model is selected, and the accurate forecasting of energy demand are crucial issues that must be focused on.
To appropriately handle such a situation, a novel ensemble forecasting approach of energy demand based on the artificial bee colony (ABC) algorithm was constructed. The main aim was to seek forecasting models that are liable to train and exhibit high precision. Furthermore, seven forecast models were applied and compared to find the most suitable one, i.e., autoregressive integrated moving average (ARIMA), second exponential smoothing (SES), support vector regression (SVR), back propagation neural network (BPNN), radical basis function neural network (RBFNN), generalized regression neural network (GRNN), and extreme learning machine (ELM) models, respectively. In practice, seldom can a model achieve a good performance under any circumstances [10]. Ensemble learning functions by building and combining multiple machine learning machines to complete learning tasks, which can achieve a higher generalization ability and prediction accuracy [11,12]. Considering the rationality and adaptability of weight allocation, the ABC intelligent algorithm was introduced to integrate the prediction results of the base model. Different forecast models were selected for the combinatorial integration forecast to determine the model that exhibits the best performance.
The motivations and contributions of this paper were as follows. First, in view of the complexity of energy demand forecasting issue, this paper proposes a novel ensemble forecasting framework, which includes factor selection, data preprocessing, forecasting model training and testing process, and forecasting future energy demand. Second, considering the merits of the ABC algorithm in seeking the best integrated solution, this paper introduces the ABC algorithm to integrate the forecasting results of the base model for accurate energy demand forecasting. Third, it is difficult to characterize the impact of exogenous factors on energy demand in a single time series. To this end, this paper considers multiple factors to establish the forecasting model of energy demand. Additionally, this paper predicts the future energy demand of China based on the ensemble model.
The layout of this paper is organized as follows. Section 2 discusses the related studies focusing on energy demand forecasting. Section 3 discusses the formulation of the energy demand forecasting framework and the application of machine learning methods to the problem of energy demand forecasting. Section 4 discusses the empirical results and summary. Section 5 provides the concluding remarks.

2. Literature Review

2.1. Energy Demand Influencing Factors

In the last few decades, energy demand forecasting has turned into an active research area due to its significant impact on the energy security and socioeconomic development of a country [13,14]. A series of prominent research studies have reported on energy demand forecasting, including factor selection [7,15,16] and method determination [13,17].
Owing to the complexity of energy demand forecasting [18,19], it is an extremely difficult task to develop an energy demand forecasting model considering all of the influencing factors. Therefore, studies on the selection of influencing factors are mainly divided into single variables and multivariate variables. Due to the method characteristics, single-variable prediction mainly focuses on the traditional econometric model [20,21,22,23] and decomposition-integrated prediction model [24,25,26,27]. For example, Tso and Yau [28] examined the effects of socioeconomic indicators on the electricity demand. Adom and Bekoe [29] adopted macro socioeconomic indicators, such as GDP, population, and degree of urbanization. To build a model to forecast energy demand, Ghanbari et al. [30] collected data on potentially useful factors, such as the GDP, consumer price index (CPI), energy intensity, energy efficiency, and population, and utilized the feature selection technique to eliminate low impact factors and select the most influential ones. Wu and Peng [31] utilized multiple factors, such as economic growth, total population, fixed-asset investments, energy efficiency, and energy structure (ES), to estimate the energy demand of China. The influencing factors were identified (Table 1).
The factors influencing energy demand forecasting are different due to different energy types and forecasting ranges. Clear differences in terms of the influencing factors among different energy types are observed, such as the electricity demand, natural gas demand, and transportation energy demand. Besides, due to different forecasting ranges, differences in the selection of influencing factors are observed. For example, most of the annual forecasts are based on macroeconomic indicators while the influencing factors of daily forecasts focus on external factors, such as weather and energy prices. According to the range and type of energy forecast, macroeconomic indicators, such as the GDP, population, and industrial structure, for energy demand forecasting are focused herein. As mentioned above, this paper will make full use of expert knowledge and the related literature to identify the influencing factors of energy demand and establish a systematic and reasonable influencing factors list.

2.2. Energy Demand Forecasting Method

Due to the characteristics and properties of energy demand data, energy demand forecasting faces severe challenges and requirements on the applicability of the prediction model. Currently, various models have been used to forecast the energy demand [7,40,42,43,44,45], which can be divided into three categories: Statistical, mathematical programming (MP), and computational intelligence (CI). Statistical methods investigate the accumulation, examination, elucidation, presentation, and association of data [46], including linear regression, multiple regression, stepwise regression, and nonparametric regression. All of these methods have been widely applied in previous studies; these methods are capable of yielding better results for solving linear problems [29,35].
With the emergence of artificial intelligence, various learning technologies, such as the heuristic algorithm [30,40], support vector machine (SVM) [47,48], artificial neural network (ANN) [49], extreme learning machine (ELM) [50], and ensemble techniques, have become quite popular. Actually, real-life problems exhibit nonlinear characteristics during forecasting, especially for energy demand [46]. MP methods provide the best solutions from a set of available alternatives under some constraints. Among these models, nonlinear programming models have been widely used to deal with nonlinear forecasting problems while computational methods have been used for prediction problems where mathematical formulae and the prior known data on the relationship between inputs and outputs are unknown. ANN and SVM are the most popular forecasting techniques for dealing with nonlinear problems. Although each of the forecasting model exhibits its own merits and demerits (Table 2), CI methods have already attracted the most attention due to their high reliability and forecasting accuracy. These CI-based models have been vastly implemented for energy demand forecasting, including SVR, recurrent neural network (RNN), and convolutional neural network (CNN).
Time series analysis utilizes regression methods to establish various function equations of time series for trend prediction. Some classical time series analysis models, such as the exponential smoothing (ES) method [51] and ARIMA [52], have been widely used in the fields of social sciences. However, in the face of the large number of nonlinear complex time series prevalent in social and economic phenomena, the performance of a conventional mathematical statistics prediction method is not sufficiently good. On the contrary, machine learning models exhibit good performance for dealing with nonlinear problems, especially complex and nonlinear timing data, such as SVMs, neural networks, and various derivation methods. However, machine learning models have strict requirements on the data structure. When the amount of data is small, over-fitting may occur for the establishment of prediction models, leading to a weak generalization ability. In this case, computational learning theory results revealed that ensemble forecasting methods exhibit good prediction performance.
Ensemble learning functions by building and combining multiple machine learning machines to complete learning tasks, which can achieve a higher generalization ability and prediction accuracy [11,12]. As Bates and Granger [10] pointed out, a linear combination of the base models would yield a smaller forecasting variance than a single forecasting model. Since the linear weighting ensemble method cannot adjust the weights of single models adaptively, the adaptability of the overall prediction model is relatively poor [53]. Galicia et al. adopted the weighted least square method to forecast electricity demand [54]. Liu et al. [55] constructed an ensemble forecasting model combining the grey model, ARIMA, and second-order polynomial regression model with particle swarm optimization to predict CO2 emission in China. Wang et al. [56], Zhu and Wei [57], and Qu et al. [58] adopted a genetic algorithm, particle swarm optimization algorithm, and bat algorithm, respectively, for the ensemble forecasting model. The artificial bee colony (ABC) algorithm combines the advantages of local deep search and global-wide search and exhibits good performance for seeking the best integrated solution, and the ABC algorithm is simple to operate and easy to implement [59].
In this paper, a predictive analysis framework based on the ABC algorithm is proposed in this paper for small sample, nonlinear, and chaotic time series. Seven base models, including SES, ARIMA, SVR, BPNN, GRNN, RBFNN, and ELM, are used for energy demand forecasting, and the ABC algorithm is implemented to combine the base model predicted results.

3. Forecasting Framework and Methods

In this section, the framework and methods of energy demand forecasting are mainly discussed. First, the ensemble framework of energy demand forecasting is introduced, and the applied base models are described. Finally, crucial information on the ABC ensemble algorithm is provided.

3.1. Ensemble Framework of Energy Demand Forecasting

To better forecast energy demand under multi-factor disturbance, this paper presents the ensemble framework of energy demand forecasting (Figure 1) and the pseudo-code of the ensemble forecasting model, as shown in Appendix B Table A3. The ensemble framework is divided into four stages: (1) Influence index selection, (2) data preprocessing, (3) forecasting model training and testing, and (4) future energy demand forecasting.

3.1.1. Factor Selection

Energy systems are complex nonlinear systems, and energy demand is affected by various socioeconomic factors, such as the GDP, CPI, economic structure, and energy efficiency. To more accurately forecast China’s future energy demand, it is crucial to systematically analyze the influencing factors of energy demand. By reviewing the literature, this paper selected the indicators that affect energy demand (Table 3).
In addition, correlations, r, between the energy demand and selected factors are listed, and p is used to test the hypothesis that there is no relationship between the observed phenomena (null hypothesis). All of the selected factors pass the correlation test under the condition of α = 95%. The urbanization rate is one of the strongest factors of energy demand. Furthermore, a significant negative correlation between technological progress and energy demand is observed (Figure 2). In fact, the correlation plot (Figure 2) revealed that the ES exhibits a weak positive correlation with energy demand.

3.1.2. Data Preprocessing

Actually, the collected datasets are extremely sensitive to various differences (such as data frequency, noise, original units, magnitude, and missing values). These differences or errors in the raw data might result in a considerable deviation in the forecasted results. Therefore, it is crucial to preprocess the data to ensure reliability during knowledge mining. Generally, data preprocessing involves the following processes.
  • Data cleaning: Data cleaning is mainly utilized to fill in missing values, remove noise, monitor outliers, and deal with differences in datasets.
  • Data transformation: The data transformation process involves several methods, including the transformation of multiple files into a unified available data format as well as feature extraction.
  • Data standardization: Due to the different dimensions of data, large differences in the magnitude of collected data exist, often leading to large deviations in data analysis. Therefore, it is crucial to conduct the standardized processing of data to eliminate effects of dimensions and magnitude. The following Formula (1) is adopted in this paper:
    x i k = X i k min ( X i ) max ( X i ) min ( X i ) ,
    where x i k represents the data of the k - t h element corresponding to the i - t h influencing factor after standardization, and Xi represents the sequence corresponding to the i - t h influencing factor.
  • Data partitioning: To train and test the prediction model, observation values need to be divided into training sets and test sets. The data from the training set are used to train the prediction model to obtain optimal parameters, and then the data from the test set are used to test the generalization performance of the model.
The processed data are presented in the Appendix A Table A2.

3.1.3. Forecasting Model Training and Testing

At this stage, the collected data are mainly used to train and test the forecasting model. The model is trained by training set data. If the training results meet the requirements, the model parameters are saved; otherwise, the model parameters with the best prediction accuracy and generalization performance are retrained until they are obtained.

3.1.4. Forecasting Future Energy Demand

To predict the future energy demand, the input data (influencing factors) need to be predicted. For the time series data of influencing factors, the trend extrapolation method for forecasting is adopted (Figure 3). The future energy demand data can be obtained by inserting the data into the integrated model.

3.2. Base Models

The time series of energy demand exhibit nonlinear and chaotic characteristics. It is affected by various factors, such as the economic development level, technology innovation, and total population. Meanwhile, the annual energy demand belongs to the small sample, presenting higher requirements on the performance of the prediction model. Hence, the traditional econometric models and emerging machine learning models were selected as base models in this section. The detailed information about the selected base is as follows.

3.2.1. Autoregressive Integrated Moving Average

The ARIMA model is a combination of the autoregressive moving average model with differencing. It can be described as ARIMA (p, d, q), where p and q represent the lagged values of autoregressive and moving average sections, respectively [60]. d denotes the number of differences in the time series. Generally, the ARIMA model can be expressed by the following formulations:
X t = θ 0 + φ 1 X t 1 + φ 2 X t 2 + + φ p X t p θ 1 ε t 1 θ 2 ε t 2 θ q ε t q ,
where X t denotes the actual value at time period t; φ i ( i = 1 , 2 , , p ) and θ j ( j = 1 , 2 , , q ) are parameters to be estimated; and ε are the random errors, which are assumed to be independently and identically distributed with a mean of zero and a constant variance of δ 2 .

3.2.2. Second Exponential Smoothing

Exponential smoothing is a classical time-series prediction method with a small sample, which is widely used due to its high operability. SES eliminates the random fluctuation in the historical statistical sequence and finds the main trend of its development. It is suitable for simple time series analysis and short- and medium-term prediction. Figure 4 shows the flow chart of the SES model.
Assume that X 0 , X 1 , , X n are the observations of the time series, and S t ( 1 ) , S t ( 2 ) are the first and second exponential smoothing values of the observation, respectively:
S t ( 2 ) = α S t ( 1 ) + ( 1 α ) S t 1 ( 2 ) ,
S t ( 1 ) = α X t + ( 1 α ) S t 1 ( 1 ) .
The prediction formula is expressed by Equation (5):
X ^ t + T = a t + b t T ,
a t = 2 S t ( 1 ) S t ( 2 ) ,
b t = α 1 α ( S t ( 1 ) S t ( 2 ) ) ,
where α is the smoothing factor, and T is the lead time of the forecast.

3.2.3. Support Vector Machine

SVM developed by Cortes C and Vapnik V is the most popular, robust, and widely used intelligence method [61]. SVM exhibits a good performance for dealing with small samples, as well as high-dimensional nonlinear problems. Assume that there is a training sample set { ( x i , y i ) , i = 1 , 2 , , l } , where x i ( x i R d ) with d features, and y i R represents the dependent variables associated with each input. The SVM model tries to formulate a mapping f ( x ) : R d R , where f(x) is expressed as follows:
f ( x ) = w T x + b ,
where w and b are the model parameters to be solved.
It is imperative to adopt a transformation function, ϕ : R d R s ( s > d ) , to all points in the input space for capturing nonlinear data features. The modified mapping function is expressed as follows:
f ( x ) = w ϕ ( x ) + b .
The ε linear insensitive loss function is defined as follows:
L ( f ( x ) , y , ε ) = { 0 , | y f ( x ) | ε | y f ( x ) | ε , | y f ( x ) | > ε ,
where f(x) denotes the forecasting value, and y represents the observed values.
Formally, the SVR model can be expressed as a convex optimal problem as follows:
min 1 2 w 2 + C i = 1 l ( ε i + ε i * ) s . t . { y i w ϕ ( x ) b ε i + ε i * , i = 1 , 2 , , l y i + w ϕ ( x ) + b ε i + ε i * , i = 1 , 2 , , l ε i 0 , ε i * 0 ,
where ε i , ε i * are the slack variables, C is the penalty factor, and ε is the error requirement of the regression function. The above problem in Equation (10) can also be expressed in its dual form as follows:
max [ 1 2 i = 1 l j = 1 l ( a i a i * ) ( a j a j * ) K ( x i , x j ) i = 1 l ( a i + a i * ) ε + i = 1 l ( a i a i * ) y i ] s . t . { i = 1 l ( a i a i * ) = 0 0 a i C 0 a i * C ,
where K ( x i , x j ) = Φ ( x i ) Φ ( x j ) is the kernel function.
By setting the optimal solution of the above equation as a = [ a 1 , a 2 , , a l ] and a * = [ a 1 * , a 2 * , , a l * ] , Equations (13) and (14) can be obtained:
w * = i = 1 l ( a i a i * ) Φ ( x i ) ,
b * = 1 N n s v { 0 < a i < C l [ y i x i S V l ( a i a i * ) K ( x i , x j ) ε ] + 0 < a i < C l [ y i x i S V l ( a j a j * ) K ( x i , x j ) ε ] } ,
where N n s v is the number of support vectors.
Therefore, the regression function is shown in Equation (15):
f ( x ) = w * ϕ ( x ) + b * = i = 1 l ( a i a i * ) Φ ( x i ) ϕ ( x ) + b * = i = 1 l ( a i a i * ) K ( x i , x ) + b * .

3.2.4. Artificial Neural Networks

ANNs has been a research hotspot in the field of artificial intelligence since the 1980s. The ANN abstracts the neural network of a human brain from the perspective of information processing, builds some simple models, and forms different networks according to different connection modes. A neural network is an operational model comprising a large number of nodes (or neurons) connected to each other. Each node represents a specific type of an output function called an activation function. The connection between each node represents a weighted value of the signal passing through the connection, known as weight, which is equivalent to the memory of the ANN. The output of the network varies according to the connection mode, weight value, and excitation function [62,63]. In this paper, three models of BPNN, RBFNN, and GRNN, respectively, are mainly introduced.

3.2.5. Extreme Learning Machine

ELM is a simple efficient algorithm that does not require the tuning of parameters and exhibits extremely rapid learning speeds [64]. The network of an ELM training model is a single-hidden-layer feed-forward neural network structure (Figure 5).
Take N data samples ( x i , y i ) , i = 1 , 2 , , N as an example, where the input data are x i = [ x i , 1 , x i , 2 , , x i , N ] and the output data are y i = [ y i , 1 , y i , 2 , , y i , M ] . The ELM algorithm is as follows:
Assume that single-layer feed-forward neural networks (SLFNs) with L hidden layer nodes and activation function g ( a , b , x ) can fit N data samples with non-errors, indicating that the presence of the parameters a i , b i , β i , i = 1 , 2 , , L makes Equation (16) true:
f L ( x j ) = i = 1 L β i g ( a i , b i , x j ) = y j , j = 1 , 2 , , N .
Equation (16) can be represented by a matrix as follows:
H β = Y [ g ( a 1 , b 1 , x 1 ) g ( a L , b L , x L ) g ( a 1 , b 1 , x N ) g ( a L , b L , x N ) ] N × L [ β 1 T β L T ] L × M = [ Y 1 T Y N T ] N × M ,
where H is the output matrix of the hidden layer, the i-th column represents the output of ( x 1 , , x N ) at the i-th hidden layer node, and the j-th row represents the output of ( x 1 , , x N ) on all hidden layer nodes.
In most cases, the number of the hidden layer nodes, L, is less than the number of samples, N, making it difficult for the constructed neural network to approximate Equation (16). By using E to represent the error between the output of the training sample and the actual output, Equation (18) is obtained:
E = [ e 1 T , , e 1 T ] N × M T .
Combining Equations (17) and (18), Equation (17) can be expressed as Equations (19)–(21):
H β = T ,
j = 1 N t j y j < ε ,
min β H β T ,
β ^ = H + T ,
where ε indicates any small number, which can ensure tj can approach yj infinitely. β ^ is the optimal solution of Equation (22), and H+ is the Moore–Penrose generalized inverse of the hidden layer output matrix, H.

3.3. Artificial Bee Colony Ensemble Algorithm

The ABC algorithm is a random search meta-heuristic global optimization algorithm, which divides worker bees into three categories: Employed bees, onlooker bees, and scout bees [65]. Employed bees and onlooker bees are used for mining honey sources while scout bees search for new honey sources. The location of nectar sources is the result of the optimization problem. First, a group of initial solutions (honey source) are randomly generated from the feasible domain, and all employed bees conduct a domain search. Second, the scout bees randomly search the honey source and record the current best solution. Then, the onlooker bees choose honey sources according to some strategies and conduct a random search to record the best solution. Finally, the honey source that falls into the local optimal point is replaced by the randomly generated honey source, and the global optimal solution is finally obtained after several iterations [66]. Figure 6 shows the schematic of the ABC algorithm.
To better demonstrate the algorithm flow, this paper gives the pseudo-code of the ABC algorithm in the Appendix B Table A4.

4. Empirical Analysis

4.1. Datasets

In this paper, an energy demand series was investigated for forecasting analysis. The primary energy consumption/demand data of China from 1978 to 2017 were collected from the Wind database, BP statistical review of world energy (https://www.bp.com/en/global/corporate/energy-economics/statistical-review-of-world-energy.html), and the China National Bureau of Statistics database (http://www.stats.gov.cn/). All the original data are listed in the Appendix A Table A1. Figure 7 shows the raw demand series and energy demand increase rate. Meanwhile, factors, such as the GDP, industrial structure, energy structure, technological innovation, urbanization rate, population, and CPI, were selected as exogenous variables. All of the series are yearly data from 1978 to 2017 including 40 observations. For the training and testing model, the series wre partitioned into two parts. The observations from 1978 to 2012 were regarded as the training samples, and the remainder were testing samples.

4.2. Error Metric and Statistic Test

The chosen error metric should reflect the forecasting performance of the models from different aspects. In view of this, three indicators were adopted herein: Root mean square error (RMSE) and mean absolute percentage error (MAPE). These indicators have been widely utilized in recent years:
R M S E = 1 n t = 1 n ( y t y ^ t ) 2 ,
M A P E = 1 n t = 1 n | y t y ^ t y t | ,
where n is the size of the predictions, and yt and y ^ t represent the observed value and predicted value, respectively.
Besides, to test the difference in the model prediction performance from the statistical aspect, the Diebold–Mariano (DM) statistical test was introduced to determine whether the prediction accuracy of model A is significantly better than that of model B. The null hypothesis of DM is that the prediction accuracy of model A is no greater than that of model B, that is, the prediction error e A , t = x t x ^ A , t of model A is greater than or equal to the prediction error e B , t = x t x ^ B , t of model B [67]. Accordingly, DM test statistics can be given by Equations (25)–(27):
D M = 1 n t = 1 n g t ( γ 0 + 2 l = 1 γ l ) / n ~ N ( 0 , 1 ) ,
g t = e A , t 2 e B , t 2 ,
γ l = cov ( g t , g t l ) .
The unilateral test of DM statistics can effectively identify the superiority of model A over model B according to the DM statistics and its p value.

4.3. Parameter Settings

To test the superiority of the proposed framework, six forecasting base models and one simple average ensemble model were established and applied as benchmarks. First, the ARIMA and SES models that exhibit better performance for small-sample nonlinear time series were selected. Then, five emerging machine learning models, i.e., SVR, BPNN, GRNN, RBFNN, and ELM models, were adopted.
For the ARIMA model, the autocorrelation function (ACF) and partial autocorrelation function (PACF) are widely used to determine the orders of the autoregressive and moving averages process [68]. ACF measures the correlation between time-series and its lags while PACF are the correlation coefficients between time-series and its lags without the influence of members in between. The lags of the autoregressive process are usually determined by the PACF diagram while the lag value of the moving averages process is determined by the ACF diagram. In this paper, parameters p and q were all set to be 5 according to the ACF and PACF diagrams. As the second-order difference of the original sequence is stable, d was set as 2. As for the SES model, the smoothing factor α was set to 0.7.
For artificial neural networks, it is widely accepted that a feed-forward network with one hidden layer and enough neurons in the hidden layers can fit any finite input-output mapping problem [64]. Therefore, the BPNN was set as a standard three-layer neural network, including an input layer, one hidden layer, and output layer [67]. The number of hidden layer nodes was set to 20 according to Zhao et al. [67] as a small number of hidden nodes results in an inaccurate fitting and too much results in local optimums. There are four layers, i.e., input layer, pattern layer, summation layer, and output layer, respectively, in the GRNN model. The number of nodes in the input layer is equal to the dimensions of the input vector. The spread of the radial basis function in the RBF model was set to 1, and the number of input neurons was equal to the number of columns in the data matrix.
Also according to Zhao et al. [67], Godarzi et al. [69], and Yu et al. [70], the kernel function was set to the Gaussian kernel function, and C and gamma were respectively set as iqr(Y)/1.349 and 1, where iqr(Y) is the interquartile range of the processed target series. Besides, the number of neurons in the hidden layer was equal to the number of training samples in the ELM model. The sigmoid function was selected as the activation function.
For the ABC ensemble algorithm, the size of the bee colony was set to 100, and the maximum number of loop iterations was 1000. All of the mentioned forecasting models were run 20 times by using MATLAB R2016a software: (2016a, MathWorks, Natick, MA, USA). The average results are shown in the following sections.

4.4. Forecasting Error and Statistical Test

In this section, according to the error metric criteria (i.e., RMSE and MAPE), the forecasting error of out-of-sample data for eight benchmark forecasting models and the established model is shown. Then, the DM test was conducted for testing the significance level of the difference between any two models.
Figure 8a plots the actual values of the energy demand (line) and the forecasted values (bar) of each model, and Figure 8b shows the absolute errors of each forecasting model. From the diagram of the prediction fitting curve (Figure 8a), the benchmark models and the proposed ensemble forecasting model are effective, reflecting that the selected models are rational. In addition, the out-of-sample prediction performance of the GRNN and RBFNN models is not very good. Such a conclusion can also be verified from Figure 8b. The absolute errors of GRNN and RBFNN fluctuate considerably. Compared to all benchmark models, the ensemble model proposed herein exhibits the smallest fluctuation of the absolute prediction error (Figure 8b).
For the in-depth analysis of the prediction performance, two evaluation indexes were calculated. Table 4 summarizes the performance of each model, including RMSE and MAPE, as well as the mean value of the statistics.
First, by the comparison of the base prediction models, including SES, ARIMA, SVR, and ANN, the ELM model exhibits the best prediction performance probably because the parameters do not need to be adjusted in the training process for the use of the ELM algorithm, and the unique optimal solution can be obtained by setting the number of neurons in the hidden layer. In addition, the ELM algorithm exhibits advantages of a rapid learning speed and good generalization performance. For the remaining base models, ARIMA exhibits better performance than SES, SVR, BPNN, GRNN, and RBFNN on the part of RMSE and MAPE, reflecting that ARIMA can forecast small-sample time series.
Once the ensemble prediction model is considered, the prediction model (the ensemble model with the ABC algorithm, E_ABC) proposed herein exhibits the best prediction performance. This model exhibits the lowest RMSE and MAPE. The comparison of the two integrated models revealed that E_ABC exhibits a better performance than E_AVE with respect to the criteria of RMSE and MAPE. For example, the RMSE values for E_ABC and E_AVE are 9.46 and 25.52, respectively. The RMSE of model E_ABC is less than half that of model E_AVE. Meanwhile, the MAPE values for E_ABC and E_AVE are 0.21% and 0.51%, respectively. The MAPE of E_AVE is considerably greater than that of E_ABC.
Besides, the prediction performance of the simple average ensemble model (E_AVE) is worse than that of some base prediction models (i.e., ELM and ARIMA model), possibly because the simple average ensemble method ignores the correlation between base models, and the number of base models is relatively small. The ensemble model with the ABC algorithm (E_ABC) makes use of the prediction results of each base model to train the model and to minimize the absolute error. The optimal training results can be obtained without considering the correlation between base models.
Finally, the results revealed that the integrated model E_ABC exhibits the best performance due to its lowest RMSE and MAPE. Hence, the integrated model E_ABC is powerful for energy demand forecasting.
Table 5 summarizes the empirical results from the DM test, as well as the p values of the relevant statistics between any two models. For explanation purposes, in Table 5, the p value in row 2, column 2 is 0.0081 and less than 0.01, indicating that the test rejects the null hypothesis (i.e., there is a significant difference between the forecasted results of ARIMA and SES) at a 99% confidence level.
By focusing on the DM test in Table 5, in the last row, the p values represent the statistical test results of the proposed ensemble model (E_ABC) and other benchmark models. The p values are all less than 0.1, indicating that there is a significant difference between the forecasting results of the integrated model E_ABC and those of eight benchmark models at a 90% confidence level. For example, the p values in row 9, columns 2, 3, 4, and 7 are less than 0.01; hence, the proposed model is better than SES, ARIMA, SVR, and RBFNN at a 99% confidence level. Considering the forecasting accuracy of all models in Table 5, the forecasting accuracy of the integrated model E_ABC is considerably better than that of the above-mentioned eight models from the statistical perspective. Generally, the integrated model E_ABC is confirmed to exhibit good performance for energy demand forecasting according to the error metric and DM statistic test.
Except for the forecasting accuracy comparison with the benchmark models listed in Table 4 and Table 5, this paper compares the experimental results with other related works [30,31,32,33,35]. As presented in Equation (24), the value of MAPE eliminates the magnitude of the predicted target and can directly reflect the prediction accuracy of the model. Table 6 presents the MAPE value of the various energy demand forecasting experiments (including different forecasting target and data set). The MAPE of the E-ABC model is the lowest compared with the others.

4.5. Future Energy Demand Forecasting Results

In this section, the integrated model E_ABC with a good prediction performance was applied to forecast the future energy demand of China. However, the out-of-sample information was required to forecast the energy demand from 2018 to 2022.
According to the 13th Five Year Plan proposed by the State Council, the economic growth rate will be greater than 6.5% by 2020, which is authoritative and representative to a certain extent. Thus, in this study, a GDP growth rate of 6.5% was used to calculate the GDP from 2018 to 2022. With respect to the other influencing factors, all of them belong to small samples; hence, the trend extrapolation method typically performs better than the others. In view of this, we applied the trend extrapolation and rolling regression methods to predict the data of each influencing factor for the next 5 years in this paper (Table 7).
Table 8 summarizes the forecasting results and growth rate for the proposed ensemble model (E_ABC) in this study, and Figure 9 summarizes the growth rate and energy demand, including the actual and forecasted data. The forecasting results revealed that the energy demand of China will maintain a steady growth trend. By 2022, the energy demand of China will reach 3429 million tons of the standard coal equivalent, corresponding to an increase of 9.48% compared to the energy demand in 2017 and an average annual growth rate of 1.897%. With respect to the growth rate of the future energy demand, the forecasted results revealed a steady fluctuation trend: The growth rate increases to 3.02% in 2018 and then decreases to 1.38% from 2019 to 2022. From Figure 9, the growth rate of the energy demand is clearly leveling off gradually.

4.6. Discussion

By comparing the forecasting performance of the two categories models (single model and ensemble model), the ensemble models were verified to perform the best in forecasting energy demand, since the ensemble model achieves the highest forecasting accuracy in terms of MAPE and RMSE, which further supports that the ensemble model can be chosen as a powerful algorithm in forecasting energy demand.
When two ensemble models were compared, the results indicated that the prediction model (E_ABC) proposed herein exhibits the best prediction performance, which exhibited the lowest RMSE and MAPE. Meanwhile, the MAPE values for E_ABC and E_AVE are 0.21% and 0.51%, respectively. The MAPE of E_AVE is considerably greater than that of E_ABC. The ensemble model with the ABC algorithm (E_ABC) makes use of the prediction results of each base model to train the model and to minimize the absolute error.
Furthermore, the forecasting accuracy of the integrated model E_ABC is considerably better than that of the above-mentioned eight models from the statistical perspective. Generally, the integrated model E_ABC is confirmed to exhibit a good performance for energy demand forecasting according to the error metric and DM statistic test.
In summary, according to the above experiments analyzed in Section 4.4 and Section 4.5, three main results can be drawn as follows: (1) The ensemble model established in this study is significantly superior to some other benchmark prediction models just in terms of the forecasting accuracy and hypothesis test; (2) the proposed ensemble approach with the ABC algorithm can be employed as a promising framework for energy demand forecasting in terms of the forecasting accuracy and hypothesis test; and (3) the results for the forecasting of the future energy demand by the ensemble model revealed that the energy demand of China will maintain a steady growth trend.

5. Conclusions and Further Research

The scientific accurate prediction of energy demand is crucial for the rational formulation of energy policies, and also for the basis of ensuring the security of the energy supply. However, it is difficult to accurately predict energy demand because of the complex relationship between the energy demand and a series of influencing factors. In this context, a novel ensemble forecasting framework was proposed, which offers a complement to both traditional and data-driven forecasting techniques.
In this approach, the artificial bee colony algorithm (ABC), a type of a meta-heuristic global optimization algorithm with a random search, was adopted to integrate the monolithic forecast models. As the ABC algorithm combines the advantages of a local deep search and global-wide search, an ensemble model with better predictive performance can be obtained. In the empirical study, the performance of the proposed integrated model was better than those of the benchmark machine learning models (i.e., SES, ARIMA, SVR, ANN, and ELM) and the simple average integrated prediction model in terms of the evaluation criteria (i.e., RMSE, MAPE, and statistic test), indicating that the proposed approach can be used as a promising tool for energy demand forecasting. Besides, the results obtained for the forecasting of the future energy demand by the ensemble model (E_ABC) revealed that the energy demand of China will maintain a steady growth trend. By 2022, the energy demand of China will reach 3429 million tons of the standard coal equivalent, corresponding to an increase of 9.48% compared to the energy demand in 2017 and an average annual growth rate of 1.897%.
Although the proposed integrated model exhibits good prediction performance, there is still considerable room for improvement. Various factors are well known to affect the energy demand, not only those selected herein. If those factors can be considered in the proposed integrated model, the prediction performance may be considerably better. In addition, energy demand forecasting faces more challenges due to the data frequency and data type. Hence, the use of text mining technology to expand data types and broaden the data frequency has become the focus of our future research.

Author Contributions

Conceptualization, J.H. and X.S.; Methodology, J.H. and X.S.; Investigation, J.H. and Q.F.; Data Curation, J.H. and Q.F.; Writing—Original Draft Preparation, J.H., X.S. and Q.F.; Supervision, Funding acquisition, Project Administration, Writing—review and editing, X.S. All authors have read and agreed to the published version of the manuscript.

Acknowledgments

We gratefully acknowledge the financial support from National Natural Science Foundation of China (No. 71771206, 71425002) and the State Grid Corporation of China’s technology project “China Power Technology and Equipment ‘Going Global’ Market Forecasting Model and Analysis”.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Total energy demand and related factors of China between 1978 and 2017.
Table A1. Total energy demand and related factors of China between 1978 and 2017.
YearGDPISESTIURPOPCPIED
Unit108 Yuan///%104/Million Toe
19783645.000.7225.9015.6817.9296,259.00184.00396.6
19794062.000.6925.1014.4219.9997,542.00208.00408.2
19804545.000.7023.8013.2619.3998,705.00238.00417.4
19814892.000.6822.8012.1520.16100,100.00264.00411.6
19825323.000.6721.4011.6621.13101,700.00288.00429.5
19835963.000.6720.5011.0721.62103,000.00316.00456.9
19847208.000.6819.809.8423.01104,400.00361.00490.2
19859098.900.7219.308.4323.71105,851.00446.00529.9
198610,376.200.7319.507.7924.52107,507.00497.00555.3
198712,174.600.7319.107.1225.32109,300.00565.00598.8
198815,180.400.7419.106.1325.81111,026.00714.00643.1
198917,179.700.7519.205.6426.21112,704.00788.00674.6
199018,872.900.7318.705.2326.41114,333.00833.00683.2
199122,005.600.7519.104.7226.94115,823.00932.00718.0
199227,194.500.7819.404.0127.46117,171.001116.00755.6
199335,673.200.8020.103.2527.99119,517.001393.00812.7
199448,637.500.8019.302.5228.51119,850.001833.00862.7
199561,339.900.8019.302.1429.04121,121.001355.00888.8
199671,813.600.8020.501.8830.48122,389.002789.00935.1
199779,715.000.8222.201.7031.91123,626.003002.00940.6
199885,195.500.8222.601.6033.35124,761.003159.00941.6
199990,564.400.8423.501.5534.78125,786.003346.00974.3
2000100,280.100.8524.401.4536.22126,743.003632.001007.9
2001110,863.100.8624.201.3637.66127,627.003869.001064.6
2002121,717.400.8624.701.3139.09128,453.004106.001161.0
2003137,422.000.8723.701.3440.53129,227.004411.001353.5
2004161,840.200.8723.801.3241.76129,988.004925.001583.8
2005187,318.900.8822.401.2642.99130,756.005463.001800.4
2006219,438.500.8922.201.1844.34131,448.006138.001974.7
2007270,232.300.8922.101.0445.89132,129.007081.002147.8
2008319,515.500.8922.000.9146.99132,802.008183.002229.0
2009349,081.400.9019.900.8848.34133,450.009514.002328.1
2010413,030.300.9021.400.7949.95134,091.0010,919.002491.1
2011489,300.600.9021.400.7151.27134,735.0013,134.002690.3
2012540,367.400.9021.800.7452.57135,404.0014,699.002797.4
2013595,244.400.9122.400.7053.73136,072.0016,190.002905.3
2014643,974.000.9123.100.6654.77136,782.0017,778.002970.6
2015689,052.100.9224.200.6256.10137,462.0019,397.003005.9
2016744,127.200.9224.700.5957.35138,271.0021,228.003053.0
2017827,121.700.9225.120.5458.52139,008.0022,902.003132
Note: Gross Domestic Product (GDP); Industrial Structure (IS); Energy Structure (ES); Technological Innovation (TI); Urbanization rate (UR); Population (Pop); Consumer Price Index (CPI); Energy Demand (ED); Ton Oil Equivalent (toe).
Table A2. The preprocessed data.
Table A2. The preprocessed data.
YearGDPISESTIURPOPCPIED
Training Set19780.000.201.001.000.000.000.000.00
19790.000.080.890.920.050.030.000.00
19800.000.120.710.840.040.060.000.01
19810.000.060.570.770.060.090.000.01
19820.000.000.380.730.080.130.000.01
19830.000.010.250.700.090.160.010.02
19840.000.050.150.610.130.190.010.03
19850.010.190.080.520.140.220.010.05
19860.010.240.110.480.160.260.010.06
19870.010.250.060.430.180.310.020.07
19880.010.300.060.370.190.350.020.09
19890.020.320.070.340.200.380.030.10
19900.020.240.000.310.210.420.030.10
19910.020.340.060.280.220.460.030.12
19920.030.450.100.230.230.490.040.13
19930.040.530.190.180.250.540.050.15
19940.050.520.080.130.260.550.070.17
19950.070.520.080.110.270.580.050.18
19960.080.530.250.090.310.610.110.20
19970.090.590.490.080.340.640.120.20
19980.100.610.540.070.380.670.130.20
19990.110.660.670.070.420.690.140.21
20000.120.710.790.060.450.710.150.22
20010.130.740.760.050.490.730.160.24
20020.140.760.830.050.520.750.170.28
20030.160.800.690.050.560.770.190.35
20040.190.780.710.050.590.790.210.43
20050.220.820.510.050.620.810.230.51
20060.260.860.490.040.650.820.260.58
20070.320.880.470.030.690.840.300.64
20080.380.880.460.020.720.850.350.67
20090.420.890.170.020.750.870.410.71
20100.500.900.380.020.790.880.470.77
20110.590.900.380.010.820.900.570.84
20120.650.900.430.010.850.920.640.88
Testing Set20130.720.930.510.010.880.930.700.92
20140.780.960.610.010.910.950.770.94
20150.830.970.760.010.940.960.850.95
20160.900.980.830.000.970.980.930.97
20171.001.000.890.001.001.001.001.00

Appendix B

Table A3. Pseudo-code of the ensemble forecasting model.
Table A3. Pseudo-code of the ensemble forecasting model.
Pseudo-Code of the Ensemble Forecasting Model
Input:training data set D = { X , Y } , the number of base model T
Output:ensemble forecasting model H
Pseudo-code
1:Step 1: training the base model
2:For t = 1 to T do
3:Training the forecasting model t with training data set
4:Get the fitting value of model t,
5:End
6:Step2: find the optimal weights for based models
7:Initialize the weights of based models { w 1 ,   w 2 , , w T }
8:Find the optimal base model weights by minimizing the error function J with ABC:
9: J = t = 1 T w t Y ^ t Y
10:Step3 get the ensemble forecasting model H:
11:Make predictions using the final model: Y ^ = t = 1 T w * Y ^ t
Table A4. Pseudo-code of the ensemble forecasting model.
Table A4. Pseudo-code of the ensemble forecasting model.
Pseudo-Code of the Artificial Bee Colony Algorithm
Input:int N; // the size of the bee colony
int M; // the maximum number of cycles
int L; // the control parameter to abandon the nectar source
Output:set X*; // the optimal solution set
Pseudo-code
1:initialize the population of solutions
2:evaluate the population
3:cycle = 1
4:Repeat
5:send employed bees
6:calculate the probability of the current nectar sources
7:send onlooker
8:send scout
9:memorize the best nectar sources position achieved currently
10:cycle = cycle + 1
11:until cycle = M

References

  1. Le, T.-H.; Nguyen, C.P. Is energy security a driver for economic growth? Evidence from a global sample. Energy Policy 2019, 129, 436–451. [Google Scholar] [CrossRef]
  2. Sharimakin, A.; Glass, A.J.; Saal, D.S.; Glass, K. Dynamic multilevel modelling of industrial energy demand in Europe. Energy Econ. 2018, 74, 120–130. [Google Scholar] [CrossRef] [Green Version]
  3. Ji, Q.; Zhang, H.; Zhang, D. The impact of OPEC on East Asian oil import security: A multidimensional analysis. Energy Policy 2019, 126, 99–107. [Google Scholar] [CrossRef]
  4. Cohen, E. Development of Israel’s natural gas resources: Political, security, and economic dimensions. Resour. Policy 2018, 57, 137–146. [Google Scholar] [CrossRef]
  5. Wang, Q.; Li, S.; Li, R. Forecasting energy demand in China and India: Using single-linear, hybrid-linear, and non-linear time series forecast techniques. Energy 2018, 161, 821–831. [Google Scholar] [CrossRef]
  6. Ji, Q.; Bouri, E.; Roubaud, D.; Kristoufek, L. Information interdependence among energy, cryptocurrency and major commodity markets. Energy Econ. 2019, 81, 1042–1055. [Google Scholar] [CrossRef]
  7. Hribar, R.; Potočnik, P.; Šilc, J.; Papa, G. A comparison of models for forecasting the residential natural gas demand of an urban area. Energy 2019, 167, 511–522. [Google Scholar] [CrossRef]
  8. He, Y.; Lin, B. Forecasting China’s total energy demand and its structure using ADL-MIDAS model. Energy 2018, 151, 420–429. [Google Scholar] [CrossRef]
  9. Wang, Z.-X.; Li, Q.; Pei, L.-L. Grey forecasting method of quarterly hydropower production in China based on a data grouping approach. Appl. Math. Model. 2017, 51, 302–316. [Google Scholar] [CrossRef]
  10. Bates, J.M.; Granger, C.W.J. The Combination of Forecasts. OR 1969, 20, 451–468. [Google Scholar] [CrossRef]
  11. Samuels, J.D.; Sekkel, R.M. Model Confidence Sets and forecast combination. Int. J. Forecast. 2017, 33, 48–60. [Google Scholar] [CrossRef]
  12. Hsiao, C.; Wan, S.K. Is there an optimal forecast combination? J. Econ. 2014, 178, 294–309. [Google Scholar] [CrossRef] [Green Version]
  13. Bedi, J.; Toshniwal, D. Deep learning framework to forecast electricity demand. Appl. Energy 2019, 238, 1312–1326. [Google Scholar] [CrossRef]
  14. Kim, T.-Y.; Cho, S.-B. Predicting residential energy consumption using CNN-LSTM neural networks. Energy 2019, 182, 72–81. [Google Scholar] [CrossRef]
  15. Wei, L.; Li, G.; Zhu, X.; Sun, X.; Li, J. Developing a hierarchical system for energy corporate risk factors based on textual risk disclosures. Energy Econ. 2019, 80, 452–460. [Google Scholar] [CrossRef]
  16. Kim, K.; Hur, J. Weighting Factor Selection of the Ensemble Model for Improving Forecast Accuracy of Photovoltaic Generating Resources. Energies 2019, 12, 3315. [Google Scholar] [CrossRef] [Green Version]
  17. Cui, J.; Yu, R.; Zhao, D.; Yang, J.; Ge, W.; Zhou, X. Intelligent load pattern modeling and denoising using improved variational mode decomposition for various calendar periods. Appl. Energy 2019, 247, 480–491. [Google Scholar] [CrossRef]
  18. Ji, Q.; Zhang, D. How much does financial development contribute to renewable energy growth and upgrading of energy structure in China? Energy Policy 2019, 128, 114–124. [Google Scholar] [CrossRef]
  19. Ji, Q.; Li, J.; Sun, X. Measuring the interdependence between investor sentiment and crude oil returns: New evidence from the CFTC’s disaggregated reports. Financ. Res. Lett. 2019, 30, 420–425. [Google Scholar] [CrossRef]
  20. Ma, Y.-R.; Ji, Q.; Pan, J. Oil financialization and volatility forecast: Evidence from multidimensional predictors. J. Forecast. 2019, 38, 564–581. [Google Scholar]
  21. Berk, I.; Ediger, V.Ş. Forecasting the coal production: Hubbert curve application on Turkey’s lignite fields. Resour. Policy 2016, 50, 193–203. [Google Scholar] [CrossRef]
  22. Mehmanpazir, F.; Khalili-Damghani, K.; Hafezalkotob, A. Modeling steel supply and demand functions using logarithmic multiple regression analysis (case study: Steel industry in Iran). Resour. Policy 2019, 63, 101409. [Google Scholar] [CrossRef]
  23. Wang, X.; Lei, Y.; Ge, J.; Wu, S. Production forecast of China’s rare earths based on the Generalized Weng model and policy recommendations. Resour. Policy 2015, 43, 11–18. [Google Scholar] [CrossRef] [Green Version]
  24. Wang, C.; Zhang, H.; Fan, W.; Ma, P. A new chaotic time series hybrid prediction method of wind power based on EEMD-SE and full-parameters continued fraction. Energy 2017, 138, 977–990. [Google Scholar] [CrossRef]
  25. Yu, L.; Wang, S.; Lai, K.K. Forecasting crude oil price with an EMD-based neural network ensemble learning paradigm. Energy Econ. 2008, 30, 2623–2635. [Google Scholar] [CrossRef]
  26. Tang, L.; Wu, Y.; Yu, L.A. A randomized-algorithm-based decomposition-ensemble learning methodology for energy price forecasting. Energy 2018, 157, 526–538. [Google Scholar] [CrossRef]
  27. Kim, S.; Lee, G.; Kwon, G.-Y.; Kim, D.-I.; Shin, Y.-J. Deep Learning Based on Multi-Decomposition for Short-Term Load Forecasting. Energies 2018, 11, 3433. [Google Scholar] [CrossRef] [Green Version]
  28. Tso, G.K.F.; Yau, K.K.W. Predicting electricity energy consumption: A comparison of regression analysis, decision tree and neural networks. Energy 2007, 32, 1761–1768. [Google Scholar] [CrossRef]
  29. Adom, P.K.; Bekoe, W. Conditional dynamic forecast of electrical energy consumption requirements in Ghana by 2020: A comparison of ARDL and PAM. Energy 2012, 44, 367–380. [Google Scholar] [CrossRef]
  30. Ghanbari, A.; Kazemi, S.M.R.; Mehmanpazir, F.; Nakhostin, M.M. A Cooperative Ant Colony Optimization-Genetic Algorithm approach for construction of energy demand forecasting knowledge-based expert systems. Knowl.-Based Syst. 2013, 39, 194–206. [Google Scholar] [CrossRef]
  31. Wu, Q.; Peng, C. A hybrid BAG-SA optimal approach to estimate energy demand of China. Energy 2017, 120, 985–995. [Google Scholar] [CrossRef]
  32. Yu, S.-W.; Zhu, K.-J. A hybrid procedure for energy demand forecasting in China. Energy 2012, 37, 396–404. [Google Scholar] [CrossRef]
  33. Karadede, Y.; Ozdemir, G.; Aydemir, E. Breeder hybrid algorithm approach for natural gas demand forecasting model. Energy 2017, 141, 1269–1284. [Google Scholar] [CrossRef]
  34. Angelopoulos, D.; Siskos, Y.; Psarras, J. Disaggregating time series on multiple criteria for robust forecasting: The case of long-term electricity demand in Greece. Eur. J. Oper. Res. 2019, 275, 252–265. [Google Scholar] [CrossRef]
  35. Piltan, M.; Shiri, H.; Ghaderi, S.F. Energy demand forecasting in Iranian metal industry using linear and nonlinear models based on evolutionary algorithms. Energy Convers. Manag. 2012, 58, 1–9. [Google Scholar] [CrossRef]
  36. Sonmez, M.; Akgüngör, A.P.; Bektaş, S. Estimating transportation energy demand in Turkey using the artificial bee colony algorithm. Energy 2017, 122, 301–310. [Google Scholar] [CrossRef]
  37. Yuan, X.-C.; Sun, X.; Zhao, W.; Mi, Z.; Wang, B.; Wei, Y.-M. Forecasting China’s regional energy demand by 2030: A Bayesian approach. Resour. Conserv. Recycl. 2017, 127, 85–95. [Google Scholar] [CrossRef]
  38. He, Y.; Zheng, Y.; Xu, Q. Forecasting energy consumption in Anhui province of China through two Box-Cox transformation quantile regression probability density methods. Measurement 2019, 136, 579–593. [Google Scholar] [CrossRef]
  39. He, Y.X.; Liu, Y.Y.; Xia, T.; Zhou, B. Estimation of demand response to energy price signals in energy consumption behaviour in Beijing, China. Energy Convers. Manag. 2014, 80, 429–435. [Google Scholar] [CrossRef]
  40. Forouzanfar, M.; Doustmohammadi, A.; Hasanzadeh, S.; Shakouri, G.H. Transport energy demand forecast using multi-level genetic programming. Appl. Energy 2012, 91, 496–503. [Google Scholar] [CrossRef]
  41. Liao, H.; Cai, J.-W.; Yang, D.-W.; Wei, Y.-M. Why did the historical energy forecasting succeed or fail? A case study on IEA’s projection. Technol. Forecast. Soc. Chang. 2016, 107, 90–96. [Google Scholar] [CrossRef]
  42. Ahmad, T.; Chen, H. Deep learning for multi-scale smart energy forecasting. Energy 2019, 175, 98–112. [Google Scholar] [CrossRef]
  43. Schaer, O.; Kourentzes, N.; Fildes, R. Demand forecasting with user-generated online information. Int. J. Forecast. 2019, 35, 197–212. [Google Scholar] [CrossRef] [Green Version]
  44. Xu, S.; Chan, H.K.; Zhang, T. Forecasting the demand of the aviation industry using hybrid time series SARIMA-SVR approach. Transp. Res. Part E Logist. Transp. Rev. 2019, 122, 169–180. [Google Scholar] [CrossRef]
  45. Ünler, A. Improvement of energy demand forecasts using swarm intelligence: The case of Turkey with projections to 2025. Energy Policy 2008, 36, 1937–1944. [Google Scholar] [CrossRef]
  46. Debnath, K.B.; Mourshed, M. Forecasting methods in energy planning models. Renew. Sustain. Energy Rev. 2018, 88, 297–325. [Google Scholar] [CrossRef] [Green Version]
  47. Yuan, X.; Chen, C.; Yuan, Y.; Huang, Y.; Tan, Q. Short-term wind power prediction based on LSSVM–GSA model. Energy Convers. Manag. 2015, 101, 393–401. [Google Scholar] [CrossRef]
  48. Zendehboudi, A.; Baseer, M.A.; Saidur, R. Application of support vector machine models for forecasting solar and wind energy resources: A review. J. Clean. Prod. 2018, 199, 272–285. [Google Scholar] [CrossRef]
  49. Wei, D.; Wang, J.; Ni, K.; Tang, G. Research and Application of a Novel Hybrid Model Based on a Deep Neural Network Combined with Fuzzy Time Series for Energy Forecasting. Energies 2019, 12, 3588. [Google Scholar] [CrossRef] [Green Version]
  50. Law, A.; Ghosh, A. Multi-label classification using a cascade of stacked autoencoder and extreme learning machines. Neurocomputing 2019, 358, 222–234. [Google Scholar] [CrossRef]
  51. Bergmeir, C.; Hyndman, R.J.; Benitez, J.M. Bagging exponential smoothing methods using STL decomposition and Box-Cox transformation. Int. J. Forecast. 2016, 32, 303–312. [Google Scholar] [CrossRef] [Green Version]
  52. Han, X.; Li, R. Comparison of Forecasting Energy Consumption in East Africa Using the MGM, NMGM, MGM-ARIMA, and NMGM-ARIMA Model. Energies 2019, 12, 3278. [Google Scholar] [CrossRef] [Green Version]
  53. Zhao, W.; Zhao, J.; Yao, X.; Jin, Z.; Wang, P. A Novel Adaptive Intelligent Ensemble Model for Forecasting Primary Energy Demand. Energies 2019, 12, 1347. [Google Scholar] [CrossRef] [Green Version]
  54. Galicia, A.; Talavera-Llames, R.; Troncoso, A.; Koprinska, I.; Martínez-Álvarez, F. Multi-step forecasting for big data time series based on ensemble learning. Knowl.-Based Syst. 2019, 163, 830–841. [Google Scholar] [CrossRef]
  55. Liu, L.; Zong, H.; Zhao, E.; Chen, C.; Wang, J. Can China realize its carbon emission reduction goal in 2020: From the perspective of thermal power development. Appl. Energy 2014, 124, 199–212. [Google Scholar] [CrossRef]
  56. Wang, J.-J.; Wang, J.-Z.; Zhang, Z.-G.; Guo, S.-P. Stock index forecasting based on a hybrid model. Omega 2012, 40, 758–766. [Google Scholar] [CrossRef]
  57. Zhu, B.; Wei, Y. Carbon price forecasting with a novel hybrid ARIMA and least squares support vector machines methodology. Omega 2013, 41, 517–524. [Google Scholar] [CrossRef]
  58. Qu, Z.; Zhang, K.; Mao, W.; Wang, J.; Liu, C.; Zhang, W. Research and application of ensemble forecasting based on a novel multi-objective optimization algorithm for wind-speed forecasting. Energy Convers. Manag. 2017, 154, 440–454. [Google Scholar] [CrossRef]
  59. Xue, Y.; Jiang, J.; Zhao, B.; Ma, T. A self-adaptive artificial bee colony algorithm based on global best for global optimization. Soft Comput. 2017, 22, 2935–2952. [Google Scholar] [CrossRef]
  60. Xu, X.; Wei, Z.; Ji, Q.; Wang, C.; Gao, G. Global renewable energy development: Influencing factors, trend predictions and countermeasures. Resour. Policy 2019, 63, 101470. [Google Scholar] [CrossRef]
  61. Shi, Y.; Tian, Y.; Kou, G.; Peng, Y.; Li, J. Optimization Based Data Mining Theory and Applications; Springer: London, UK, 2011. [Google Scholar]
  62. Wang, C.; Zhang, H.; Fan, W.; Fan, X. A new wind power prediction method based on chaotic theory and Bernstein Neural Network. Energy 2016, 117, 259–271. [Google Scholar] [CrossRef]
  63. Qureshi, A.S.; Khan, A.; Zameer, A.; Usman, A. Wind power prediction using deep neural network based meta regression and transfer learning. Appl. Soft Comput. 2017, 58, 742–755. [Google Scholar] [CrossRef]
  64. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  65. Pérez, C.J.; Vega-Rodríguez, M.A.; Reder, K.; Flörke, M. A Multi-Objective Artificial Bee Colony-based optimization approach to design water quality monitoring networks in river basins. J. Clean. Prod. 2017, 166, 579–589. [Google Scholar] [CrossRef]
  66. Lin, Y.; Jia, H.; Yang, Y.; Tian, G.; Tao, F.; Ling, L. An improved artificial bee colony for facility location allocation problem of end-of-life vehicles recovery network. J. Clean. Prod. 2018, 205, 134–144. [Google Scholar] [CrossRef]
  67. Zhao, Y.; Li, J.; Yu, L. A deep learning ensemble approach for crude oil price forecasting. Energy Econ. 2017, 66, 9–16. [Google Scholar] [CrossRef]
  68. Peimankar, A.; Weddell, S.J.; Jalal, T.; Lapthorn, A.C. Multi-objective ensemble forecasting with an application to power transformers. Appl. Soft Comput. 2018, 68, 233–248. [Google Scholar] [CrossRef]
  69. Godarzi, A.A.; Amiri, R.M.; Talaei, A.; Jamasb, T. Predicting oil price movements: A dynamic Artificial Neural Network approach. Energy Policy 2014, 68, 371–382. [Google Scholar] [CrossRef] [Green Version]
  70. Yu, L.; Zhao, Y.; Tang, L. A compressed sensing based AI learning paradigm for crude oil price forecasting. Energy Econ. 2014, 46, 236–245. [Google Scholar] [CrossRef]
Figure 1. Ensemble framework of energy demand forecast.
Figure 1. Ensemble framework of energy demand forecast.
Energies 13 00550 g001
Figure 2. Correlation plot between the energy demand and selected factors.
Figure 2. Correlation plot between the energy demand and selected factors.
Energies 13 00550 g002
Figure 3. The trend extrapolation method of energy demand influencing factors.
Figure 3. The trend extrapolation method of energy demand influencing factors.
Energies 13 00550 g003
Figure 4. Flow chart of the SES model.
Figure 4. Flow chart of the SES model.
Energies 13 00550 g004
Figure 5. Structure of the single-hidden-layer feed-forward neural network.
Figure 5. Structure of the single-hidden-layer feed-forward neural network.
Energies 13 00550 g005
Figure 6. Schematic of the artificial bee colony algorithm.
Figure 6. Schematic of the artificial bee colony algorithm.
Energies 13 00550 g006
Figure 7. Primary energy demand time series from 1978 to 2017.
Figure 7. Primary energy demand time series from 1978 to 2017.
Energies 13 00550 g007
Figure 8. Forecasting results (a) and absolute errors (b) of each forecasting model.
Figure 8. Forecasting results (a) and absolute errors (b) of each forecasting model.
Energies 13 00550 g008
Figure 9. Forecasting results and the growth rate of the future energy demand.
Figure 9. Forecasting results and the growth rate of the future energy demand.
Energies 13 00550 g009
Table 1. Influencing factors of energy demand.
Table 1. Influencing factors of energy demand.
LiteraturesForecast RangeCase StudyFactors
Hribar, Potočnik, Šilc, and Papa [7]Daily/CityNatural gas demandWeather, socioeconomic factors
Adom and Bekoe [29]Yearly/CountryElectrical demandMacro socioeconomic indicators, population, gross national product, degree of urbanization, industry efficiency, industry value-added
Ghanbari, Kazemi, Mehmanpazir, and Nakhostin [30]Yearly/CountryEnergy demandPopulation, GDP, CPI, imports, exports, energy intensity, energy efficiency
Wu and Peng [31]Yearly/CountryEnergy demandEconomic growth, total population, fixed-asset investments, energy efficiency, energy structure, household energy consumption per capita
Yu and Zhu [32]Yearly/CountryEnergy demandEconomic growth, total population, economic structure, urbanization rate, energy structure, energy price
Karadede et al. [33]Yearly/CountryNatural gas demandGross national product, population, growth rate
Angelopoulos et al. [34]Yearly/CountryElectricity demandGDP, unemployment rate, population, weather-related criteria, electricity price, energy efficiency criterion
Piltan et al. [35]Yearly/CountryEnergy demandNumber of employees, investment value, gas price, electricity price
Sonmez et al. [36]Yearly/CountryTransport energy demandGross domestic product, population
Yuan et al. [37]Yearly/CountryEnergy demandEconomic level, industrial structure, demographic change, urbanization process, technological progress
He et al. [38]Yearly/ProvinceEnergy demandHistorical energy consumption, population, GDP growth rate, total GDP, the three major industrial GDP, CPI
He et al. [39]Yearly/CityEnergy demandPopulation, GDP growth rate, GDP, the three major industrial GDP, CPI
Forouzanfar et al. [40]Yearly/CountryTransport energy demandPopulation, gross domestic product, number of vehicles
Liao et al. [41]Yearly/CountryEnergy demandGDP, population, oil price
Table 2. Comparison of different forecasting models for energy demand.
Table 2. Comparison of different forecasting models for energy demand.
ModelsTrend FeaturesForecast PeriodNumber of Variables
LinearNonlinearLongShortMultipleSingle
Regression-based
ARIMA
SVM
ANN
ELM
Note: symbol ✓ represents the superiority of forecasting performance.
Table 3. Selected factor of the energy demand forecasting model.
Table 3. Selected factor of the energy demand forecasting model.
No.Factorsrp
1Gross domestic product (GDP)0.9650.0000
2Industrial structure (IS)0.8870.0000
3Energy structure (ES)0.3280.0387
4Technological innovation (TI)−0.6970.0000
5Urbanization rate (UR)0.9780.0000
6Population (Pop)0.8730.0000
7Consumer price index (CPI)0.9590.0000
8Energy demand (ED)1.000——
Note: Correlation in r is considered significant if the off-diagonal elements of p are less than the significance level (0.05).
Table 4. Forecasting accuracy of all models.
Table 4. Forecasting accuracy of all models.
ModelRMSEMAPE
SES46.681.41%
ARIMA18.310.51%
SVR101.903.31%
BPNN56.421.76%
GRNN152.154.62%
RBFNN70.281.51%
ELM16.570.45%
E_AVE25.520.51%
E_ABC9.460.21%
Note: The boldface value represents the best performance among the seven models in terms of RMSE and MAPE.
Table 5. DM statistical test result.
Table 5. DM statistical test result.
SESARIMASVRBPNNGRNNRBFNNELME_AVE
ARIMA0.00811
SVR0.865520.99625
BPNN0.007570.114780.00005
GRNN0.690700.889840.450130.95854
RBFNN0.994070.999770.979850.999910.93608
ELM0.007820.120050.000310.554760.068430.00012
E_AVE0.002830.026310.000050.270890.050170.000050.16362
E_ABC0.00071 ***0.00123 ***0.00005 ***0.09451 *0.04439 **0.00005 ***0.01202 **0.01955 **
Note that (*), (**), and (***) represent rejection of the null hypothesis at confidence levels of 90%, 95%, and 99%, respectively.
Table 6. MAPE comparison with other forecasting models.
Table 6. MAPE comparison with other forecasting models.
ArticleModelTime RangeTarget/CountryMAPE (%)
[30]Cooperative-ACO-GA1971–2007Oil/Turkey0.72
Cooperative-ACO-GANatural gas/Turkey1.18
Cooperative-ACO-GAElectricity/Turkey0.90
[31]BAG-SA1990–2015Energy/China0.28
[32]PSO-GA-linear1980–2009Energy/China0.45
PSO-GA-exponential0.95
PSO-GA-quadratic0.29
[33]BGA2001–2014Natural gas/Turkey1.88
BGA-SA1.43
[35]PSO1980–2003electricity consumption/Turkey5.24
Real coding GA5.93
Present studyE_ABC1978–2017Energy/China0.21
Note: GA indicates Genetic Algorithm; PSO denotes Particle Swarm Optimization; ACO stands for Ant Colony Optimization; BAG denotes Bat Algorithm, Gaussian Perturbations; SA indicates Simulated Annealing; BGA indicates Breeder Genetic Algorithm.
Table 7. Prediction results of each input variable under ARIMA.
Table 7. Prediction results of each input variable under ARIMA.
YearGDP (108 Yuan)IS (/)ES (/)TI (/)UR (%)POP (104)CPI (/)
2018900,3090.926725.340.528159.71139,99124,602.38
2019958,8290.932125.470.434360.75140,97126,353.20
20201,021,1530.936325.380.332261.88142,02228,016.92
20211,087,5280.942825.110.217363.02143,14929,666.45
20221,158,2170.951924.720.087564.15144,34731,267.32
Table 8. Forecasting results of the future energy demand by the E_ABC model.
Table 8. Forecasting results of the future energy demand by the E_ABC model.
YearEnergy Demand (Million Toe)Increase Rate
20173132--
201832263.02%
201932741.47%
202033321.78%
202133821.50%
202234291.38%
Total--9.48%

Share and Cite

MDPI and ACS Style

Hao, J.; Sun, X.; Feng, Q. A Novel Ensemble Approach for the Forecasting of Energy Demand Based on the Artificial Bee Colony Algorithm. Energies 2020, 13, 550. https://doi.org/10.3390/en13030550

AMA Style

Hao J, Sun X, Feng Q. A Novel Ensemble Approach for the Forecasting of Energy Demand Based on the Artificial Bee Colony Algorithm. Energies. 2020; 13(3):550. https://doi.org/10.3390/en13030550

Chicago/Turabian Style

Hao, Jun, Xiaolei Sun, and Qianqian Feng. 2020. "A Novel Ensemble Approach for the Forecasting of Energy Demand Based on the Artificial Bee Colony Algorithm" Energies 13, no. 3: 550. https://doi.org/10.3390/en13030550

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop