Next Article in Journal
Participation of Aggregated DERs to the Ancillary Services Market: A Monte Carlo Simulation-Based Heuristic Greedy-Indexing Model
Previous Article in Journal
Mutual Diffusion of Model Acceptor/Donor Bilayers under Solvent Vapor Annealing as a Novel Route for Organic Solar Cell Fabrication
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forecasting Building Energy Consumption Using Ensemble Empirical Mode Decomposition, Wavelet Transformation, and Long Short-Term Memory Algorithms

1
Taiwan Building Technology Center, National Taiwan University of Science and Technology, Taipei 106, Taiwan
2
Department of Industrial Management, National Taiwan University of Science and Technology, Taipei 106, Taiwan
3
Department of Logistics Engineering, Universitas Pertamina, Jakarta 12220, Indonesia
*
Author to whom correspondence should be addressed.
Energies 2022, 15(3), 1035; https://doi.org/10.3390/en15031035
Submission received: 21 November 2021 / Revised: 3 January 2022 / Accepted: 6 January 2022 / Published: 29 January 2022
(This article belongs to the Topic Exergy Analysis and Its Applications)

Abstract

:
A building, a central location of human activities, is equipped with many devices that consume a lot of electricity. Therefore, predicting the energy consumption of a building is essential because it helps the building management to make better energy management policies. Thus, predicting energy consumption of a building is very important, and this study proposes a forecasting framework for energy consumption of a building. The proposed framework combines a decomposition method with a forecasting algorithm. This study applies two decomposition algorithms, namely the empirical mode decomposition and wavelet transformation. Furthermore, it applies the long short term memory algorithm to predict energy consumption. This study applies the proposed framework to predict the energy consumption of 20 buildings. The buildings are located in different time zones and have different functionalities. The experiment results reveal that the best forecasting algorithm applies the long short term memory algorithm with the empirical mode decomposition. In addition to the proposed framework, this research also provides the recommendation of the forecasting model for each building. The result of this study could enrich the study about the building energy forecasting approach. The proposed framework also can be applied to the real case of electricity consumption.

1. Introduction

Energy is one of the basic needs of human life. In the last few decades, energy consumption has increased significantly due to several factors, such as a population increase, indoor human activities, an increasing number of buildings, and global climate change. As most human activities are conducted in buildings, it consumes most of the electricity. In the US and the European Union, 40% of energy consumption comes from buildings [1].
Buildings can be categorized into several types based on their function: industry, transportation, housing or residential, commercial, public services, agriculture, fisheries, and other sectors. Among all of these types, housing/residential buildings such as housing complexes and apartments consume up to 30% of the total energy buildings worldwide [2]. In some countries, the energy consumption of commercial buildings such as offices, hospitals, shopping centers, restaurants, warehouses, and others is relatively high because they use many devices continuously [3]. The EIA [4] predicts that energy consumption in buildings will grow by an average of 1.3% per year from 2018 to 2050.
Although energy consumption is predicted to increase continuously, more accurate energy demand forecasting is essential. An accurate energy demand forecast will be essential information to make a decision in many different fields, including scheduling, operations, monitoring, and others [5]. Forecasting has an important role in different perspectives for future energy development regionally, nationally, and globally [6]. From the perspective of city stakeholders, it is helpful to understand energy consumption and its variations across types of urban buildings and their user profiles. Energy allocation based on an accurate prediction can improve building policy [7]. Accurate energy consumption forecasting can also avoid overestimating and underestimating resource allocation [8]. From the building owner’s or operator’s point of view, energy forecasting is essential for designing daily operational rules for the building [9]. For the power company and government, city-scale energy consumption predictions help to optimize the electricity distribution and production scenario [9]. Therefore, energy management with accurate forecasting is needed because accuracy allows management to allocate energy needs due to the significant demand to avoid unexpected power outages and reduce operational costs [10]. The primary purpose of this study is to propose a forecasting framework for energy consumption.
Forecasting itself can be classified based on the timespan; short-term, mid-term, and long-term [11]. For example, short-term forecasting with a time span of hourly data for one week [12] will have higher accuracy than mid-term and long-term forecasting. It helps the building management in operation [13]. It also can improve the automation of the energy system of a building [14] at the management level.
Many studies accommodate this problem to improve an accurate energy analysis system, one of which is a research work that forecasts electrical energy consumption in buildings. However, several factors become challenges for researchers in making accurate forecasts [5]. The challenging factors include the limitation of the physical method which requires high computational complexity. On the other hand, the statistics methods cannot perform well for nonlinear hidden feature data.
The proposed forecasting framework combines a time series decomposition method with recurrent neural network algorithms to obtain an accurate forecasting model. There are two decomposition algorithms utilized in this study. They are the ensemble empirical mode decomposition (EEMD) and wavelet transformation (WT). Some previous studies have applied forecasting energy demand using the long short-term memory (LSTM) algorithm [15,16,17]. Peng, Wang, Xia, and Gao [15] and Somu, MR, and Ramamritham [17] applied the LSTM with WT. On the other hand, Gao, Ruan, and Buildings [16] enhanced the LSTM with a feature attention process. These studies show that the LSTM algorithm is a promising algorithm for energy consumption prediction. However, further improvements should be made to get a better result with a high generalization. Thus, this study proposes a forecasting framework, which consists of data preprocessing, data decomposition using EEMD and WT, and the aggregation process for the final prediction.
This study evaluates the proposed forecasting framework’s performance using six benchmark datasets and a real study case of electricity consumption of a university building in Taiwan. Each dataset consists of several buildings. Thus, a total of 20 buildings are evaluated in this study.

2. Literature Review

This section reviews some fundamental theories applied in this research. It includes time series forecasting, decomposition algorithms, and recurrent neural network.

2.1. Time Series Forecasting

Forecasting is the process of predicting future events [11]. Runge and Zmeureanu [18] define forecasting as the process of estimating a value derived from current values or historical data to determine the value that will be generated in the future. In forecasting energy consumption, the data used are time-series data. Electrical energy consumption data are time-series data observed based on the same time interval [19]. Therefore, forecasting electrical energy consumption is categorized as time-series data forecasting. Its goal is to predict events that will occur in the future based on a series of events or history [5].
Problems in forecasting are generally packaged in the form of a time period, such as days, weeks, or months [11]. Based on the time horizon, forecasting is divided into several categories [20], those are:
1.
Short-term
The coverage period of this forecast is hours, days, or weeks ahead. This type of forecasting manages daily operational activities, purchase planning, and evaluation [21].
2.
Medium-term
The scope of the forecasting period is 1–2 years in the future [11]. This forecasting category is very useful for the strategic planning of demand [21].
3.
Long-term
The coverage period of this forecast is about 1–50 years in the future. It is commonly used in a planning system or energy plant installation [21].
Many techniques have been used for forecasting data on electricity consumption in buildings. To obtain an accurate forecast, the methodology should consider data complexity, including weather conditions, lighting systems, and other systems that involve the use of electricity [22]. According to Deb, Zhang, Yang, Lee, and Shah [22], there are two categories of forecasting techniques: the data-driven technique and the deterministic technique. Table 1 shows several advantages and disadvantages of both techniques.
Bourdeau, et al. [23] divide energy forecasting methods in buildings into three categories, namely physical (white box), data-driven (black box), and hybrid (gray box). The data-driven (black box) method uses machine learning and statistics to build and model energy in buildings [23]. The physical method (white box) can be interpreted the same as the deterministic method above. In contrast, the hybrid method combines knowledge and information from physical and data-driven methods. Most of the methods used by researchers are data-driven (black box).

2.2. Recurrent Neural Network

Recurrent neural network (RNN) is a very popular model used to solve forecasting time series because it can retrieve information obtained from time-series data [24]. The RNN develops the forecasting model based on the input in the current state and inputs from the previous states [25]. RNN extends an artificial neural network (ANN) that connects neurons in the same hidden layer [26] the structure as illustrated in Figure 1.
The RNN structure consists of a loop that can store information from the past [27]. Although the RNN has good forecasting capabilities, it cannot handle data dimensions of long-term dependencies [24]. The long short-term memory (LSTM) improves RNN, overcoming the long-term dependencies problems [28]. The LSTM solves the vanishing gradient problem due to not overcoming data with long-term dependencies [24]. The LSTM uses a component called memory blocked [28]. The vanishing gradient problem can be eliminated by applying the gate mechanism and memory cells that replace the nodes in it [29] as illustrated in Figure 2.
Memory blocked is a subnet containing repeatedly connected functional modules called memory cells and gates [26]. Memory cells are tasked with remembering the state of the network and the gates formed to regulate information patterns [26]. LSTM is divided into three parts, namely input gates, output gates, and forget gates. Input gates function to control the information entering memory cells, forget gates function to remember the amount of information passing through the network, and output gates function to control the amount of information used to perform calculations [26].

2.3. Ensemble Empirical Mode Decomposition

The empirical mode decomposition (EMD) is a pre-processing method used for non-stationary data [30], such as energy-time series data in buildings. Non-stationary data will be decomposed into several components named intrinsic mode functions (IMF) as illustrated in Figure 3 [31]. The EMD method shows its superiority in time series data forecasting [30]. Thus, the ensemble empirical mode decomposition (EEMD) was proposed to improve the EMD [32]. This method was created to overcome the mixing mode problems in EMD [32,33]. It also aims to extract the existing oscillation function [30]. Mode mixing is a problem that arises because one IMF component has a different signal scale or the same signal scale in different IMFs [34]. To bypass the mixing mode, EEMD follows the EMD steps on the original time series x ( t ) ,   ( t = 1 , 2 , ,   t ) , to get a set of IMF added by Gaussian white noises. Then, the average from the IMF applies as the final decomposition result [30].
The steps of EEMD as explained below [30]:
1.
Add the white noise into the original data x ( t ) to get the new construction:
x i ( t ) = x ( t ) + w i ( t ) ,
2.
Describes the time series which have been added the white noise into nth IMF c j i ( t ) ( j = 1 , 2 , , n ) and a residue EMD, r i ( t ) .
x i ( t ) = j = 1 n c j i ( t ) + r i ( t ) ,
3.
Repeat steps 1 and 2 M-times with different white noises until you get the appropriate decomposition result.
4.
Calculate the average of the IMF trials which conducted by M-times as the final IMF.
c j ( t ) = 1 M j = 1 n c j i ( t ) ,
The result of EEMD can be described as a linear combination of IMF and residuals formulated by Equation (4).
x ( t ) = j = 1 n c j ( t ) + r ( t )
The amount of IMF denoted by n, c j ( t ) ( t = 1 , 2 , t ) is the jth IMF extracted in the jth decomposition in the interval t and r ( t ) the final residue. It assumed that the complexity of the data involves information and noises, while the average of the ensemble data with different noises approximates the actual signal, white noises are used to capture the true IMF, and themselves are offset by the ensemble mean. The error of adding white noises can be controlled by Equation (5).
ε n e = ε M ,
where:
  • M —number of ensemble members
  • ε —amplitude of the added noises series amplitude
  • ε n e   —final standard deviation of error
  • ε n e —also interpreted as the difference between the input signal and the relevant IMF [32].

2.4. Wavelet Transformation

Wavelet transformation (WT) is a time-frequency decomposition method that provides a time series basis, both time and frequency [35]. Time series identical to non-stationary data. The WT is a wave that moves up and down in space-time periodically [36]. WT can also be described as a short wave. This model describes the frequency of the signal timing [37]. This model is also known as pre-processing for the denoising process on time series data by decomposing the data into several series [38].
WT identifies the signal shifts and then analyzes them. The purpose of the analysis is to obtain the information and frequency spectrum simultaneously [37]. This model is considered an effective model in analyzing time-frequency after special Fourier analysis in signal processing [39].
WT has two types of models, namely Continuous Wavelet Transformation (CWT) and Discrete Wavelet Transformation (DWT) [39]. CWT is modeled by Equation (6).
C W T y ( α , τ ) = 1 | α | y ( t ) ψ * ( t τ α ) d t ,
where α is parameters of scale, τ is translation parameters, ψ * ( x ) is conjugation complex functions, and ψ ( x ) is mother wavelet.
The Discrete Wavelet Transformation (DWT) is modeled in Equation (7).
D W T y ( m , n ) = α 0 m 2 y ( t ) ψ * ( α 0 m t n τ 0 ) d t ,
where m is a constant scale (decomposition level), and n is a constant integer.

3. Methodology

This study proposes a forecasting framework for electricity consumption. The main challenge in this problem is the complexity of the data pattern. Therefore, a decomposition algorithm is applied to understand data characteristics. Afterward, it applies an LSTM algorithm is to forecast the electricity consumption. Figure 4 illustrates the proposed forecasting framework. The main contribution of this paper is the forecasting framework, which consists of decomposition algorithm. The decomposition algorithm divides time-series into several series. Thus, it needs an additional procedure to make a prediction for the data. Each series is predicted using neural network-based algorithm. Afterward, the results need an aggregation process. The proposed forecasting framework also involves the aggregation step for the data.

3.1. Data Pre-Processing

Data pre-processing is a critical part of processing complex time-series data. Therefore, the proposed framework applies data pre-processing techniques, including normalization, statistics descriptive analysis, and decomposition.

3.1.1. Data Normalization

Data normalization aims to avoid a large variety of data values. Normalization is carried out using the standard scaler method, as shown in Equation (8).
x s c a l e d = x i x m e a n s t a n d a r d   d e v i a t i o n ,

3.1.2. Statistics Descriptive Analysis

Data exploration includes analyzing the data distribution by observing the data patterns, performing descriptive statistical tests, and performing stationarity tests using the Kwiatkowski-Phillips-Schmidt-Shin (KPSS) and augmented Dickey-Fuller (ADF) to see the stationarity of the data to strengthen the theory of the method used. Furthermore, data visualization is also applied to analyze the data before forecasting based on the pattern captured from the data visualization.

3.1.3. Data Decomposition

Decomposition is useful for data that have a high level of complexity and non-stationary data to improve the performance of the forecasting method that will be used. In the decomposition, two methods are used. They are WT and EEMD.
  • Wavelet Transformation
This study uses DWT to analyze the data with a time-scale function. WT has a multi-scale resolution and has time-shifting characteristics. Moreover, scaling operations can observe signals at different scales [40]. Thus, this method is very suitable for handling non-stationary time series data.
Based on Figure 4, the data are divided into two parts, training set x t r a i n ( t ) = 1 , 2 , , n t r a i n and testing set x t e s t ( t ) = 1 , 2 , , n t e s t . The training set is used to build the forecasting model after the WT step. The WT has two components, namely approximation, and detailed series. The approximation series captures the low-frequency features, while the detailed series captures the high-frequency features of the original data. The approximation series is a further decomposed process with WT. The high-frequency noise representing fluctuations and irregularities is extracted and filtered [41]. DWT consists of two basic components of wavelet functions, namely father wavelet φ and mother wavelet ψ as shown in Equations (9) and (10) [38].
φ j , k ( x ) = 2 j 2 φ ( 2 j t k ) ,
ψ j , k ( x ) = 2 j 2 ψ ( 2 j t k ) ,
where j = 1 , , J is the scaling parameter on the decomposition of each j th level. The k is the translation parameter. The approximation series is transformed by the father wavelet from the original data. The detailed series is transformed by the mother wavelet. The detailed series deals with oscillations of length 2 j 2 j + 1 . The approximation and detail series are modeled in Equations (11) and (12).
D j , t = y ( x ) φ j , t ( x ) dx ,
A j , t = y ( x ) ψ j , t ( x ) dx
Finally, the time-series data on electricity consumptions consists of approximation and detail series as represented in Equation (13).
y ( x ) = A j ( x ) + D j ( x ) + D j 1 ( x ) + + D 1 ( x ) ,
2.
Ensemble Empirical Mode Decomposition
The decomposition using EEMD is conducted as follows [30,42]:
(1)
Add the White Gaussian Noise series ε j ( t ) into the train set x t r a i n ( t )   and become the new series of x t r a i n   j ( t ) .
(2)
Decomposed the x t r a i n   j ( t )   into several IMFs c j ( t ) ,   j = 1 , 2 , , n t r a i n and a residue r ( t ) .
(3)
Then, repeated the steps 1 and 2 on each j = 1 , 2 , ,   N E by adding the white Gaussian noise series for every repeated process. N E are the amount of repetition.
(4)
Take the average of all IMFs and the average of the residue as the final results.
(5)
The time series data after EEMD is the sum of all IMFs components and residues as shown in Equation (14).
x t r a i n ( t ) = j = 1 n c j ( t ) + r ( t ) ,

3.2. Forecasting Using LSTM Algorithm

The proposed forecasting framework applies many-to-one RNN structure where each node in the hidden node uses the LSTM structure. Figure 5 illustrates the framework of RNN and the LSTM features.
The LSTM model consists of several stages [43]. It starts by determining the information that can pass in the cell state. The decision is controlled by the forget gate for time t , f t , as shown in Equation (15).
f t = σ ( W f x x t + W f h h t 1 + b f ) ,
where x t is the input of time t , h t 1 is the output from the hidden note at time t 1 , W f x is the weight between nodes in the input layer and forget gate, W f h is the weight between nodes in the hidden layer and forget gate, and b f is bias in forget gate.
The next step is determining information that must be entered into the cell state using Equation (16).
i t = σ ( W i x x t + W i h h t 1 + b i ) ,
where the i t is the output from input gate at time t , W i x is the weight between nodes in the input layer and input gate, W i h is the weight between nodes in the hidden layer and input gate, and b i is bias in the input gate.
In addition, the candidate C ˜ t is generated using Equation (17).
C ˜ t = tan h ( W c x x t + W c h h t 1 + b c ) ,
where W c x is the weight between nodes in the input layer to the output cell,   W c h is the weight between nodes in the hidden layer to the output cell, and b c is the bias of the cell.
Forget the unwanted information by multiplying the old cell state C t 1 with f t and adding some new information to the cell state by i t C ˜ t as shown in Equation (18).
C t = f t C t 1 + i t C ˜ t ,
Calculate the final output by compressing the tanh layer with C t then multiplying by the output gate o t as shown by Equations (19) and (20).
o t = σ ( W o x x t + W o h h t 1 + b o ) ,
h t = o t tan h ( C t )

4. Experimental Result

4.1. Dataset Description

This study applies the proposed forecasting framework to six benchmark datasets and one dataset taken from an education building in Taiwan. The six benchmark datasets consist of 16 buildings from different locations with different time zones and building functionalities. The dataset from Taiwan consists of four buildings data. There are 20 buildings studied in this research. These varieties make the data have different patterns and characteristics. The benchmark dataset consists of industrial buildings, education, commercial building, government, and residential. These data are available at https://github.com/buds-lab/the-building-data-genome-project [44] (accessed on 13 March 2021) and http://traces.cs.umass.edu/index.php/Smart/Smart [45] (accessed on 13 March 2021). The data were initially collected by using a building electricity meter [44]. First, the data are used to analyze the building in question. Then the data are extracted every certain period, one of which is the analysis of the energy system of the building [44]. The unit is a kilowatt-hour (kWh), and each dataset is distinguished based on the level of consumption. Table 2 shows the data specifications. The data are available as hourly data. The forecasting aims to predict one hour. In managing electricity consumption in a building, time hour interval is enough to set the devices.

4.2. Statistic Description Analysis

Data exploration using a statistics description approach is a simple yet effective way to analyze the data concentration and distribution. Table 3 summarizes the statistics descriptive of the data. Table 3 represents the level of energy consumption of each building. The average value shows that dataset 5, the education industry, has very high consumption because it is used for residential in the University, multipurpose buildings, laboratories, classrooms, and offices. Multipurpose and residential buildings in the University have the highest average energy consumption. On the contrary, the apartment and primary/secondary class buildings have the lowest energy consumption.
Furthermore, a data stationarity test is carried out to show whether each series is stationary or non-stationary. This study performs the Kwiatkowski-Phillips-Schmidt-Shin (KPSS) test and the augmented Dickey-Fuller (ADF) test for the stationarity test. The KPSS test is a unit root test for the stationarity of data on a trend. It is often referred to as a trend-stationary test [46]. The ADF test is a statistical test that is more popular than the KPSS test [47]. The ADF test has the same function as the KPSS test, but it tests whether the data are linear or the difference is stationary. The difference between trend or difference stationary is that the stationary trend has a deterministic average, or the shocks caused to the data will return to its path. While difference stationary has a stochastic average property that has permanent shocks, it has the weakest stationary nature of the data [47].
The hypothesis for the KPSS test is as follows:
H 0 —The data are stationary (trend stationary)
H 1 —The data have a unit root. Thus, the data are non-stationary
Table 4 shows the results of the KPSS test, which consists of the p-value and the critical value using four different settings. If the p-value < 0.05, then we reject H 0 , which means the data are non-stationary. Otherwise, the data are stationary. Based on the KPPS test, Office_Evelyn, UnivLab_Susan, and PrimClass_Umar are stationary.
On the other hand, the hypothesis for the SDF test is as follows:
H 0 —the data have a unit root. Thus, the data are non-stationary
H 1 —the data are stationary (difference stationary)
Unlike the KPSS test, the hypothesis of the ADF test is the opposite. If the p-value < 0.05, then reject H 0 which means the data are stationary (difference stationary). Otherwise, the data are non-stationary. The ADF test result in Table 4 shows that only UnivDorm_Malachi is non-stationary.
To see the level of the stationarity of data based on the KPSS and ADF tests, this study refers to the possible outputs generated based on:
  • Case 1: KPSS and ADF tests are non-stationary, then the series is non-stationary.
  • Case 2: KPSS and ADF tests are stationary, so the series is stationary.
  • Case 3: KPSS test is stationary, and ADF is non-stationary, then the series is trend-stationary. The treatment must be done by eliminating the trend to make the series stationary.
  • Case 4: KPSS is non-stationary, and ADF is stationary, so the series the difference stationary. Differentiation is used to create a stationary series.
The combined result of the KPSS and ADF tests in Table 4 reveals that UnivDorm_Malachi is non-stationary. The Office_Evelyn, UnivLab_Susan, and PrimeClass_Umar are stationary. The other series is the difference stationary category (weak), which is the lowest stationary category.
Data visualization in Figure 6 presents the pattern of each series. The charts are available in a span of one year of hourly data resolution. Figure 6 shows that several buildings have similarities in data patterns. Some buildings have similar consumption patterns as Apt_Moon and Apt_Phobos buildings, but both buildings have very low consumption levels. In addition, UnivDorm_Malachi, as the only non-stationary series, has an uneven consumption pattern. Some buildings have stable consumption patterns such as Office_Bobbi, Office_Evelyn, UnivClass_Boyd, UnivLab_Bethany, UnivDorm_Mitch, Office_Glenn, Office_Stella, UnivClass_Seb, UnivLab_Susan, UnivMulti_Ares, UnivMulti_Zeus, PrimClass_Ulysses, PrimClass_Ulysses. Buildings UnivDorm_Athena, UnivExhibit_Hermes, UnivMulti_Ares, UnivMulti_Zeus have a very high consumption rate and the highest among other buildings.

4.3. Parameter Setting

This study applies the proposed forecasting framework to predict one hour ahead data based on one-year data for each building. Herein, the data period is hourly. In the forecasting process, the data are divided into training, validation, and testing data. Training data are 70% of the data population of each building, validation data are 20%, and testing data are 10%. The data splitting process is carried out sequentially, accounting for 70% of the first data for training. Then, after 70% of the training data, 20% of the data is used for validation, and the remaining 10% of the data for testing data is calculated after 20% of the validation data until the last data. The validation is the result of the splitting of data testing carried out to overcome overfitting in the data, and the model does not understand the character of the data [48].
The LSTM involves several parameters, including the number of hidden nodes, lookback periods, activation function, and recurrent activation. To get the best parameter setting for the forecasting model, this study evaluates several parameter settings as listed in Table 5. Each parameter set is used for all buildings, with five replications for each building.
Table 6 shows the best parameter setting for LSTM. This parameter is used of forecasting without decomposition, with EEMD-LSTM, and WT-LSTM. According to Table 6, the forecasting is conducted using 24 h of look back data. It means that predicting the next one hour requires the previous 24 h of historical data. The number of hidden nodes indicates the number of nodes contained in the hidden layers of the LSTM network. The optimizer is Nadam. Nadam is one of the adaptive movement estimations (Adam) that added momentum to the Nesterov accelerated gradient (NAG). This optimizer works for the gradient descent process in the process of finding the minimum function by increasing the speed of convergence and the quality of the studied model. Epoch is number of iterations used for the forward and backward process on developing the model based on the training dataset. Activation is used in the cell state of the LSTM structure, while recurrent activation is used for the gates in the LSTM including input gate, forget gate and output gate.
The forecasting model with EEMD-LSTM involves more parameters. They are the noise width (white noise’s standard deviation) and ensemble member (trial). This study uses 0.05 as the noise width and 100 for the ensemble member [32]. Furthermore, the WT-LSTM is set using a single level of decomposition, symmetric signal extension model, and Daubechies (db1) wavelet object.

4.4. Forecasting Results Analysis

To evaluate the robustness of the proposed forecasting framework, each forecasting method is run ten times for each data series. This study evaluates the results based on the mean absolute percentage error (MAPE), mean squared error (MSE), and mean absolute error (MAE). These parameters are calculated using Equations (21)–(23).
M A P E = 1 n i = 1 n | y i y ^ i y i |
M S E = i = 1 n ( y i y ^ i ) 2 n  
M A E = i = 1 n | y i y ^ i | n
where y i is the actual value of the period i , y ^ i is the predicted value of the period i , and n is the number of periods. To evaluate the performance of the LSTM, this study also compares the LSTM with the recurrent neural network (RNN) algorithm.
Table 7, Table 8, Table 9 and Table 10 shows the average and standard deviation value of the results. Table 11 shows the comparison of the three approaches based on the MAPE value. These results reveal that the eight different buildings, including office buildings, lecture laboratories, lecture classrooms, dormitory or residential in university areas, apartments, multipurpose, and secondary classrooms, have very error rates. The training error of the EEMD-LSTM method has the lowest average MAPE value compared to the LSTM and WT-LSTM. It is followed by the testing error of the EEMD-LSTM, which is still the lowest among the other methods. It can be caused by the decomposition method itself, which helps in the denoising process of the time-series data used.
Figure 7 shows the comparison between the actual and predicted value of the energy consumption. It reveals that for most of the buildings, the predicted value is very close to the actual value. Furthermore, the standard deviation of the EEMD-LSTM algorithm shown in Table 9 is relatively small. It indicates that the EEMD-LSTM algorithm has a stable result.
Furthermore, this study also applies a time-series cross-validation to evaluate the forecasting result. Figure 8 shows the framework of the time-series cross-validation applied in this study. Table 12, Table 13, Table 14 and Table 15 summarize the results of the time-series cross-validation using five-fold cross-validation. In addition, Table 16 shows the comparison of all algorithms. These tables show that the EEMD-LSTM algorithm obtains the best result for most of the buildings in terms of the average of MAPE, MSE, and MAE.
The EEMD decomposition is carried out to the maximum decomposition based on the amount of data used on its implementation. Meanwhile, the LSTM method has the lower error rates than the WT-LSTM because the wavelet object selection uses Daubechies 1 (db1). The db1 is the same as the oldest wavelet object, namely Haar. This object is the most commonly used object to simplify the implementation, and db1 only has two filters. Each filter is used to catch the low pass filter and high pass filter. The low pass filter is used to decompose the signal that is close to the original data. It is called the approximation series. The high pass filter contains noise named as detail series. The decomposition level used single or one level. Therefore, it only produces an approximation series and a detailed series which causes the signal decomposition to be less smooth and results in a fairly high error compared to the LSTM.
Based on the forecasting results, the apartments buildings consisting of Apt_Moon and Apt_Phobos have the highest error percentage rates in training, validation, and testing. This is caused by the raw data of the two buildings having an extreme slope, touching below 1 kWh. Those two buildings have the lowest level of consumption among the other two buildings in the same category. The uneven distribution of the raw data causes high errors during the forecasting process. Both buildings have significantly different data ranges in several periods. For several periods, the data are relatively stable, while at other periods significantly decreases. The forecasting model is generated based on the training set, which is relatively stable data, while the validation and testing set has a significantly different pattern. Therefore, the validation and testing errors are significantly higher than the training error. The result of five-fold cross-validation also shows that the average MAPE of the third fold has the highest value. At the third fold, the training and test set has a significantly different pattern as illustrated in Figure 9.
Buildings such as college classrooms and laboratories have more stable forecasting results with small errors than other buildings, followed by offices. Another building that has a high error is the primary/secondary classroom. It is caused by the fairly high distribution of data based on the buildings in that category. Stationary buildings such as UnivClass_Boyd, Office_Stella and PrimClass_Uma tend to have small and stable errors based on data stationarity. The UnivClass_Boyd has the smallest error among other buildings. Likewise, UnivDorm_Malachi is the only non-stationary building that has very small and stable errors in training, validation, and testing.
Further comparison is conducted using statistics Levene’s test and T-test to evaluate how significant the differences are between the LSTM without decomposition and with decomposition. The statistics test uses MAPE results as the data. This study runs the statistical tests using SPSS.
Levene’s test checks whether the variance of the two samples is the same or not with the following hypothesis below.
H 0 —both samples have the same variance
H 1 —both samples have different variances
If the p-value (sig.) is less than 0.05, then reject H 0 meaning that the results from the two methods have different variances. Otherwise, they have the same variances.
The T-test compares the results based on their means with the following hypothesis.
H 0 —both samples have the similar means
H 1 —both samples have different means
If the p-value (sig.) is less than 0.05, then reject H 0 meaning that the results from the two methods are significantly different. Otherwise, their results are not significantly different.
The statistics test results summarized in Table 17 reveal that the LSTM and EEMD-LSTM methods have significantly different variance and average. However, LSTM and WT-LSTM do not significantly differ in terms of the variance and average. Furthermore, the results of EEMD-LSTM and WT-LSTM are significantly different in terms of variance and average.
The test results conclude that the EEMD-LSTM method has a significant difference compared with the LSTM and WT-LSTM methods. On the other hand, the LSTM method is not significantly different from the WT-LSTM method. This conclusion is also supported by the comparison of the average MAPE Test for each method, where the MAPE test for the EEMD-LSTM method tends to be lower. However, the MAPE of LSTM and WT-LSTM methods are not significantly different. The results also reveal that the EEMD-LSTM method obtains better results than the LSTM and WT-LSTM methods.
Based on the previous studies conducted by Imani and Ghassemian [49] and Fang, Gao, and Ruan [14], the error of LSTM without decomposition is between 3% to 16%. Meanwhile, for this research, the forecasting results using the LSTM method has an average error of 15.3%. The results show that EEMD has successfully improved the forecasting accuracy of the LSTM method.
Finally, this research also provides the recommendation of the forecasting method for each building, as listed in Table 18. However, for the APT-Moon and APT_Phobos, other algorithms should be considered because the MAPE of the EEMD-LSTM is still above 20%.

5. Conclusions

Energy consumption is predicted to increase continuously. Building as a facility where most of the human activities happen consumes energy, especially electricity. Without proper energy management, energy consumption of a building might be inefficient. On the other hand, managing energy consumption needs information on the future energy demand. Therefore, this study proposes a forecasting framework to predict the energy consumption of a building.
The proposed forecasting framework is developed based on LSTM with decomposition methods to improve the performance of the LSTM. There are three methods evaluated in this study. They are the LSTM without decomposition, LSTM with EEMD as the decomposition, and the LSTM with WT as the decomposition method. The LSTM without decomposition is the baseline for both EEMD-LSTM and WT-LSTM.
This study applies the proposed forecasting methods on 20 series of energy consumption data from 20 different buildings located in many different countries and different functions. The experiment shows that the EEMD-LSTM method obtains the smallest error for 85% of the series. While 15% of the series obtained the smallest error when using the LSTM method. The average error for LSTM, EEMD-LSTM and WT-LSTM are 15.31%, 7.57%, and 16.36%, respectively. The statistics T-test also reveals that the EEMD-LSTM method is significantly better than the other two methods.
Meanwhile, the results of the LSTM and WT-LSTM are not significantly different. The EEMD and WT methods have significantly different values because the EEMD decomposed the data based on the sample data used. On the other hand, the WT decomposition only uses a single-level decomposition, and the object wavelets only have two filters. Further study should increase the decomposition level and the number of filters to improve the performance of WT. Other hybrid forecasting methods and differences also should be considered. In addition to the forecasting algorithm, further research should consider more datasets from different types and locations of the building to get a more general forecasting framework.

Author Contributions

Conceptualization, A.D. and S.-Y.C.; methodology, F.E.Z.; software, A.D. and M.F.; validation, A.D. and F.E.Z.; formal analysis, A.D. and F.E.Z.; investigation, A.D., S-Y. Chou, and F.E.Z.; resources, S.-Y.C.; data curation, A.D.; writing, A.D., F.E.Z. and M.F.; visualization, M.F.; supervision, S-Y. Chou; project administration, S-Y. Chou; funding acquisition, S-Y. Chou. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded in part by the Taiwan Building Technology Center, National Taiwan University of Science and Technology, from the Featured Areas Research Center Program within the framework of the Higher Education Sprout Project by the Ministry of Education in Taiwan under the grant 110P011.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Benchmark data are available on https://github.com/buds-lab/the-building-data-genome-project [44] (accessed on 13 March 2021) and http://traces.cs.umass.edu/index.php/Smart/Smart [45] (accessed on 13 March 2021).

Acknowledgments

This work was supported in part by the Taiwan Building Technology Center, National Taiwan University of Science and Technology, from the Featured Areas Research Center Program within the framework of the Higher Education Sprout Project by the Ministry of Education in Taiwan under the grant 110P011.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cao, X.; Dai, X.; Liu, J. Building energy-consumption status worldwide and the state-of-the-art technologies for zero-energy buildings during the past decade. Energy Build. 2016, 128, 198–213. [Google Scholar] [CrossRef]
  2. Saidur, R.; Masjuki, H.H.; Jamaluddin, M.Y. An application of energy and exergy analysis in residential sector of Malaysia. Energy Policy 2007, 35, 1050–1063. [Google Scholar] [CrossRef]
  3. Allouhi, A.; El Fouih, Y.; Kousksou, T.; Jamil, A.; Zeraouli, Y.; Mourad, Y. Energy consumption and efficiency in buildings: Current status and future trends. J. Clean. Prod. 2015, 109, 118–130. [Google Scholar] [CrossRef]
  4. EIA. International Energy Outlook 2019; U.S. Energy Information Administration (EIA): Washington, DC, USA, 2019.
  5. Bedi, J.; Toshniwal, D. Empirical mode decomposition based deep learning for electricity demand forecasting. IEEE Access 2018, 6, 49144–49156. [Google Scholar] [CrossRef]
  6. Ahmad, T.; Chen, H. A review on machine learning forecasting growth trends and their real-time applications in different energy systems. Sustain. Cities Soc. 2020, 54, 102010. [Google Scholar] [CrossRef]
  7. Kontokosta, C.E.; Tull, C. A data-driven predictive model of city-scale energy use in buildings. Appl. Energy 2017, 197, 303–317. [Google Scholar] [CrossRef] [Green Version]
  8. Hong, W.-C.; Dong, Y.; Zhang, W.Y.; Chen, L.-Y.; Panigrahi, B.K. Cyclic electric load forecasting by seasonal SVR with chaotic genetic algorithm. Int. J. Electr. Power Energy Syst. 2013, 44, 604–614. [Google Scholar] [CrossRef]
  9. Kneifel, J.; Webb, D. Predicting energy performance of a net-zero energy building: A statistical approach. Appl. Energy 2016, 178, 468–483. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Barhmi, S.; Elfatni, O.; Belhaj, I. Forecasting of wind speed using multiple linear regression and artificial neural networks. Energy Syst. 2019, 11, 935–946. [Google Scholar] [CrossRef]
  11. Montgomery, D.; Jennings, C.; Kulahci, M. Introduction to Time Series Analysis and Forecasting; John Wiley & Sons: Hoboken, NJ, USA, 2008; p. 472. [Google Scholar]
  12. Mocanu, E.; Nguyen, P.H.; Gibescu, M.; Kling, W.L. Deep learning for estimating building energy consumption. Sustain. Energy Netw. 2016, 6, 91–99. [Google Scholar] [CrossRef]
  13. Kong, W.; Dong, Z.Y.; Jia, Y.; Hill, D.J.; Xu, Y.; Zhang, Y. Short-term residential load forecasting based on lstm recurrent neural network. IEEE Trans. Smart Grid 2019, 10, 841–885. [Google Scholar] [CrossRef]
  14. Fang, C.; Gao, Y.; Ruan, Y. Improving forecasting accuracy of daily energy consumption of office building using time series analysis based on wavelet transform decomposition. IOP Conf. Ser. Earth Environ. Sci. 2019, 294, 012031. [Google Scholar] [CrossRef] [Green Version]
  15. Peng, L.; Wang, L.; Xia, D.; Gao, Q.J.E. Effective energy consumption forecasting using empirical wavelet transform and long short-term memory. Energy 2022, 238, 121756. [Google Scholar] [CrossRef]
  16. Gao, Y.; Ruan, Y.J.E. Buildings, interpretable deep learning model for building energy consumption prediction based on attention mechanism. Energy Build. 2021, 252, 111379. [Google Scholar] [CrossRef]
  17. Somu, N.; MR, G.R.; Ramamritham, K.J.A.E. A hybrid model for building energy consumption forecasting using long short term memory networks. Appl. Energy 2020, 261, 114131. [Google Scholar] [CrossRef]
  18. Runge, J.; Zmeureanu, R. Forecasting energy use in buildings using artificial neural networks: A review. Energies 2019, 12, 3254. [Google Scholar] [CrossRef] [Green Version]
  19. Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  20. Koprinska, I.; Rana, M.; Agelidis, V.G. Correlation and instance based feature selection for electricity load forecasting. Knowl.-Based Sys. 2015, 82, 29–40. [Google Scholar] [CrossRef]
  21. Friedrich, L.; Afshari, A. Short-term forecasting of the abu dhabi electricity load using multiple weather variables. Energy Procedia 2015, 75, 3014–3026. [Google Scholar] [CrossRef] [Green Version]
  22. Deb, C.; Zhang, F.; Yang, J.; Lee, S.E.; Shah, K.W. A review on time series forecasting techniques for building energy consumption. Renew. Sustain. Energy Rev. 2017, 74, 902–924. [Google Scholar] [CrossRef]
  23. Bourdeau, M.; Zhai, X.; Nefzaoui, E.; Guo, X.; Chatellier, P. Modeling and forecasting building energy consumption: A review of data-driven techniques. Sustain. Cities Soc. 2019, 48, 101533. [Google Scholar] [CrossRef]
  24. Mandic, D.P.; Chambers, J.A. Recurrent Neural Networks for Prediction: Learning Algorithms, Architectures and Stability; Wiley: Hoboken, NJ, USA, 2001. [Google Scholar]
  25. Kumar, D.N.; Raju, K.S.; Sathish, T. River flow forecasting using recurrent neural networks. Water Resour. Manag. 2004, 18, 143–161. [Google Scholar] [CrossRef]
  26. Hua, Y.; Zhao, Z.; Li, R.; Chen, X.; Liu, Z.; Zhang, H. Deep learning with long short-term memory for time series prediction. IEEE Commun Mag. 2019, 57, 114–119. [Google Scholar] [CrossRef] [Green Version]
  27. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  28. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  29. Aungiers, J. LSTM Neural Network for Time Series Prediction. Available online: https://www.jakob-aungiers.com/articles/a/LSTM-Neural-Network-for-Time-Series-Prediction (accessed on 13 March 2021).
  30. Wu, Y.-X.; Wu, Q.-B.; Zhu, J.-Q. Improved EEMD-based crude oil price forecasting using LSTM networks. Phys. A Stat. Mech. its Appl. 2019, 516, 114–124. [Google Scholar] [CrossRef]
  31. Xu, Y.; Zhou, C.; Geng, J.; Gao, S.; Wang, P. A method for diagnosing mechanical faults of on-load tap changer based on ensemble empirical mode decomposition, volterra model and decision acyclic graph support vector machine. IEEE Access 2019, 7, 84803–84816. [Google Scholar] [CrossRef]
  32. Wu, Z.; Huang, N.E. Ensemble empirical mode decomposition: A noise-assisted data analysis method. Adv. Adapt. Data Anal. 2009, 1, 1–41. [Google Scholar] [CrossRef]
  33. Huang, N.; Shen, Z.; Long, S. A new view of non-linear water waves: The hilbert spectrum. Annu. Rev. Fluid Mech. 2003, 31, 417–457. [Google Scholar] [CrossRef] [Green Version]
  34. Zhang, X.; Lai, K.K.; Wang, S.-Y. A new approach for crude oil price analysis based on empirical mode decomposition. Energy Econ. 2008, 30, 905–918. [Google Scholar] [CrossRef]
  35. Terzija, N. Robust Digital Image Watermarking Algorithms for Copyright Protection. Ph.D. Thesis, University of Duisburg-Essen, Duisburg, Germany, 2006. [Google Scholar]
  36. Sripathi, D. Efficient Implementations of Discrete Wavelet Transforms using FPGAs; Florida State University, Florida, USA. 2003. Available online: http://purl.flvc.org/fsu/fd/FSU_migr_etd-1599 (accessed on 18 July 2021).
  37. Sugiartawan, P.; Pulungan, R.; Kartika, A. Prediction by a hybrid of wavelet transform and long-short-term-memory neural networks. Int. J. Adv. Comput. Sci. Appl. 2017, 8, e0142064. [Google Scholar] [CrossRef] [Green Version]
  38. Jin, J.; Kim, J. Forecasting natural gas prices using wavelets, time series, and artificial neural networks. PLoS ONE 2015, 10, e0142064. [Google Scholar] [CrossRef]
  39. Chang, Z.; Zhang, Y.; Chen, W. Electricity price prediction based on hybrid model of adam optimized LSTM neural network and wavelet transform. Energy 2019, 187, 115804. [Google Scholar] [CrossRef]
  40. Liu, Y.; Guan, L.; Hou, C.; Han, H.; Liu, Z.; Sun, Y.; Zheng, M. Wind Power Short-Term Prediction Based on LSTM and Discrete Wavelet Transform. Appl. Sci. 2019, 9, 1108. [Google Scholar] [CrossRef] [Green Version]
  41. Wang, F.; Yu, Y.; Zhang, Z.; Li, J.; Zhen, Z.; Li, K. Wavelet decomposition and convolutional lstm networks based improved deep learning model for solar irradiance forecasting. Appl. Sci. 2018, 8, 1286. [Google Scholar] [CrossRef] [Green Version]
  42. Qin, Q.; Lai, X.; Zou, J. Direct multistep wind speed forecasting using lstm neural network combining eemd and fuzzy entropy. Appl. Sci. 2019, 9, 126. [Google Scholar] [CrossRef] [Green Version]
  43. Liu, W.; Liu, W.D.; Gu, J. Forecasting oil production using ensemble empirical model decomposition based long short-term memory neural network. J. Pet. Sci. Eng. 2020, 189, 107013. [Google Scholar] [CrossRef]
  44. Miller, C.; Meggers, F. The building data genome project: An open, public data set from non-residential building electrical meters. Energy Procedia 2017, 122, 439–444. [Google Scholar] [CrossRef]
  45. Barker, S.; Mishra, A.; Irwin, D.; Cecchet, E.; Shenoy, P.; Albrecht, J. Smart*: An open data set and tools for enabling research in sustainable homes. SustKDD 2012, 111, 108. [Google Scholar]
  46. Kwiatkowski, D.; Phillips, P.C.; Schmidt, P.; Shin, Y. Testing the null hypothesis of stationarity against the alternative of a unit root: How sure are we that economic time series have a unit root? J. Econ. 1992, 54, 159–178. [Google Scholar] [CrossRef]
  47. Singh, A.A. Gentle Introduction to Handling a Non-Stationary Time Series in Python. Available online: https://www.analyticsvidhya.com/blog/2018/09/non-stationary-time-series-python/# (accessed on 18 July 2021).
  48. Mueller, J.P.; Massaron, L. Machine Learning for Dummies; John Wiley & Sons: Hoboken, NJ, USA, 2021. [Google Scholar]
  49. Imani, M.; Ghassemian, H. Lagged load wavelet decomposition and lstm networks for short-term load forecasting. In Proceedings of the 2019 4th International Conference on Pattern Recognition and Image Analysis (IPRIA), Tehran, Iran, 6–7 March 2019; pp. 6–12. [Google Scholar]
Figure 1. Structure of recurrent neural network [26].
Figure 1. Structure of recurrent neural network [26].
Energies 15 01035 g001
Figure 2. Structure of long short-term memory (LSTM) [26].
Figure 2. Structure of long short-term memory (LSTM) [26].
Energies 15 01035 g002
Figure 3. Intrinsic mode function (IMF) [5].
Figure 3. Intrinsic mode function (IMF) [5].
Energies 15 01035 g003
Figure 4. The proposed forecasting framework.
Figure 4. The proposed forecasting framework.
Energies 15 01035 g004
Figure 5. The RNN and LSTM structures.
Figure 5. The RNN and LSTM structures.
Energies 15 01035 g005
Figure 6. Energy consumption pattern in each building.
Figure 6. Energy consumption pattern in each building.
Energies 15 01035 g006aEnergies 15 01035 g006b
Figure 7. Comparison of the Actual and Predicted Value.
Figure 7. Comparison of the Actual and Predicted Value.
Energies 15 01035 g007aEnergies 15 01035 g007bEnergies 15 01035 g007cEnergies 15 01035 g007d
Figure 8. Framework for the time-series cross-validation.
Figure 8. Framework for the time-series cross-validation.
Energies 15 01035 g008
Figure 9. The pattern differences between training, validation, and testing set for Apt_moon dataset.
Figure 9. The pattern differences between training, validation, and testing set for Apt_moon dataset.
Energies 15 01035 g009
Table 1. The Advantages and Disadvantages of Data-Driven and Deterministic Technique.
Table 1. The Advantages and Disadvantages of Data-Driven and Deterministic Technique.
TechniquesAdvantagesDisadvantages
Data-drivenFaster calculations with real-time dataHistorical data are needed
Can be applied for non-linear problems Solution for a case may not be suitable for other cases
DeterministicRefers to the physical science of the buildingIt is difficult to create scenarios that refer to the actual model
Does not require historical dataRequires detail building properties data
One solution can be applied to many cases-
Source: [22].
Table 2. Dataset specifications.
Table 2. Dataset specifications.
DatasetUnitsData StartData EndFunctionalitiesTimezone
1UnivClass_Boyd2012-01-01 0:002012-12-13 23:00College ClassroomAmerica/Los_Angeles
UnivLab_Bethany2012-01-01 0:002012-12-13 23:00College LaboratoryAmerica/Los_Angeles
Office_Evelyn2012-01-01 0:002012-12-13 23:00OfficeAmerica/Los_Angeles
Office_Bobbi2012-01-01 0:002012-12-13 23:00OfficeAmerica/Los_Angeles
2UnivDorm_Malachi2014-05-01 0:002015-04-30 23:00ResidentialAmerica/Chicago
UnivDorm_Mitch2014-05-01 0:002015-04-30 23:00ResidentialAmerica/Chicago
3UnivClass_Seb2014-12-01 0:002015-11-30 23:00College ClassroomEurope/London
UnivLab_Susan2014-12-01 0:002015-11-30 23:00College LaboratoryEurope/London
Office_Stella2014-12-01 0:002015-11-30 23:00OfficeEurope/London
Office_Glenn2014-12-01 0:002015-11-30 23:00OfficeEurope/London
4Apt_Moon2015-01-01 0:002015-12-31 23:00ApartmentAmerica/New_York
Apt_Phobos2015-01-01 0:002015-12-31 23:00ApartmentAmerica/New_York
5UnivExhibit_Hermes2019-01-02 0:002020-01-31 23:00ExhibitionAsia/Taiwan
UnivDorm_Athena2019-01-02 0:002020-01-31 23:00ResidentialAsia/ Taiwan
UnivMulti_Zeus2019-01-02 0:002020-01-31 23:00MultipurposeAsia/ Taiwan
UnivMulti_Ares2019-01-02 0:002020-01-31 23:00MultipurposeAsia/ Taiwan
6PrimClass_Uma2012-02-02 0:002013-01-31 23:00Primary/Secondary ClassroomAsia/Singapore
PrimClass_Umar2012-02-02 0:002013-01-31 23:00Primary/Secondary ClassroomAsia/Singapore
UnivDorm_Una2012-02-02 0:002013-01-31 23:00ResidentialAsia/Singapore
PrimClass_Ulysses2012-02-02 0:002013-01-31 23:00Primary/Secondary ClassroomAsia/Singapore
Source: [44,45], and an education building in Taiwan.
Table 3. Descriptive statistics.
Table 3. Descriptive statistics.
DatasetUnitsMeanStdevVarianceMinMax
1UnivClass_Boyd22.1991.9123.65516.78828.452
UnivLab_Bethany72.94617.247297.45143.634124.788
Office_Evelyn217.39885.8117,363.4598.444479.186
Office_Bobbi74.4228.538814.4333.225157.675
2UnivDorm_Malachi88.06131.071965.42935.25180.25
UnivDorm_Mitch62.8217.281298.63732.25116.25
3UnivClass_Seb63.36525.922671.97310141
UnivLab_Susan19.15711.801139.27456
Office_Stella66.44226.473700.82920143
Office_Glenn34.03515.484239.7535.185.26
4Apt_Moon1.0770.930.8640.0186.872
Apt_Phobos1.3821.2411.5390.0128.715
5UnivExhibit_Hermes0.2930.2400.0580.0442.68
UnivDorm_Athena111.90846.1382128.69126.246347.304
UnivMulti_Zeus90.11233.6311131.06135.927222.969
UnivMulti_Ares28.45413.916193.6559.05594.036
6PrimClass_Uma12.3759.0181.184139
PrimClass_Umar21.73917.24297.221173
UnivDorm_Una23.5358.67375.213149
PrimClass_Ulysses25.27926.157684.2071103
Table 4. Results of KPSS and ADF test.
Table 4. Results of KPSS and ADF test.
SeriesKPSS TestADF TestCombination of KPSS and ADF Results
Test Statisticsp-ValueInterpretationTest Statisticsp-ValueInterpretation
UnivClass_Boyd1.3250.01non-stationary−7.2681.61 × 10−10stationarydifference stationary
UnivLab_Bethany0.3190.01non-stationary−10.6564.54 × 10−19stationarydifference stationary
Office_Evelyn0.1430.055stationary−13.552.42 × 10−25stationarystationary
Office_Bobbi0.3010.01non-stationary−10.9837.37 × 10−20stationarydifference stationary
UnivDorm_Malachi1.190.01non-stationary−2.5110.1128005non-stationarynon-stationary
UnivDorm_Mitch1.0910.01non-stationary−3.8780.0022008stationarydifference stationary
UnivClass_Seb0.7150.01non-stationary−11.9244.96 × 10−22stationarydifference stationary
UnivLab_Susan0.0870.1stationary−11.1912.36 × 10−20stationarystationary
Office_Stella2.4710.01non-stationary−5.5871.36 × 10−6stationarydifference stationary
Office_Glenn1.630.01non-stationary−5.765.70 × 10−7stationarydifference stationary
Apt_Moon3.3290.01non-stationary−4.2090.000635stationarydifference stationary
Apt_Phobos3.6360.01non-stationary−4.1590.0007739stationarydifference stationary
UnivExhibit_Hermes1.9660.01non-stationary−8.0451.81 × 10−12stationarydifference stationary
UnivDorm_Athena1.8050.01non-stationary−4.7277.49 × 10−5stationarydifference stationary
UnivMulti_Zeus2.7870.01non-stationary−6.3642.43 × 10−8stationarydifference stationary
UnivMulti_Ares1.6630.01non-stationary−6.6714.58 × 10−9stationarydifference stationary
PrimClass_Uma0.2320.01non-stationary−11.2232.00 × 10−20stationarydifference stationary
PrimClass_Umar0.1190.1stationary−12.6121.65 × 10−23stationarystationary
UnivDorm_Una0.5770.01non-stationary−3.6430.0049932stationarydifference stationary
PrimClass_Ulysses0.2160.01non-stationary−11.6452.11 × 10−21stationarydifference stationary
Table 5. Trial of parameter combinations.
Table 5. Trial of parameter combinations.
CombinationsHidden NodesLook BackActivationRecurrent Activation
16424sigmoidsigmoid
26424sigmoidrelu
36424sigmoidtanh
46424tanhSigmoid
56424tanhrelu
66424tanhtanh
76424relusigmoid
86424relutanh
964168sigmoidsigmoid
1064168sigmoidtanh
1164168tanhSigmoid
1264168tanhtanh
1312824sigmoidsigmoid
1412824sigmoidrelu
1512824sigmoidtanh
1612824tanhSigmoid
1712824tanhrelu
1812824tanhtanh
1912824relusigmoid
2012824relutanh
21128168sigmoidsigmoid
22128168sigmoidtanh
23128168tanhSigmoid
24128168tanhtanh
Table 6. Best parameters setting for the LSTM.
Table 6. Best parameters setting for the LSTM.
ParametersValues
Layers1 layer
Hidden nodes64 nodes
Look back24 h
Epochs50 epochs
OptimizerNadam
ActivationTanh
Recurrent activationSigmoid
Table 7. The MAPE, MSE, and MAE of forecasting using LSTM without decomposition.
Table 7. The MAPE, MSE, and MAE of forecasting using LSTM without decomposition.
Dataset MAPE (%)MAEMSE
MetricsTrainValTestTrainValTestTrainValTest
Univclass_Boydmean1.951.892.080.440.430.430.340.310.32
std0.020.020.100.000.010.020.000.010.02
Univlab_Bethanymean2.783.002.802.112.192.059.149.678.58
std0.040.040.040.030.030.030.200.420.31
Office_Evelynmean4.614.636.0110.0510.1013.12438.18352.39674.85
std0.140.320.120.280.710.2317.1348.7617.05
Office_Bobbimean5.727.877.834.295.185.4036.7548.3953.76
std0.080.250.170.050.130.110.902.032.06
Univdorm_Malachimean4.764.715.773.784.866.3235.4347.0768.26
std0.080.110.120.050.080.100.550.851.99
Univdorm_Mitchmean8.107.987.855.744.554.5363.8539.9238.46
std0.080.310.190.040.180.120.852.652.42
Univclass_Sebmean3.644.753.972.182.402.588.6411.6312.99
std0.090.250.110.050.120.070.391.020.86
Univlab_Susanmean7.6013.6511.111.682.601.979.8921.0411.58
std0.580.670.850.070.090.120.460.920.71
Office_Stellamean3.495.324.072.362.263.1212.3911.2919.02
std0.070.270.120.050.070.100.630.441.11
Office_Glennmean6.776.717.712.841.632.0916.745.808.60
std0.100.360.160.030.070.030.180.310.20
Apt_Moonmean34.31128.0339.380.210.340.540.160.430.73
std2.3210.510.930.000.000.010.000.010.01
Apt_Phobosmean47.92116.1249.060.420.380.760.580.401.09
std5.8213.421.260.000.000.000.010.010.01
Univexhibit_Hermesmean5.959.274.980.020.050.010.010.020.00
std0.120.270.160.000.000.000.000.000.00
Univdorm_Athenamean8.058.359.698.409.338.32120.67155.49117.45
std0.120.190.240.120.240.133.357.313.10
Univmulti_Zeusmean5.295.668.025.155.185.4253.8257.5266.14
std0.100.200.680.050.130.320.542.402.78
Univmulti_Aresmean7.3611.2115.421.744.015.389.3034.0757.41
std0.170.190.250.030.050.070.201.021.33
Primclass_Umamean19.6917.9017.271.531.581.5610.118.459.41
std0.450.640.560.030.040.040.280.410.46
Primclass_Umarmean20.5914.9624.422.672.733.8429.9224.1741.45
std0.730.600.710.070.070.070.791.401.16
Univdorm_Unamean18.3315.3316.712.293.643.3515.7628.3325.60
std0.430.690.360.040.190.100.493.151.63
Primclass_Ulyssesmean27.0922.9628.853.334.675.2466.4588.1999.04
std0.910.770.650.090.190.121.444.822.22
Table 8. The MAPE, MSE, and MAE of forecasting using RNN without decomposition.
Table 8. The MAPE, MSE, and MAE of forecasting using RNN without decomposition.
Dataset MAPE (%)MAEMSE
MetricsTrainValTestTrainValTestTrainValTest
Univclass_Boydmean1.762.052.200.400.470.460.290.360.35
std0.030.020.050.010.010.010.010.000.01
Univlab_Bethanymean2.733.032.942.062.212.158.779.518.99
std0.080.080.070.050.050.050.310.410.53
Office_Evelynmean4.544.076.409.698.7013.80401.08192.37692.24
std0.100.250.140.240.640.2513.1824.3627.48
Office_Bobbimean5.478.108.024.075.275.5033.2948.6154.71
std0.130.470.320.070.200.150.902.752.25
Univdorm_Malachimean4.754.946.053.715.126.6533.9351.3575.23
std0.160.150.140.080.130.141.051.792.95
Univdorm_Mitchmean7.518.518.395.274.834.8156.0044.7342.79
std0.140.280.110.090.150.071.522.301.23
Univclass_Sebmean3.694.934.112.232.512.688.9712.8814.04
std0.270.260.260.150.120.150.940.981.09
Univlab_Susanmean7.7613.9810.541.662.591.889.1520.2510.53
std0.350.810.660.040.070.060.350.640.46
Office_Stellamean3.335.374.022.242.253.0910.6710.9318.45
std0.080.320.100.040.110.080.380.770.72
Office_Glennmean6.547.528.342.691.772.2315.336.489.48
std0.331.090.580.100.180.110.750.680.50
Apt_Moonmean36.55159.7442.120.210.360.550.150.430.74
std2.5916.610.850.010.010.010.010.010.01
Apt_Phobosmean54.35151.3352.990.410.390.800.530.391.13
std8.9723.681.770.010.010.010.020.020.03
Univexhibit_Hermesmean6.159.735.440.030.050.010.010.020.00
std0.110.150.170.000.000.000.000.000.00
Univdorm_Athenamean7.638.4810.137.879.468.73108.82159.33127.89
std0.340.290.470.220.310.264.329.617.01
Univmulti_Zeusmean5.215.797.845.045.305.2152.1058.0663.32
std0.130.220.570.120.160.271.912.342.67
Univmulti_Aresmean7.3511.5915.581.704.125.448.7334.7857.09
std0.410.230.230.070.070.080.290.911.47
Primclass_Umamean18.7118.1317.581.491.661.629.689.399.71
std0.660.860.600.050.050.040.380.400.39
Primclass_Umarmean19.8015.3126.642.562.783.9428.1224.2542.46
std1.020.731.230.100.110.130.561.091.24
Univdorm_Unamean16.6015.8718.102.003.783.6312.9528.4928.56
std0.550.640.450.050.180.110.572.291.50
Primclass_Ulyssesmean26.5523.1131.513.224.715.3262.1788.4398.03
std1.620.901.170.120.200.181.933.713.50
Table 9. The MAPE, MSE, and MAE of forecasting using EEMD-LSTM model.
Table 9. The MAPE, MSE, and MAE of forecasting using EEMD-LSTM model.
Dataset MAPE (%)MAEMSE
MetricsTrainValTestTrainValTestTrainValTest
Univclass_Boydmean0.630.711.660.140.160.340.040.040.19
std0.010.040.120.000.010.020.000.000.03
Univlab_Bethanymean1.001.091.050.740.780.761.021.061.00
std0.020.010.010.010.000.000.040.010.01
Office_Evelynmean1.991.472.654.013.085.5649.5517.35113.79
std0.040.010.020.080.030.023.200.241.50
Office_Bobbimean1.902.302.431.361.501.653.253.854.90
std0.040.020.010.030.020.010.140.070.06
Univdorm_Malachimean1.952.364.801.432.224.974.227.7131.57
std0.170.451.150.110.351.180.511.8312.41
Univdorm_Mitchmean2.753.003.521.891.641.957.584.966.12
std0.040.180.220.040.090.120.290.470.50
Univclass_Sebmean1.411.721.560.830.830.981.161.171.98
std0.020.020.020.010.010.010.040.020.07
Univlab_Susanmean4.407.029.230.841.231.291.944.702.79
std0.260.370.950.060.050.120.260.070.34
Office_Stellamean1.332.041.500.910.821.141.401.182.20
std0.070.070.100.050.020.080.180.050.28
Office_Glennmean2.382.983.240.940.680.851.690.751.12
std0.070.150.530.030.030.130.190.070.30
Apt_Moonmean23.93118.9921.520.080.160.240.020.070.13
std1.260.860.360.000.000.000.000.000.00
Apt_Phobosmean34.1270.8919.590.170.140.300.070.040.18
std1.141.720.110.000.000.010.000.000.01
Univexhibit_Hermesmean5.247.915.200.010.020.010.000.000.00
std0.150.120.230.000.000.000.000.000.00
Univdorm_Athenamean2.792.585.252.852.884.4013.0513.8628.51
std0.050.021.150.060.040.860.470.4410.02
Univmulti_Zeusmean1.832.533.111.712.201.985.267.988.44
std0.010.160.220.020.120.160.150.700.54
Univmulti_Aresmean3.394.526.220.731.551.981.195.007.10
std0.080.100.330.020.020.070.070.080.38
Primclass_Umamean7.036.766.840.650.670.691.671.491.48
std0.100.090.110.000.010.020.020.060.09
Primclass_Umarmean8.306.849.601.081.091.523.733.726.72
std0.210.140.180.030.010.010.230.080.10
Univdorm_Unamean5.304.995.520.881.321.282.424.184.51
std0.120.060.250.020.010.040.090.250.22
Primclass_Ulyssesmean11.9710.6615.311.411.752.087.8710.1313.39
std0.310.621.600.030.080.180.340.521.06
Table 10. The MAPE, MSE, and MAE of forecasting using WT-LSTM model.
Table 10. The MAPE, MSE, and MAE of forecasting using WT-LSTM model.
Dataset MAPE (%)MAEMSE
MetricsTrainValTestTrainValTestTrainValTest
Univclass_Boydmean2.272.132.320.520.490.490.450.410.43
std0.030.030.150.010.010.030.010.010.04
Univlab_Bethanymean3.673.773.722.792.782.7117.4016.8015.66
std0.130.060.080.090.040.021.400.640.61
Office_Evelynmean6.936.468.5214.9413.8718.26778.93577.211101.00
std0.280.140.350.120.180.2627.2948.7616.85
Office_Bobbimean7.138.969.245.315.876.3653.0164.2276.37
std0.500.520.510.300.170.174.161.832.95
Univdorm_Malachimean5.695.376.704.485.527.3246.4259.4392.47
std0.120.040.110.110.020.161.740.434.79
Univdorm_Mitchmean8.958.188.126.224.634.6471.1240.0739.44
std0.050.090.090.040.030.040.770.320.23
Univclass_Sebmean4.956.385.582.933.273.5817.1323.8326.35
std0.290.070.150.160.030.121.900.562.36
Univlab_Susanmean9.1116.5914.422.043.222.5713.3029.6617.03
std0.180.320.290.050.050.070.340.880.51
Office_Stellamean4.936.945.583.282.954.2421.6619.0432.83
std0.270.200.150.190.060.121.560.281.34
Office_Glennmean7.918.018.893.311.982.4422.419.1212.21
std0.010.150.050.010.010.010.100.070.05
Apt_Moonmean40.73192.7646.550.230.380.570.180.460.77
std0.443.320.480.000.000.000.000.000.00
Apt_Phobosmean67.95184.6259.320.480.400.820.680.401.17
std4.9415.601.250.000.000.010.010.010.02
Univexhibit_Hermesmean10.0615.137.550.040.070.020.010.030.00
std0.250.060.040.000.000.000.000.000.00
Univdorm_Athenamean9.6810.2012.479.9211.2510.68163.75229.93189.17
std0.240.080.370.170.110.244.293.248.39
Univmulti_Zeusmean6.126.728.396.046.335.8376.1693.8980.47
std0.070.050.040.060.030.050.920.901.85
Univmulti_Aresmean11.5715.1219.662.605.326.8117.1054.4190.39
std0.270.140.500.050.020.130.430.583.07
Primclass_Umamean19.6118.8018.361.691.771.8111.6112.2112.83
std0.360.340.320.040.030.030.260.330.23
Primclass_Umarmean22.2617.0928.613.133.104.4532.2729.2947.92
std0.810.430.490.080.030.011.140.720.16
Univdorm_Unamean21.5417.9618.512.924.313.8219.6333.7528.21
std1.520.901.090.290.200.312.042.443.73
Primclass_Ulyssesmean28.7222.0529.783.654.595.3871.4292.63109.26
std0.980.500.380.180.110.101.132.142.52
Table 11. Comparison of algorithms based on MAPE (%).
Table 11. Comparison of algorithms based on MAPE (%).
DatasetTrainingValidationTesting
RNNLSTMEEMD LSTMWT LSTM RNN LSTMEEMD LSTMWT LSTM RNN LSTMEEMD LSTM WT LSTM
Univclass_Boyd1.761.950.63 *2.272.051.890.71 *2.132.22.081.66 *2.32
Univlab_Bethany2.732.781 *3.673.0331.09 *3.772.942.81.05 *3.72
Office_Evelyn4.544.611.99 *6.934.074.631.47 *6.466.46.012.65 *8.52
Office_Bobbi5.475.721.9 *7.138.17.872.3 *8.968.027.832.43 *9.24
Univdorm_Malachi4.754.761.95 *5.694.944.712.36 *5.376.055.774.8 *6.7
Univdorm_Mitch7.518.12.75 *8.958.517.983 *8.188.397.853.52 *8.12
Univclass_Seb3.693.641.41 *4.954.934.751.72 *6.384.113.971.56 *5.58
Univlab_Susan7.767.64.4 *9.1113.9813.657.02 *16.5910.5411.119.23 *14.42
Office_Stella3.333.491.33 *4.935.375.322.04 *6.944.024.071.5 *5.58
Office_Glenn6.546.772.38 *7.917.526.712.98 *8.018.347.713.24 *8.89
Apt_Moon36.5534.3123.93 *40.73159.74128.03118.9 *192.7642.1239.3821.52 *46.55
Apt_Phobos54.3547.9234.12 *67.95151.33116.1270.89 *184.6252.9949.0619.59 *59.32
Univexhibit_Hermes6.155.955.24 *10.069.739.277.91 *15.135.444.98 *5.27.55
Univdorm_Athena7.638.052.79 *9.688.488.352.58 *10.210.139.695.25 *12.47
Univmulti_Zeus5.215.291.83 *6.125.795.662.53 *6.727.848.023.11 *8.39
Univmulti_Ares7.357.363.39 *11.5711.5911.214.52 *15.1215.5815.426.22 *19.66
Primclass_Uma18.7119.697.03 *19.6118.1317.96.76 *18.817.5817.276.84 *18.36
Primclass_Umar19.820.598.3 *22.2615.3114.966.84 *17.0926.6424.429.6 *28.61
Univdorm_Una16.618.335.3 *21.5415.8715.334.99 *17.9618.116.715.52 *18.51
Primclass_Ulysses26.5527.0911.97 *28.7223.1122.9610.66 *22.0531.5128.8515.31 *29.78
*: best value.
Table 12. The cross-validation result of forecasting using LSTM without decomposition.
Table 12. The cross-validation result of forecasting using LSTM without decomposition.
Dataset MAPE (%)MAEMSE
MetricsTrainValTestTrainingMetricsTrainValTestTesting
Univclass_Boydmean1.981.981.990.450.460.440.340.350.33
std0.070.130.130.010.040.040.020.050.06
Univlab_Bethanymean2.963.063.082.242.322.2810.5210.8410.39
std0.230.160.290.170.180.301.742.172.75
Office_Evelynmean5.634.805.3712.8210.5511.57669.14437.24468.07
std1.302.401.683.555.983.80296.19523.21320.99
Office_Bobbimean5.796.877.224.345.045.0738.4548.0647.68
std0.130.910.810.160.710.533.4813.678.96
Univdorm_Malachimean4.865.295.883.554.635.9130.7645.0172.17
std0.520.801.180.681.722.159.9627.5148.69
Univdorm_Mitchmean8.138.098.395.505.585.4360.9158.4156.46
std0.611.080.920.240.991.083.7917.4221.47
Univclass_Sebmean4.284.534.642.552.722.7111.9714.1213.77
std0.820.680.640.440.580.414.466.714.08
Univlab_Susanmean9.6312.2212.632.082.502.3813.4518.0816.36
std2.952.932.300.490.730.595.2410.418.70
Office_Stellamean3.894.665.732.662.723.1915.4914.9019.74
std0.521.252.370.420.550.594.565.225.80
Office_Glennmean7.197.308.493.192.542.4720.1313.6212.25
std0.661.021.760.491.050.824.8610.118.00
Apt_Moonmean28.0582.00110.990.230.310.380.150.320.43
std16.1558.7079.860.010.080.110.050.190.23
Apt_Phobosmean42.1078.63130.290.500.450.530.710.600.65
std17.1148.48117.080.090.160.160.160.330.30
Univexhibit_Hermesmean5.458.019.190.020.030.050.010.010.02
std1.342.834.450.010.030.040.010.010.02
Univdorm_Athenamean8.779.7510.538.4110.5211.53121.42184.66255.98
std1.102.131.290.441.222.9411.4936.83165.99
Univmulti_Zeusmean5.775.636.235.245.395.5957.6258.0967.20
std0.570.521.130.250.560.536.0411.5515.04
Univmulti_Aresmean7.5910.4412.371.623.223.997.6825.1335.01
std0.832.342.290.381.361.253.3617.9917.34
Primclass_Umamean20.3524.1520.051.742.051.7211.9116.0210.39
std1.9212.263.630.270.980.382.7614.374.17
Primclass_Umarmean19.9423.8122.262.773.563.3829.9245.4236.31
std2.0510.095.260.271.200.474.9931.719.52
Univdorm_Unamean15.9924.2621.372.323.463.4515.6930.0426.86
std3.1912.877.790.270.780.493.5215.268.09
Primclass_Ulyssesmean28.5432.6230.103.954.904.7579.57103.8688.49
std2.3416.086.460.451.840.808.3262.3021.51
Table 13. The cross-validation result of forecasting using RNN without decomposition.
Table 13. The cross-validation result of forecasting using RNN without decomposition.
Dataset MAPE (%)MAEMSE
MetricsTrainValTestTrainValTestTrainValTest
Univclass_Boydmean1.772.152.190.400.490.490.290.400.40
std0.070.140.190.020.040.060.020.070.09
Univlab_Bethanymean2.893.163.252.172.382.409.9311.1511.05
std0.350.330.450.230.310.411.983.223.55
Office_Evelynmean5.475.295.7011.9811.1611.90543.37413.15453.10
std1.192.952.082.976.884.49198.38495.12346.28
Office_Bobbimean5.567.097.434.145.175.2235.5049.3849.74
std0.200.910.690.140.780.502.4514.418.89
Univdorm_Malachimean5.686.076.653.875.016.4631.9747.9184.84
std1.542.031.751.011.462.2911.4324.3265.81
Univdorm_Mitchmean7.819.269.505.246.336.0755.5970.7467.28
std0.711.351.070.411.261.155.3623.6324.65
Univclass_Sebmean4.404.945.052.642.972.9713.0116.8416.74
std0.820.781.020.510.740.635.509.197.23
Univlab_Susanmean9.9012.9313.141.962.472.3110.9516.4514.57
std2.873.403.410.330.680.582.398.787.26
Office_Stellamean3.925.046.282.632.863.3613.8814.9920.39
std0.721.683.360.450.740.853.946.177.30
Office_Glennmean7.118.389.453.082.802.7318.6815.3714.02
std0.952.062.110.410.970.833.589.648.07
Apt_Moonmean34.71106.15130.170.220.340.400.150.330.44
std28.4878.76104.870.030.090.110.050.190.22
Apt_Phobosmean47.12116.85163.980.450.490.570.600.620.67
std25.0286.25137.740.050.170.150.090.380.31
Univexhibit_Hermesmean5.918.429.810.020.040.050.010.010.02
std1.752.844.800.010.030.050.010.010.03
Univdorm_Athenamean8.709.4510.878.2110.4012.28118.95178.36312.43
std1.441.472.041.050.854.1328.9624.06274.21
Univmulti_Zeusmean5.996.066.695.385.786.1358.7065.7182.24
std0.780.630.960.520.680.969.3312.7228.88
Univmulti_Aresmean7.8210.7412.561.623.334.057.4026.6936.11
std1.562.532.390.481.461.303.6619.5117.74
Primclass_Umamean20.9427.5221.811.772.231.8411.3416.5910.79
std2.6414.964.160.241.010.372.0914.083.30
Primclass_Umarmean21.2627.6025.242.783.833.5627.9249.3637.78
std3.5914.096.780.321.370.474.6635.839.91
Univdorm_Unamean14.3125.3320.671.983.463.3712.1029.6725.06
std3.5114.225.940.320.820.193.4616.162.29
Primclass_Ulyssesmean33.9442.5437.124.055.485.1177.71116.3790.65
std6.5326.5211.830.672.331.0315.0183.5227.24
Table 14. The cross-validation result of forecasting using EEMD-LSTM.
Table 14. The cross-validation result of forecasting using EEMD-LSTM.
Dataset MAPE (%)MAEMSE
MetricsTrainValTestTrainValTestTrainValTest
Univclass_Boydmean0.630.710.880.140.160.190.030.040.06
std0.030.130.300.010.030.060.000.010.03
Univlab_Bethanymean1.051.071.230.790.800.891.181.161.39
std0.070.060.320.060.090.210.200.320.58
Office_Evelynmean2.472.012.425.164.254.9285.8269.4764.00
std0.741.301.231.683.082.4251.94115.6959.93
Office_Bobbimean1.962.172.431.401.561.653.464.214.71
std0.080.180.240.070.230.180.351.220.85
Univdorm_Malachimean2.071.785.791.361.505.133.594.4547.03
std0.420.263.180.310.582.941.413.0644.22
Univdorm_Mitchmean2.872.855.971.861.893.707.327.0724.62
std0.310.592.700.090.372.000.592.5221.42
Univclass_Sebmean1.501.592.270.880.911.261.301.402.78
std0.090.100.900.050.110.450.150.371.69
Univlab_Susanmean4.545.536.190.851.041.051.863.223.16
std0.621.181.150.090.330.250.252.241.94
Office_Stellamean1.372.6111.090.931.304.761.443.9350.78
std0.122.3212.870.100.854.550.295.5464.65
Office_Glennmean2.312.348.091.000.772.351.941.2012.21
std0.150.326.120.120.231.940.450.8014.86
Apt_Moonmean17.6359.0377.370.080.130.250.020.050.13
std14.9948.5038.830.010.050.130.010.030.10
Apt_Phobosmean26.9149.4677.260.210.190.340.100.100.28
std14.4931.3444.270.050.090.290.040.090.39
Univexhibit_Hermesmean5.186.9315.170.010.020.050.000.000.01
std0.551.638.850.000.010.040.000.000.01
Univdorm_Athenamean3.142.844.942.943.095.2813.8815.4757.58
std0.380.412.770.220.273.442.042.5275.26
Univmulti_Zeusmean2.062.143.901.791.933.335.836.3422.32
std0.250.442.180.120.192.030.791.1126.55
Univmulti_Aresmean3.543.967.390.671.152.000.913.227.38
std0.420.783.800.150.480.430.462.532.59
Primclass_Umamean7.7411.8710.410.721.000.881.823.832.32
std1.019.634.440.100.660.260.404.301.42
Primclass_Umarmean8.3211.3810.391.081.571.433.899.906.80
std0.906.473.010.100.720.270.769.332.96
Univdorm_Unamean5.219.437.850.921.481.482.446.415.02
std0.688.073.610.090.690.390.556.021.85
Primclass_Ulyssesmean13.4217.7418.041.642.292.149.6120.1313.61
std1.4911.357.260.201.260.641.6119.116.44
Table 15. The cross-validation result of forecasting using WT-LSTM.
Table 15. The cross-validation result of forecasting using WT-LSTM.
Dataset MAPE (%)MAEMSE
MetricsTrainValTestTrainValTestTrainValTest
Univclass_Boydmean2.302.242.270.520.520.510.460.460.44
std0.060.140.150.010.040.050.020.060.08
Univlab_Bethanymean4.044.074.313.063.123.1720.7521.1320.06
std0.420.240.450.340.350.415.216.864.97
Office_Evelynmean8.647.178.5019.3415.3617.991159.95742.98961.04
std1.572.001.904.385.634.44421.95513.37404.27
Office_Bobbimean7.768.378.945.736.116.1862.4071.4070.87
std0.261.130.950.211.050.694.8824.6116.28
Univdorm_Malachimean5.885.836.544.165.106.5738.9153.3287.12
std0.390.661.410.611.982.5211.4032.0658.59
Univdorm_Mitchmean9.708.509.216.275.745.8072.2658.6661.10
std1.131.381.690.240.951.324.7115.5224.02
Univclass_Sebmean5.606.126.473.353.673.7723.8029.2829.66
std0.991.060.850.630.730.348.7913.466.30
Univlab_Susanmean12.1814.8515.932.533.002.9016.7723.8321.88
std2.003.732.790.300.830.542.4912.1710.01
Office_Stellamean5.796.067.413.963.644.3631.2626.8836.13
std0.570.581.350.580.740.638.419.8710.53
Office_Glennmean8.398.139.233.772.872.8028.0718.3516.85
std0.311.051.090.491.130.925.6312.8111.52
Apt_Moonmean37.94114.79131.960.260.330.410.180.340.46
std27.4289.4095.900.020.100.120.040.210.25
Apt_Phobosmean58.37100.42134.130.600.470.580.920.620.69
std22.8958.9363.670.140.190.210.300.400.34
Univexhibit_Hermesmean8.4312.0613.580.030.050.060.010.020.03
std2.175.106.910.020.030.050.010.020.03
Univdorm_Athenamean11.6811.1013.3011.1212.0714.95210.67245.96432.71
std2.332.111.942.281.154.7691.0543.07312.91
Univmulti_Zeusmean6.706.497.296.136.386.8077.4989.44106.63
std0.610.380.720.250.820.873.9924.3526.57
Univmulti_Aresmean12.6214.3616.272.524.355.2215.0542.1857.08
std0.652.452.160.511.691.476.1628.2626.77
Primclass_Umamean21.3224.9521.782.062.212.0415.0917.8313.81
std2.6913.292.990.340.980.373.3013.894.91
Primclass_Umarmean23.7425.8626.543.503.964.1537.8045.5245.54
std4.309.475.030.490.990.327.9123.527.67
Univdorm_Unamean17.3722.5023.062.593.373.9416.2926.2530.73
std4.039.726.840.340.610.894.3412.4512.21
Primclass_Ulyssesmean34.3037.1334.275.025.575.52103.54118.14109.34
std4.4017.177.330.701.720.8116.7557.7130.24
Table 16. Comparison of algorithms based on MAPE (%) using time series cross-validation.
Table 16. Comparison of algorithms based on MAPE (%) using time series cross-validation.
DatasetTrainingValidationTesting
RNNLSTMEEMD LSTMWT LSTMRNNLSTMEEMD LSTMWT LSTMRNNLSTMEEMD LSTMWT LSTM
Univclass_Boyd1.771.980.63 *2.32.151.980.71 *2.242.191.990.88 *2.27
Univlab_Bethany2.892.961.05 *4.043.163.061.07 *4.073.253.081.23 *4.31
Office_Evelyn5.475.632.47 *8.645.294.82.01 *7.175.75.372.42 *8.5
Office_Bobbi5.565.791.96 *7.767.096.872.17 *8.377.437.222.43 *8.94
Univdorm_Malachi5.684.862.07 *5.886.075.291.78 *5.836.655.885.79 *6.54
Univdorm_Mitch7.818.132.87 *9.79.268.092.85 *8.59.58.395.97 *9.21
Univclass_Seb4.44.281.5 *5.64.944.531.59 *6.125.054.642.27 *6.47
Univlab_Susan9.99.634.54 *12.1812.9312.225.53 *14.8513.1412.636.19 *15.93
Office_Stella3.923.891.37 *5.795.044.662.61 *6.066.285.73 *11.097.41
Office_Glenn7.117.192.31 *8.398.387.32.34 *8.139.458.498.09 *9.23
Apt_Moon34.7128.0517.63 *37.94106.158259.03 *114.79130.2110.9977.37 *131.96
Apt_Phobos47.1242.126.91 *58.37116.8578.6349.46 *100.42163.9130.2977.26 *134.13
Univexhibit_Hermes5.915.455.18 *8.438.428.016.93 *12.069.819.19 *15.1713.58
Univdorm_Athena8.78.773.14 *11.689.459.752.84 *11.110.8710.534.94 *13.3
Univmulti_Zeus5.995.772.06 *6.76.065.632.14 *6.496.696.233.9 *7.29
Univmulti_Ares7.827.593.54 *12.6210.7410.443.96 *14.3612.5612.377.39 *16.27
Primclass_Uma20.9420.357.74 *21.3227.5224.1511.87 *24.9521.8120.0510.41 *21.78
Primclass_Umar21.2619.948.32 *23.7427.623.8111.38 *25.8625.2422.2610.39 *26.54
Univdorm_Una14.3115.995.21 *17.3725.3324.269.43 *22.520.6721.377.85 *23.06
Primclass_Ulysses33.9428.5413.42 *34.342.5432.6217.74 *37.1337.1230.118.04 *34.27
*: best value
Table 17. Statistics’ test results.
Table 17. Statistics’ test results.
ComparisonMean Std. Deviationp-Value of Levene’s Testp-Value of T-Test
LSTM0.1530.145-
EEMD-LSTM0.0760.062
WT-LSTM0.1640.135
LSTM vs. EEMD-LSTM- 0.008   Reject   H 0 0.037   Reject   H 0
LSTM vs. WT-LSTM- 0.852   Accept   H 0 0.813   Accept   H 0
EEMD-LSTM vs. WT-LSTM- 0.008   Reject   H 0 0.013   Reject   H 0
Table 18. Recommendation of the forecasting method.
Table 18. Recommendation of the forecasting method.
UnitsFunctionalitiesTime ZonesRecommendation
Office_BobbiOfficeAmerica/Los_AngelesEEMD-LSTM
Office_EvelynOfficeAmerica/Los_AngelesEEMD-LSTM
UnivClass_BoydCollege ClassroomAmerica/Los_AngelesEEMD-LSTM
UnivLab_BethanyCollege LaboratoryAmerica/Los_AngelesEEMD-LSTM
UnivDorm_MalachiResidentialAmerica/ChicagoEEMD-LSTM
UnivDorm_MitchResidentialAmerica/ChicagoEEMD-LSTM
Office_GlennOfficeEurope/LondonEEMD-LSTM
Office_StellaOfficeEurope/LondonEEMD-LSTM
UnivClass_SebCollege ClassroomEurope/LondonEEMD-LSTM
UnivLab_SusanCollege LaboratoryEurope/LondonEEMD-LSTM
Apt_MoonApartmentAmerica/New_YorkEEMD-LSTM
Apt_PhobosApartmentAmerica/New_YorkEEMD-LSTM
UnivDorm_AthenaResidentialAsia/TaiwanEEMD-LSTM
UnivExhibit_HermesExhibitionAsia/TaiwanEEMD-LSTM
UnivMulti_AresMultipurposeAsia/TaiwanEEMD-LSTM
UnivMulti_ZeusMultipurposeAsia/TaiwanEEMD-LSTM
PrimClass_UlyssesSecondary ClassroomAsia/SingaporeEEMD-LSTM
PrimClass_UmaSecondary ClassroomAsia/SingaporeEEMD-LSTM
PrimClass_UmarSecondary ClassroomAsia/SingaporeEEMD-LSTM
UnivDorm_UnaResidentialAsia/SingaporeEEMD-LSTM
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chou, S.-Y.; Dewabharata, A.; Zulvia, F.E.; Fadil, M. Forecasting Building Energy Consumption Using Ensemble Empirical Mode Decomposition, Wavelet Transformation, and Long Short-Term Memory Algorithms. Energies 2022, 15, 1035. https://doi.org/10.3390/en15031035

AMA Style

Chou S-Y, Dewabharata A, Zulvia FE, Fadil M. Forecasting Building Energy Consumption Using Ensemble Empirical Mode Decomposition, Wavelet Transformation, and Long Short-Term Memory Algorithms. Energies. 2022; 15(3):1035. https://doi.org/10.3390/en15031035

Chicago/Turabian Style

Chou, Shuo-Yan, Anindhita Dewabharata, Ferani E. Zulvia, and Mochamad Fadil. 2022. "Forecasting Building Energy Consumption Using Ensemble Empirical Mode Decomposition, Wavelet Transformation, and Long Short-Term Memory Algorithms" Energies 15, no. 3: 1035. https://doi.org/10.3390/en15031035

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop