Next Article in Journal
Quantifying National-Scale Changes in Agricultural Land Exposure to Fluvial Flooding
Previous Article in Journal
Democratising Heritage Values: A Methodological Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparison of the Deep Learning Performance for Short-Term Power Load Forecasting

Department of Software Engineering, Artificial Intelligence Convergence College, Chonnam National University, Buk-gu, Gwangju 61186, Korea
Sustainability 2021, 13(22), 12493; https://doi.org/10.3390/su132212493
Submission received: 18 August 2021 / Revised: 27 October 2021 / Accepted: 2 November 2021 / Published: 12 November 2021

Abstract

:
Electricity demand forecasting enables the stable operation of electric power systems and reduces electric power consumption. Previous studies have predicted electricity demand through a correlation analysis between power consumption and weather data; however, this analysis does not consider the influence of various factors on power consumption, such as industrial activities, economic factors, power horizon, and resident living patterns of buildings. This study proposes an efficient power demand prediction using deep learning techniques for two industrial buildings with different power consumption patterns. The problems are presented by analyzing the correlation between the power consumption and weather data by season for industrial buildings with different power consumption patterns. Four models were analyzed using the most important factors for predicting power consumption and weather data (temperature, humidity, sunlight, solar radiation, total cloud cover, wind speed, wind direction, humidity, and vapor pressure). The prediction horizon for power consumption forecasting was kept at 24 h. The existing deep learning methods (DNN, RNN, CNN, and LSTM) cannot accurately predict power consumption when it increases or decreases rapidly. Hence, a method to reduce this prediction error is proposed. DNN, RNN, and LSTM were superior when using two-year electricity consumption rather than one-year electricity consumption and weather data.

1. Introduction

The average global temperature is steadily rising due to global warming. In July 2018, the Korean government announced a policy to reduce 37% greenhouse gases by 2030 [1]. Therefore, the Korean government has proposed technologies and policies to reduce coal power generation and drastically increase renewable energy. In 2018, Korea experienced the ‘worst heatwave ever recorded’ based on 110 years of recorded meteorological observations. Figure 1 shows the results of the power consumption analysis based on a variety of applications in Korea in 2018 [2]. Figure 1a shows the power consumption that includes industrial (55.70%), general (22.20%), residential (13.90%), agricultural (3.50%), educational (1.60%), late-night (2.40%), and street-light applications (0.70%). Figure 1b shows the rate of increase in power consumption by application in 2018 compared to the previous year (2017). The results were analyzed as follows.
  • The rate of power consumption by application increased; this included agricultural (7.30%), residential (6.30%), general (5.10%), educational (4.30%), industrial (2.50%), late-night (2%), and street-lighting applications (0.70%), which are the most sensitive to seasonal factors.
  • The annual domestic power consumption exceeded 70,000 GWh for the first time in 2018. This is expected to increase steadily due to future heat waves.
  • Industrial use increased by 2.5% compared to the previous year, but it accounted for 55.70%. Thus, the power consumption increased considerably to over 2.92 million GWh. It was determined that the boom in the exports of major energy-consuming industries, such as domestic semiconductors and petrochemicals, was affected.
  • This study proves that seasonal and economic factors influence power consumption [3,4,5].
In addition, the existing electricity demand forecasting studies have been steadily progressing according to the forecast horizon, data characteristics, and data combinations.
1.
The forecast horizon of electricity demand is essential because of the maintenance schedule of the energy management system (EMS) [6], the generation of power system utilities, expansion of operations and plans [7], load switching, safety, market demand assessment, cost reduction, and guarantee of continuous power supply [8]. Forecast horizons can be divided into four types [9,10,11].
Very short-term load forecasting (VSTLF) predicts power consumption and demand in real-time by performing predictions for 1 h, 30 min, and 15 min to stably operate power.
Short-term load forecasting (STLF) is the most used widely used technique that can perform predictions from 1 h to 1 week, and the forecasting stage is time-based. This prediction plays an essential role in operating the power system and plans the arrival and departure of units and optimal load distribution.
Medium-term load forecasting (MTLF) can perform predictions from 1 month to 1 year and includes a daily forecasting step. This prediction is essential for medium-term planning, including the economic operation and repair planning of a system that are directly related to the system’s reliability.
Long-term load forecasting (LTLF) performs long-term predictions of over one year. It has a forecast range of over ten years, with a monthly forecasting step. This prediction is essential in network development, such as increasing the number of power plants, transmission lines, and distribution equipment.
2.
Data characteristics can be divided into data-driven models (single model) and hybrids (combination model).
Data-driven models predict the collected data by using probability and statistics (exponential smoothing [12], autoregressive integrated moving average [13]), classification models (k-nearest neighbors [14], decision trees [15], support vector machines [16], particle swarm optimization [17]), and artificial intelligence (artificial neural networks [18], deep neural networks [19], recurrent neural network [20], long short-term memory [21], and convolution neural networks [22]). This single model has no particular computational complexity, but its results are unreliable and inaccurate. Hence, the combined model is a good alternative to improve the load prediction accuracy and stability by combining the advantages and disadvantages of a single model [23].
As for the hybrid model, various models that combine a single model were proposed: (1) EMD–SVM model to predict power demand [24]; (2) random forest-expert model [25], PSO–SVM model [26], and neuro-fuzzy inference–whale optimization algorithm model [27] for short-term load prediction; (3) firefly algorithm–SVR model to predict electrical load related to air conditioning [28]; (4) a combination of CNN and LSTM to extract complex features related to individual energy consumption prediction [29]; and (5) DeepAR, which is an AR-based RNN that combines statistical AR methods and RNN with deep learning [30] (similar to LSTNet [31], TPA [32], DSANet [33], MTNet [34], DSSM [35], DeepTCN [36], DFM-RF [37], DeepGLO [38], and ForecastNet [39]).
3.
A model can be categorized as univariate or multivariate according to the data composition (combination). A univariate model performs predictions by only using power consumption. A multivariable model performs predictions by combining the power consumption and weather data (temperature, humidity, sunlight, solar radiation, total cloudiness, wind speed, wind direction, humidity, and vapor pressure). In addition, many recent multivariate models have been used for short-term wind speed prediction. A model applying solar production, temperature, insolation, humidity, and wind speed [40] has been used.
The critical factors for forecasting the building power consumption vary according to the weather and indoor conditions, the size and use of the buildings, and the prediction horizons [41,42,43,44]. Among these factors, (1) meteorological conditions are the main predictive factors that can change the indoor conditions and the activities that govern the building power consumption. For example, ambient temperature, cloud cover, humidity, and insolation affect the load pattern due to the comfort of the occupants and adjustment of the lighting levels [45,46]. (2) Commercial buildings that are more than ten times larger than typical residential buildings will most likely consume more electricity than weather variation factors. For example, the load patterns of the buildings vary according to usage and time. This includes single-family houses, multifamily houses, cultural and assembly facilities, religious facilities, education, and research facilities, factories, and warehouse facilities. The varying complexity of load patterns makes it difficult to forecast the building power consumption accurately. (3) Mid/long-term and short-term power consumption forecasting can efficiently operate peer-to-peer (P2P) power transactions and EMS that are suitable for power supply and demand. In particular, changes in consumers’ power consumption patterns are uncertain factors for future power demand. Accurate demand forecasting begins by responding to fluctuations in electricity demand, considering weather/environmental changes, and reflecting recent demand patterns. The response to real-time prediction is brought about by frequency control and economic dispatch. Short-term forecasts account for power generation plans, and mid-term forecasts account for the maintenance of power facilities. Meanwhile, long-term forecasts account for construction plans of generators and transmission networks.
In this study, the seasonal factors and the building power consumption pattern are reflected and proposed to forecast the power consumption accurately.
  • Four models (TM1–TM4) are proposed, and an analysis of the correlation between the seasonal factors, weather data, and power consumption is conducted. These are univariate and multivariate models that have collected power consumption and weather data (temperature, humidity, sunlight, solar radiation, the total cloud cover, wind speed, wind direction, humidity, and vapor pressure) for 1–2 years.
  • The prediction horizon for power consumption forecasting was 24 h, which ensured that power suppliers operated peer-to-peer power transactions and energy management systems efficiently.
  • The proposed four models are compared and analyzed by applying a DNN, RNN, CNN, and LSTM to deep learning, which is currently an issue.
  • Two industrial buildings in the agricultural complex in Naju, Jeollanam-do, were selected to test and verify the proposed method. Building B is an industrial building with constant power consumption, which houses a manufacturing company. Building T is an industrial building with nonuniform power consumption, which houses a livestock processing company. In addition, the experimental data adopted in this study were collected for three years (from 2017 to 2019) by applying industrial electricity real-time rates. The usage was provided by the Korea Electric Power Corporation’s (KEPCO) i-smart [47].
The structure of this study is as follows. Section 2 analyzes the power consumption for Companies B and T’s buildings, which have different power consumption patterns. Section 3 briefly explains the experiments using deep learning methods. Section 4 describes the problems of deep learning, and it explains the proposed method. Section 5 describes the experiment and the analysis of the proposed method. Finally, Section 6 presents the conclusions of this study and the future directions of research.

2. Analysis of the Power Consumption of Companies B and T

Section 2.1 explains the meteorological factors affecting power consumption, which the Korea Meteorological Administration recently announced. Section 2.2 analyzes the power consumption of Companies B and T located in Naju, Jeollanam-do, by season. Finally, Section 2.3 analyzes the correlation between the meteorological data and the power consumption by season.

2.1. Existing Research on the Building Electricity Consumption

In previous studies [48,49,50], the meteorological factors affecting the building power consumption include the temperature (average, maximum, and minimum), humidity, wind speed, cloudiness, discomfort index, perceived temperature, and precipitation [48]. In Korea, summer (July–August) and winter (December–February) highly correlate with the minimum temperature, cloudiness, a sensible temperature, and power consumption. During the changing seasons (May and October), the correlation between the meteorological factors and the power consumption is relatively low [49]. When forecasting the power consumption using only the temperature, the prediction error was 1.8%. In contrast, when weather factors such as humidity, cloudiness, sensible temperature, wind speed, and precipitation are added, the prediction error is reduced to 1.3%. When forecasting the power consumption, the prediction error can be improved by 25% depending on the combination of meteorological factors; the annual power generation amount can be reduced by approximately 1100 GWh, saving about 120 billion KRW [50]. However, this analysis has the disadvantage in that it may differ slightly from the actual electricity demand forecast since it does not consider variables such as lifestyle, industrial activities, and economic factors.

2.2. Analysis of the Power Consumption by Season

To analyze the power consumption patterns of Companies B and T, the months of April, July, October, and January were selected, which represent Korea’s spring, summer, fall, and winter, respectively. Figure 2 shows a comparison of the power consumption of each month (per 24 h) for Companies B and T during spring (April), summer (July), fall (October), and winter (January). Company B’s building uses electricity from 9 AM to 12 PM and from 1 to 6 PM on weekdays. It does not use electricity during hours other than holidays and working hours. By comparing the power consumption by season, the order is winter > fall > summer > spring. In the winter, the power consumption is high. In contrast, the power consumption is low in the summer (July) since this is the vacation season.
The power consumption of Company T has an irregular pattern. This is because Company T is an agroprocessed food company that consumes electricity according to economic factors (consumer demand) without affecting the season, time, or public holidays.
Table 1 shows the average (average), standard deviation (Std.), and the coefficient of variation (CV) for the seasonal power consumption of Companies B and T. The CV can compare the relative scope; however, it is not enough to calculate the range or variance. Therefore, to compare the power consumption of the different units of measurement, the CV is adopted (Equation (1)):
CV = σ m ¯
The larger the value of the CV, the larger the relative difference. In addition, σ and m ¯ represent the standard deviation and mean, respectively.
Company B has a relatively small power consumption than Company T; however, the mean, standard deviation, and CV are large. This is because there is a relatively significant difference between the working hours (9 AM to 6 PM) in which electricity is used and the nonworking hours wherein electricity is not in use. Company T consumes more electricity than Company B. Still, it consumes electricity regardless of the working hours, holidays, and weather factors. Therefore, the standard deviation and CV values are small.

2.3. Correlation between the Seasonal Weather Data and Power Consumption

Figure 3 shows the analysis of the correlation (R2) and trend line (Y) of the temperature and humidity by season for the Naju region where Companies B and T are located. The temperature and humidity in the spring, fall, and winter did not seem to have any correlation. The correlation (R2) was determined to be 0.76 for the summer temperature and humidity.
Figure 4 and Figure 5 show the trend line (Y) analysis and the correlations (R2) of Companies B and T’s temperature and power consumption by season, respectively; Company T has a higher correlation on average than Company B.

2.4. Concluding Remarks

Section 2 analyzed the relationships between temperature and humidity, seasonal meteorological factors, and temperature and power consumption. Although weather factors affect power consumption, it was determined that building habits, industrial activities, and economic factors have the greatest influence on power consumption.

3. Deep Learning for Power Consumption

Section 3.1 briefly describes the DNN, LSTM, RNN, and CNN. Section 3.2 presents the comparison and analysis of forecasting the power consumption using deep learning for Companies B and T.

3.1. Deep Learning

3.1.1. Deep Neural Network

DNN is an artificial neural network composed of several hidden layers between the input and output, as shown in Figure 6 [51,52]. DNNs can model complex nonlinear relationships, such as general artificial neural networks. For example, in a deep neural network structure of an object identification model, each object can be expressed as a hierarchical configuration of the essential elements of an image [53]. Here, the additional layers may aggregate the characteristics of the gradually gathered lower layers. This feature of the deep neural networks enables complex data modeling with fewer units (units, nodes) when compared to similarly performing artificial neural networks [51].

3.1.2. Recurrent Neural Network

RNN refers to a neural network. The connections between the units that constitute an artificial neural network comprise a directed cycle, as shown in Figure 7 [54]. Unlike forward neural networks [20], an RNN can use the memory inside the neural network to process arbitrary inputs. Due to these characteristics, RNNs are used in handwriting recognition fields and show a high recognition rate [55]. A variety of methods are used for structures that can construct RNNs. Hopfield networks [56], Elman networks [57], echo state networks (ESNs) [58], LSTMs [59], bidirectional RNNs [60], continuous-time RNNs (CTRNNs) [61], hierarchical RNNs (HRNNs) [62], and second-order RNNs (SORNNs) [63] are the typical examples. A gradient descent approach, Hessian-free optimization, and global optimization are typically used to train an RNN. However, the RNN has a scaling issue and is difficult to train when there are many neurons or input units [64,65].

3.1.3. Concurrent Neural Network

CNN is a multilayer perceptron designed to use minimal preprocessing, as shown in Figure 8 [66]. The CNN comprises one or several convolutional layers. There are general artificial neural network layers on top; it uses weights and pooling layers. Because of this structure, CNN can use the input data of the two-dimensional structure. CNNs perform well in video and audio fields [67,68]. CNNs can also be trained by using standard back-passing methods. CNNs are more accessible to training than the other feed-forward neural network techniques. They have the advantage of using fewer parameters. Recently, a convolutional deep belief network (CDBN) has been developed in deep learning, which is structurally similar to a CNN. It can use the two-dimensional structure of the figure well [69]. Deep belief networks (DBNs) can also take advantage of line training [70]. CDBN provides a general structure that can be used for a variety of image and signal processing techniques. It is used in several benchmark results for standard image data, such as those that belong to the Canadian Institute for Advanced Research (CIFAR) [71].

3.1.4. Long Short-Term Memory

LSTM is a method used to solve the problem of gradient extinction, which is a disadvantage of the RNN [64,65]. Compared to the basic RNN model, a study has reported that the accuracy is approximately 10% higher when the LSTM model is implemented [72,73]. The interior of the LSTM block comprises a memory cell with a circular structure and three types of gates (input, forget, and output), as shown in Figure 9. Modifications of the LSTM model include a single-layer LSTM [21], multilayer LSTM [74], gated recurrent unit (GRU) [75], bidirectional LSTM [76], encoder–decoder (ED)–LSTM [77], and con–LSTM [78]. Multiple LSTMs have an accuracy higher than single-layer LSTMs.

3.1.5. Persistence Model

The persistence model (PM) is the simplest way of producing a forecast [79,80]. A persistence model assumes that the future value of a time series is calculated under the assumption that nothing changes between the current time and the forecast time (Equation (2)).
P t ^ = O t 1
In Equation (2), O t 1 is the measured power consumption at the time (t − 1) and P t ^ is the predicted value at time t. Its accuracy decreases if the time horizon is longer than 1 h.

3.2. Comparison of the Deep Learning Performance for Companies B and T

In this section, we used the power consumption in the spring, summer, autumn, and winter in 2018 to compare the deep learning performance of Companies B and T. The experimental data were divided into learning data (70%) and test data (30%) and then simulated. Table 2 lists the training and testing options for simulating the PM, DNN, RNN, CNN, and LSTM. In the simulation, the learning rate, loss function (MSE), optimizer (ADAM) [81], and activity function (RELU) [82] were all the same. The remaining options were set for each deep learning method.
Table 3 compares the persistence model and the deep learning performance of Companies B and T by season. Figure 10 shows comparison of the persistence model and the deep learning performance of Companies B and T in April 2018. In addition, the average (Ave.), standard deviation (Std.), CV, root-mean-square error (RMSE) [83], and mean absolute percentage error (MAPE) [84] were used for the prediction errors. The values with excellent performance for each deep learning method are shown in bold. The RMSE and MAPE are shown in Equations (3) and (4), respectively.
RMSE = 1 n ( O i P i ) 2
  MAPE = 100 n i = 1 n | O i P i O i |
In Equations (3) and (4), n is the total number that is to be predicted, O i is the actual value, and P i is the predicted value. For RMSE and MAPE, the smaller the number, the better the performance.
Table 4 shows the average RMSE and MAPE values for the four seasons when using persistence model and deep learning. The RMSE of Company B with large CV fluctuations is: LSTM (16.05) > RNN (17.04) > DNN (17.11) > PM (17.93) > CNN (19.74). However, for MAPE, it is: PM (36.38) > RNN (54.97) > DNN (56.18) > LSTM (63.94) > CNN (93.74). The RMSE of Company T with small CV fluctuations is: LSTM (18.58) > DNN (19.02) > RNN (19.08) > PM (19.51) > CNN (21.24). For MAPE, LSTM (10.08) > DNN (11.70) > PM (11.77) > RNN (12.03) > CNN (13.25). Thus, Company B with large CV fluctuations has different priorities depending on the prediction error method, but Company T with small CV fluctuations has an excellent LSTM regardless of the prediction error method.

4. Proposed Method

Section 4.1 analyzes the power consumption patterns of Companies B and T, and it describes the problems. Section 4.2 proposes the prediction error correction to solve the problems when simulating traditional deep learning for Companies B and T.

4.1. Problem of Existing Deep Learning

Section 2 analyzes the relationship between the weather data and power consumption by season. This analysis reveals that the weather data, lifestyle, industrial activity, and economic factors affect power consumption. For example, Figure 11 compares the electricity usage pattern analysis for the daily averages by season for Companies B and T. Figure 11a shows that Company B has a constant power consumption pattern. Still, it fluctuates rapidly, depending on the period of power usage and lifestyle. Thus, the power consumption increases during the intensive working hours from 8 AM to 6 PM. Still, the power consumption decreases sharply during the rest of the hours (7 PM to 8 AM) and lunchtime (12 PM to 1 PM). Unlike Company B, as shown in Figure 11b, Company T does not have a uniform power consumption pattern. Still, the power consumption does not significantly increase or decrease.

4.2. Proposed Forecasting Error Correction

The application of DNN, RNN, CNN, and LSTM in different seasons revealed that the performance of the LSTM was excellent for deep learning. However, as shown in Figure 11a, when the power consumption increases and decreases rapidly, LSTM cannot accurately predict the power consumption; therefore, this study proposes a method to correct the prediction error when the power consumption rapidly increases or decreases. Figure 12 shows the flowchart proposed in this study.
First, as a data collection and preprocessing process, the electricity and weather data of Companies B and T in Naju were collected over 1 h intervals and preprocessed for simulation. Second, after splitting the electricity and weather data into four test models, each was divided into training data (70%) and test data (30%). Third, the training data of four test models were simulated using deep learning (DNN, RNN, CNN, and LSTM). The proposed prediction error correction is applied to the training data of four test models and simulated. Fourth, the learning deep learning module was applied to the test data using four test models. Finally, the performances of the traditional and proposed methods are compared with the prediction error measurement (RMSE, MAPE).

4.2.1. Data Collection and Preprocessing

The electricity and weather data were collected from 2017 to 2018 for Companies B and T in Naju, Jeollanam-do. The electricity data were collected from Companies B and T in one-hour intervals using i-Smart, which provides a real-time power consumption service from KEPCO. The weather data were collected in one-hour intervals from the Naju Regional Meteorological Administration. The collected electricity and weather data were preprocessed monthly (January, April, August, and October).

4.2.2. Proposed Four Models

This section proposes four models (TM1–TM4), as shown in Table 5, which combine the power consumption and weather data that affect the power consumption prediction. For TM1, one variable was univariate; for TM2 to TM4, multivariate variables with two or more variables were adopted. TM1 was the most frequently used power consumption variable for estimating the existing power consumption. The power consumption data for Companies B and T were collected by season in one-hour units in 2018. TM2 predicts the power consumption by collecting data over two years (from 2017 to 2018). TM3 comprises the 2018 power consumption and the weather data (e.g., temperature, humidity, sunlight, and cloud cover). Finally, TM4 includes the two-year power consumption and the weather data (e.g., temperature, humidity, sunshine, cloud cover, precipitation, wind direction, wind speed, and vapor pressure).
Figure 13 shows that the four test models split the training and test data ratio to 70:30. Four test models separated by the training data were simulated for each deep learning method using Table 2. The training data had different input variables for each TM, but all have synchronized data in units of 1 h. The test data of the four TMs predict the power consumption from 22 April 2018, 1:00 (a.m.), to 31 April 2018, 12:00 (p.m.). Figure 13a is a simulation of the April 2018 (spring) power data, with learning data from 1 April 2018, 1:00 (a.m.), to 21 April 2018, 12:00 (p.m.). Figure 13b uses data from April 2017 and April 2018, wherein two variables were synchronized each hour. The training data used in TM2 are from 1 April 2017, 1:00 (a.m.), and 1 April 2018, 1:00 (a.m.), to 21 April 2018, 12:00 (p.m.). Figure 13c,d shows five (power consumption, temperature, humidity, solar radiation, and cloud cover of April 2018) and nine variables (power consumption of two years, temperature, humidity, solar radiation, cloud cover, precipitation, wind speed, wind direction, and vapor pressure of April 2018), respectively.

4.2.3. Proposed Error Correction

Section 4.1 states that if the power consumption increases or decreases rapidly, it cannot be accurately predicted. Therefore, we propose a prediction error correction method shown in Equation (5) by considering the difference between the previous and current power consumptions.
P i ^ = O i P i + | O i O i 1 O i |
In Equation (5), O i and O i 1 represent the actual power consumption for i and i − 1 h, respectively. P i is the predicted power consumption, and P i ^ is the prediction error correction proposed in this study. Finally, the learned deep learning modules simulated four test data points for each deep learning method.

4.2.4. Prediction Error Measurement

This study adopted RMSE and MAPE to measure the prediction error of the existing and proposed methods using Equations (3) and (4) in Section 3.2. RMSE is the most commonly used prediction error measurement method [83]. MAPE is the most common measurement used to forecast error and works best if there are no extremes in the data (and no zeros) [84].

5. Simulation Results and Analysis

5.1. Test Environment

To verify the traditional and proposed methods, the experiments were performed on a workstation computer equipped with an Intel Xeon (R) W-2133, a 3.60 GHz CPU, and 32 GB of RAM (Dell Precision 5820 Tower Workstation). The operating system was Windows 10 Pro for the workstations (64 bit). The traditional and proposed methods were implemented using deep learning libraries provided by Tensor flow [85] and Keras [86].

5.2. Company B with a Constant Power Consumption Pattern

Section 2.2 states that Company B has a constant power consumption pattern. Table 6 shows the results of applying traditional deep learning and the proposed method to four TMs. Although there is a slight difference in the deep learning performance based on the month, CNN has the worst performance among the four TMs. However, the DNN and RNN exhibited an excellent TM performance. LSTM showed that the TM1, TM2, and TM3 performances are similar to those of DNN and RNN; however, TM4 did not perform well. Therefore, the power consumption prediction will perform well if the test model adopts TM1 or TM2 and simulates DNN, RNN, and LSTM.
Figure 14 shows a comparison of MAPE (%) for the traditional and the proposed methods for Company B in April 2018. Figure 14a is a performance comparison of four TMs using deep learning. The error between the traditional and the proposed methods appeared in TM4, TM3, TM2, and TM1. For DNN, RNN, and LSTM, the more variables were adopted, the greater the error. Figure 14b is a comparison of deep learning performance by TM. The error between the traditional and the proposed method appeared in the order CNN, LSTM, RNN, and DNN. In particular, CNN performed poorly in all TMs. For DNN, RNN, and LSTM, the smaller the variable, the smaller the error.
Figure 15 shows the results of comparing the performance of MAPE (%) by TMs and deep learning for the proposed method for Company B in January 2018. Figure 15a shows the deep learning performance of TMs, which exhibited the following order: DNN, RNN, LSTM, and CNN. Figure 15b shows the results of four test models for each deep learning method; the performance was excellent in the order of TM1, TM2, TM3, and TM4.
Figure 16 shows the performance analysis of the proposed DNN, RNN, CNN, and LSTM for the four TMs in January 2018. Figure 16a shows the proposed DNN performance analysis for the four TMs. The predicted values (DNN-TM1 and DNN-TM2), in which TM1 and TM2 were applied, have predictions similar to the actual value (power). The predicted values (DNN-TM3 and DNN-TM4) when applying TM3 and TM4 from 97 to 121 times showed that the predicted values were not similar to the actual values. Figure 16b shows the proposed RNN performance analysis for the four TMs. Like the proposed DNN results, the predicted values of the proposed RNNs (RNN-TM1 and RNN-TM2), in which TM1 and TM2 were applied, have been predicted to be similar to the actual values. However, the predicted value of TM3 (RNN-TM3) exhibits a higher error (RMSE: 25.70) from the actual value than that shown by the predicted value of TM4 (RNN-TM4); thus, the performance was poor. Figure 16c shows the performance analysis of the proposed CNN for the four TMs. Similar to the proposed DNN results, the predicted values (CNN-TM1, CNN-TM2) of the CNNs, in which TM1 and TM2 were applied, have been predicted to be similar to the actual values. However, the predicted value of TM4 (CNN-TM4) was too high (RMSE: 19.44) from the actual value; thus, the performance was not good compared to that of TM3 (CNN-TM3). Figure 16d shows the proposed LSTM performance analysis for the four TMs. Similar to the proposed DNN result, the predicted value (LSTM-TM1) of the LSTM, in which TM1 was applied, has been predicted to be like the actual value. However, the TM3 predicted value (LSTM-TM3) exhibited a high error (RMSE: 25.92) because the difference from the actual value was too large.
Figure 17 illustrates a performance comparison of DNN, RNN, CNN, and LSTM for TM2, which has an excellent performance among the four TMs. The deep learning performance comparison results for TM2 are in the following order: DNN-TM2 (RMSE: 15.05) > RNN-TM2 (RMSE: 15.46) > LSTM-TM2 (RMSE: 17.36) > CNN-TM2 (RMSE: 18.23).

5.3. Company T with the Irregular Power Consumption Pattern

Table 7 shows the profound learning performance results for Company T with irregular power consumption patterns. By comparing the performance of the four test models, the prediction error for each test model was not significant for Company T, unlike Company B. In other words, Company T has a slight difference in the deep learning performance based on the month, but similar results are observed for Company B, as follows: (1) The DNN and RNN are the four TMs with excellent performance. (2) The performances of the LSTM and CNN are similar to DNN and RNN in TM1–TM3, but TM4 did not perform well. Therefore, although Company T has an irregular power consumption pattern, it has an excellent performance when simulating DNN, RNN, and LSTM by adopting TM1 or TM2 for predicting the power consumption.
Figure 18 shows a comparison of MAPE (%) for the traditional and the proposed methods for Company T in April 2018. Figure 18a is a performance comparison of four TMs using deep learning. The error between the traditional and the proposed methods was almost similar to that observed for the TM performance, unlike Company B. Similar to Company B, Company T’s CNN performed poorly in all TMs. Figure 18b shows a comparison of the deep learning performance exhibited by TM. The error between the traditional and the proposed methods appeared in the following descending order: CNN, LSTM, RNN, and DNN. DNN, RNN, and LSTM performed similarly regardless of the number of variables.
Figure 19 shows the results of comparing the MAPE (%) performance by using TMs and deep learning for the method proposed for Company T in January 2018. Figure 19a shows the deep learning performance of TMs, which exhibited the following order: DNN, RNN, LSTM, and CNN. Figure 19b shows the results of four test models for each deep learning model; the performance exhibited the following order: TM2, TM3, TM1, and TM4.
Figure 20 shows the DNN, RNN, CNN, and LSTM proposed for the four TMs in January 2018 (winter) for Company T. Figure 20a shows the proposed DNN performance analysis for the four TMs. The four TMs displayed excellent performance. Thus, a building such as Company T, whose power consumption pattern is not constant, will perform well if the power consumption is predicted using the existing power consumption and weather data. In Figure 20b–d, the predicted values of RNN, CNN, and LSTM are applied with TM1–TM3, which have a similar performance to the actual values. However, the predicted values of RNN, CNN, and LSTM applied with TM4 have a large error difference when compared to the measured values; thus, the performance was poor.
Figure 21 compares the performances of the proposed DNN, RNN, CNN, and LSTM for TM2, which exhibits an excellent performance among the four TMs. Based on the deep learning performance comparison for TM2, it was shown in the order of DNN-TM2 (RMSE: 18.03) > RNN-TM2 (RMSE: 18.29) > CNN-TM2 (RMSE: 19.47) > LSTM-TM2 (RMSE: 19.70).

6. Conclusions and Future Research

Forecasting the building power consumption is necessary to improve the energy efficiency of buildings and overcome the power peak problem or energy crisis that occur every year worldwide. Building owners and emergency medical services operators need to manage their buildings’ electrical energy consumption. Since electrical energy is the main form of energy consumed in commercial buildings, the ability to predict the electrical energy consumption would benefit building owners and operators. Data-driven models for energy forecasting have been extensively studied over the past few decades due to their improved performance, robustness, and ease of deployment. Artificial neural networks are one of the most popular data-driven approaches applied to date among the different models.
In this study, power demand prediction was proposed by using deep learning for two industrial buildings with different power consumption patterns. At this time, four TMs were constructed and simulated while considering the influences on power consumption. This includes the industrial activities, economic factors, resident living patterns, and weather data for buildings, which represent the primary power consumption.
Although the performances are different for each of the four TMs and deep learning methods, TM1 (power consumption for one year) and TM2 (power consumption for two years), which do not use weather data, show excellent performance when using DNN, RNN, and LSTM. In contrast, the performance was poor if weather data that did not affect power consumption were used (such as TM4). Therefore, it is more critical to select weather data (temperature, humidity, solar radiation, and cloud cover) that only affects power consumption, such as TM3.
In future research, the proposed method in this article will be applied worldwide by overcoming the limitations of the Korean region.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the National Program for Excellence in SW, supervised by the IITP (Institute for Information & Communications Technology Planning & Evaluation) in 2021 (2021-D-01409). I would like to thank Wonhee Cho, Hyungjeong Yang, and Evan Kim, who collaborate this article.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. 2030 National Greenhouse Gas Reduction Target; Ministry of Industry: Sejong City, Korea, 2020. Available online: https://english.motie.go.kr/en/am/introduction/introduction.jsp (accessed on 27 October 2021).
  2. KOSIS: Korean Statistical Information Service. Available online: https://kosis.kr/statHtml/statHtml.do?orgId=771&tblId=DT_77101_G001002 (accessed on 12 March 2021).
  3. Kandananond, K. Forecasting electricity demand in Thailand with an artificial neural network approach. Energies 2011, 4, 1246–1257. [Google Scholar] [CrossRef]
  4. Brown, R.E.; Koomey, J.G. Electricity use in California: Past trends and present usage patterns. Energy Policy 2003, 31, 849–864. [Google Scholar] [CrossRef] [Green Version]
  5. Niu, S.; Jia, Y.; Wang, W.; He, R.; Hu, L.; Liu, Y. Electricity consumption and human development level: A comparative analysis based on panel data for 50 countries. Int. J. Electr. Power Energy Syst. 2013, 53, 338–347. [Google Scholar] [CrossRef]
  6. Sideratos, G.; Ikonomopoulos, A.; Hatziargyriou, N.D. A novel fuzzy-based ensemble model for load forecasting using hybrid deep neural networks. Electr. Power Syst. Res. 2020, 178, 106025. [Google Scholar] [CrossRef]
  7. Maldonado, S.; González, A.; Crone, S. Automatic time series analysis for electric load forecasting via support vector regression. Appl. Soft Comput. J. 2019, 83, 105616. [Google Scholar] [CrossRef]
  8. Yang, A.; Li, W.; Yang, X. Short-term electricity load forecasting based on feature selection and Least Squares Support Vector Machines. Knowl.-Based Syst. 2019, 163, 159–173. [Google Scholar] [CrossRef]
  9. Raza, M.Q.; Khosravi, A. A review on artificial intelligence based load demand forecasting techniques for smart grid and buildings. Renew Sustain Energy Rev. 2015, 50, 1352–1372. [Google Scholar] [CrossRef]
  10. Willis, H.L.; Northcote-Green, J.E.D. Spatial electric load forecasting: A tutorial review. Proc. IEEE 1983, 71, 232–253. [Google Scholar] [CrossRef]
  11. Enea, M. A review of machine learning algorithms used for load forecasting at micro-grid level. In Sinteza 2019—International Scientific Conference on Information Technology and Data Related Research; Singidunum University: Belgrade, Serbia, 2019; pp. 452–458. [Google Scholar]
  12. Brown, R.G. Smoothing Forecasting and Prediction of Discrete Time Series; Prentice-Hall: Englewood Cliffs, NJ, USA, 1963. [Google Scholar]
  13. Ohtsuka, Y.; Oga, T.; Kakamu, K. Forecasting electricity demand in Japan: A Bayesian spatial autoregressive ARMA approach. Comput. Stat. Data Anal. 2010, 54, 2721–2735. [Google Scholar] [CrossRef]
  14. Fix, E.; Hodges, J.L., Jr. Discriminatory Analysis-Nonparametric Discrimination: Consistency Properties; International Statistical Institute: Voorburg, The Netherlands, 1989; Volume 57, pp. 238–247. [Google Scholar]
  15. Yu, Z.; Haghighat, F.; Fung, B.C.M.; Yoshino, H. A decision tree method for building energy demand modeling. Energy Build. 2010, 42, 1637–1646. [Google Scholar] [CrossRef] [Green Version]
  16. Dong, B.; Cao, C.; Lee, S.E. Applying support vector machines to predict building energy consumption in tropical region. Energy Build. 2005, 37, 545–553. [Google Scholar] [CrossRef]
  17. Bonyadi, M.R.; Michalewicz, Z. Particle swarm optimization for single objective continuous space problems: A review. Evol. Comput. 2017, 25, 1–54. [Google Scholar] [CrossRef]
  18. Kalogirou, S.A.; Neocleous, C.C.; Schizas, C.N. Building heating load estimation using artificial neural networks. In Proceedings of the 17th International Conference on Parallel Architectures and Compilation Techniques, San Francisco, CA, USA, 10–14 November 1997. [Google Scholar]
  19. Bagnasco, A.; Fresi, F.; Saviozzi, M.; Silvestro, F.; Vinci, A. Electrical consumption forecasting in hospital facilities: An application case. Energy Build. 2015, 103, 261–270. [Google Scholar] [CrossRef]
  20. Graves, A.; Liwicki, M.; Fernandez, F.; Bertolami, R.; Bunke, H.; Schmidhuber, J. A Novel Connectionist System for Improved Unconstrained Handwriting Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 855–868. [Google Scholar] [CrossRef] [Green Version]
  21. Gers, F.; Schmidhuber, J.; Cummins, F. Learning to Forget: Continual Prediction with LSTM. In Proceedings of the 9th International Conference on Artificial Neural Networks, ICANN’99, Edinburgh, UK, 7–10 September 1999; pp. 850–855. [Google Scholar]
  22. Ian, G.; Yoshua, B.; Aaron, C. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  23. Zhang, J.; Wei, Y.M.; Li, D.; Tan, Z.; Zhou, J. Short term electricity load forecasting using a hybrid model. Energy 2018, 158, 774–778. [Google Scholar] [CrossRef]
  24. Xiong, T.; Bao, Y.; Hu, Z. Interval forecasting of electricity demand: A novel bivariate EMD-based support vector regression modeling framework. Int. J. Electr. Power Energy Syst. 2014, 63, 353–362. [Google Scholar] [CrossRef] [Green Version]
  25. Lahouar, A.; Slama, J.B.H. Day-ahead load forecast using random forest and expert input selection. Energy Convers. Manag. 2015, 103, 1040–1051. [Google Scholar] [CrossRef]
  26. Selakova, A.; Cvijetinovic, D.; Milovic, L.; Mellon, S.; Bekut, D. Hybrid PSO-SVM method for short-term load forecasting during periods with significant temperature variations in city of Burbank. Appl. Soft Comput. 2014, 16, 80–88. [Google Scholar] [CrossRef]
  27. Azim, H.; Davide, A.G.; Farshid, K.; Fabio, B.; Livio, D.S. Hybrid intelligent strategy for multifactor influenced electrical energy consumption forecasting. Energy Sources Part B Econ. Plan. Policy 2020, 14, 341–358. [Google Scholar]
  28. Kavousi-fard, A.; Samet, H.; Marzbani, F. A new hybrid Modified Firefly Algorithm and Support Vector Regression model for accurate Short Term Load Forecasting. Expert Syst. Appl. 2014, 41, 6047–6056. [Google Scholar] [CrossRef]
  29. Kim, T.; Cho, S. Predicting residential energy consumption using CNN-LSTM neural networks. Energy 2019, 182, 72–81. [Google Scholar] [CrossRef]
  30. Zhang, J.; Tan, Z.; Wei, Y. An adaptive hybrid model for short term electricity price forecasting. Appl. Energy 2019, 258, 114087. [Google Scholar] [CrossRef]
  31. Lago, J.; De Ridder, F.; De Schutter, B. Forecasting spot electricity prices: Deep learning approaches and empirical comparison of traditional algorithms. Appl. Energy 2018, 221, 386–405. [Google Scholar] [CrossRef]
  32. Panapakidis, I.P.; Dagoumas, A.S. Day-ahead electricity price forecasting via the application of artificial neural network based models. Appl. Energy 2016, 172, 132–151. [Google Scholar] [CrossRef]
  33. Amjady, N.; Daraeepour, A.; Keynia, F. Day-ahead electricity price forecasting by modified relief algorithm and hybrid neural network. IET Gener. Transm. Distrib. 2010, 4, 432–444. [Google Scholar] [CrossRef] [Green Version]
  34. Conejo, A.J.; Plazas, M.A.; Espinola, R.; Molina, A.B. Day-ahead electricity price forecasting using the wavelet transform and arima models. IEEE Trans. Power Syst. 2005, 20, 1035–1042. [Google Scholar] [CrossRef]
  35. Amjady, N.; Keynia, F. Day-ahead price forecasting of electricity markets by mutual information technique and cascaded neuro-evolutionary algorithm. IEEE Trans. Power Syst. 2008, 24, 306–318. [Google Scholar] [CrossRef]
  36. Luo, S.; Weng, Y. A two-stage supervised learning approach for electricity price forecasting by leveraging different data sources. Appl. Energy 2019, 242, 1497–1512. [Google Scholar] [CrossRef] [Green Version]
  37. Yang, W.; Wang, J.; Niu, T.; Du, P. A hybrid forecasting system based on a dual decomposition strategy and multi-objective optimization for electricity price forecasting. Appl. Energy 2019, 235, 1205–1225. [Google Scholar] [CrossRef]
  38. Amiri, M.; Davande, H.; Sadeghian, A.; Chartier, S. Feedback associative memory based on a new hybrid model of generalized regression and self-feedback neural networks. Neural Netw. 2010, 23, 892–904. [Google Scholar] [CrossRef]
  39. Dabrowski, J.J.; Zhang, Y.; Rahman, A. ForecastNet: A Time-Variant Deep Feed-Forward Neural Network Architecture for ulti-step Ahead Time-Series Forecasting. In Proceedings of the International Conference on Neural Information Processing, Bangkok, Thailand, 18–22 November 2020; pp. 579–591. [Google Scholar]
  40. Takeda, H.; Tamura, Y.; Sato, S. Using the ensemble Kalman filter for electricity load forecasting and analysis. Energy 2016, 104, 184–198. [Google Scholar] [CrossRef]
  41. Chen, Y.; Zhang, S.; Zhang, W.; Peng, J.; Cai, Y. Multifactor spatio-temporal correlation model based on a combination of convolutional neural network and long short-term memory neural network for wind speed forecasting. Energy Convers. Manage. 2019, 185, 783–799. [Google Scholar] [CrossRef]
  42. Ghedamsi, R.; Settou, N.; Gouareh, A.; Khamouli, A.; Saifi, N.; Recioui, B.; Dokkar, B. Modeling and forecasting energy consumption for residential buildings in Algeria using bottom-up approach. Energy Build. 2016, 121, 309–317. [Google Scholar] [CrossRef]
  43. Dong, B.; Li, Z.; Rahman, S.M.; Vega, R. A hybrid model approach for forecasting future residential electricity consumption. Energy Build 2016, 117, 341–351. [Google Scholar] [CrossRef]
  44. Tascikaraoglu, A.; Sanandaji, B.M. Short-term residential electric load forecasting: A compressive spatial-temporal approach. Energy Build. 2016, 111, 380–392. [Google Scholar] [CrossRef]
  45. Leung, M.C.; Norman, C.F.; Lai, L.L.; Chow, T.T. The use of occupancy space electrical power demand in building cooling load prediction. Energy Build. 2012, 55, 151–163. [Google Scholar] [CrossRef]
  46. Yun, K.; Luck, R.; Mago, P.J.; Cho, H. Building hourly thermal load prediction using an indexed ARX model. Energy Build 2012, 54, 225–233. [Google Scholar] [CrossRef]
  47. KEPCO’s iSMART. Available online: http://pccs.kepco.co.kr/ (accessed on 13 March 2021).
  48. YiRe, S.; Sanghoo, Y. Evaluation of weather information for electricity demand forecasting. J. Korean Data Inf. Sci. Soc. 2016, 27, 1601–1607. [Google Scholar]
  49. Youjung, J.; Hyesook, H.; Baekjo, K.; Seongkyoun, K.; Giman, H.; Wookyun, L. The change of the average discomfort index from June to September during the past 10 years. Clim. Chang. Res. 2012, 3, 89–100. [Google Scholar]
  50. Statistics KOREA Government Official Work Conference. Available online: https://www.index.go.kr/unify/idx-info.do?idxCd=4291 (accessed on 4 August 2021).
  51. Bengio, Y.L.; Courville, Y.; Vincent, P. Representation Learning: A Review and New Perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef]
  52. Schmidhuber, J. Deep Learning in Neural Networks: An Overview. 2014. Available online: http://arxiv.org/abs/1404.7828 (accessed on 4 August 2021).
  53. Szegedy, C.; Toshev, A.; Erhan, D. Deep neural networks for object detection. In Proceedings of the Advance Neural Information Processing Systems Conference, Lake Tahoe, NV, USA, 5–10 December 2013. [Google Scholar]
  54. Dupond, S. A thorough review on the current advance of neural network structures. Annu. Rev. Control 2019, 14, 200–230. [Google Scholar]
  55. Zell, A. Simulation Neuronaler Netze [Simulation of Neural Networks], 1st ed.; Addison-Wesley: Bonn, Germany, 1994. (In German) [Google Scholar]
  56. Hopfield, J.J. Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. USA 1982, 79, 2554–2558. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Cruse, H. Neural Networks as Cybernetic Systems, 2nd and revised ed.; Brains, Minds Media: Bielefeld, Germany, 2006. [Google Scholar]
  58. Jaeger, H. Echo state network. Scholarpedia 2007, 2, 2330. [Google Scholar] [CrossRef]
  59. Felix, A.G.; Schraudolph, N.N.; Jürgen, S. Learning Precise Timing with LSTM Recurrent Networks. J. Mach. Learn. Res. 2002, 3, 115–143. [Google Scholar]
  60. Berglung, M.; Raiko, T.; Honkala, M.; Leo, K.; Vetek, A.; Karhunen, J. Bidirectional Recurrent Neural Networks as Generative Models. Adv. Neural Inf. Process. Sys. 2015, 28, 856–864. [Google Scholar]
  61. Alex, S. Deriving the Recurrent Neural Network Definition and RNN Unrolling Using Signal Processing. In Proceedings of the Critiquing and Correcting Trends in Machine Learning Workshop at NeurIPS-2018, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  62. Rainer, W.P.; Jun, T. How Hierarchical Control Self-organizes in Artificial Adaptive Systems. Adapt. Behav. 2005, 13, 211–225. [Google Scholar]
  63. Giles, C.L.; Miller, C.B.; Chen, D.; Chen, H.H.; Sun, G.Z.; Lee, Y.C. Learning and Extracting Finite State Automata with Second-Order Recurrent Neural Networks. Neural Comput. 1992, 4, 393–405. [Google Scholar] [CrossRef]
  64. Hochreiter, S. Untersuchungen zu dynamischen neuronalen Netzen. Diploma Thesis, Institut f. Informatik, Technische University Munich, Munich, Germany, 1991. [Google Scholar]
  65. Hochreiter, S.; Bengio, Y.; Frasconi, P.; Schmidhuber, J. Gradient Flow in Recurrent Nets: The Difficulty of Learning Long-Term Dependencies; Wiley-IEEE Press: Hoboken, NJ, USA, 2001; pp. 237–243. [Google Scholar]
  66. Valueva, M.V.; Nagornov, N.N.; Lyakhov, P.A.; Valuev, G.V.; Chervyakov, N.I. Application of the residue number system to reduce hardware costs of the convolutional neural network implementation. In Mathematics and Computers in Simulation; Elsevier BV: Amsterdam, The Netherlands, 2020; Volume 177, pp. 232–243. [Google Scholar]
  67. Mittelman, R. Time-series modeling with undecimated fully convolutional neural networks. arXiv 2015, arXiv:1508.00317. [Google Scholar]
  68. Chen, Y.; Kang, Y.; Chen, Y.; Wang, Z. Probabilistic forecasting with temporal convolutional neural network. Neurocomputing 2009, 399, 491–501. [Google Scholar] [CrossRef] [Green Version]
  69. Honglak, L.; Grosse, R.; Ranganath, R.; Ng, A.Y. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations. In Proceedings of the 26th Annual International Conference on Machine Learning (ICML), Montreal, QC, Canada, 14–18 June 2009; pp. 609–616. [Google Scholar]
  70. Hinton, G. Deep belief networks. Scholarpedia 2009, 4, 5947. [Google Scholar] [CrossRef]
  71. Convolutional Deep Belief Networks on CIFAR-10. Available online: http://www.cs.toronto.edu/~kriz/conv-cifar10-aug2010.pdf (accessed on 5 August 2021).
  72. Graves, A.; Fernández, S.; Gomez, F. Connectionist temporal classification: Labelling unsegmented sequence data with recurrent neural networks. In Proceedings of the International Conference on Machine Learning (ICML), Orlando, FL, USA, 14–16 December 2006; pp. 369–376. [Google Scholar]
  73. Wierstra, D.; Schmidhuber, J.; Gomez, F.J. Evolino: Hybrid Neuroevolution/Optimal linear search for sequence learning. In Proceedings of the 19th International Joint Conference on Artificial Intelligence (IJCAI), Edinburgh, Scotland, UK, 30 July–5 August 2005; pp. 853–858. [Google Scholar]
  74. Afan, G.S.; Yaya, H.; Edi, A.; Wayan, S. Single layer & multilayer long short-term memory model with intermediate variables for weather forecasting. Procedia Comput. Sci. 2018, 135, 89–98. [Google Scholar]
  75. Cho, K.; Van Merrienboer, B.; Bahdanau, D.; Bengio, Y. On the Properties of Neural Machine Translation: Encoder-Decoder Approaches. In Proceedings of the SSST-8, Eighth Workshop on Syntax, Semantics and Structure in Statistical Translation, Doha, Qatar, 25 October 2014; pp. 103–111. [Google Scholar]
  76. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef] [Green Version]
  77. Zhang, T.; Zhang, E.; Zhang, H.; Shen, F.; Guo, D.; Lin, K. An encoder-decoder neural network for indefinite length digit sequences in natural scene recognition. J. Phys. Conf. Ser. 2019, 1345, 022025. [Google Scholar] [CrossRef]
  78. Seongchan, K.; Seungkyun, H.; Minsu, J.; Sakwang, S. Deep Rain: ConvLSTM network for precipitation prediction using multichannel radar data. arXiv 2017, arXiv:1711.02316. [Google Scholar]
  79. Martín, L.; Zarzalejo, L.F.; Polo, J.; Navarro, A.; Marchante, R.; Cony, M. Prediction of global solar irradiance based on time series analysis: Application to solar thermal power plants energy production planning. Sol. Energy 2010, 84, 1772–1781. [Google Scholar] [CrossRef]
  80. Kleissel, J. Solar Energy Forecasting and Resource Assessment; Academic Press: Cambridge, MA, USA, 2013. [Google Scholar]
  81. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. arXiv 2015, arXiv:1412.6980. [Google Scholar]
  82. Nair, V.; Hinton, G. Rectified linear units improve restricted Boltzmann machines. In Proceedings of the 27th International Conference on International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010. [Google Scholar]
  83. Agresti, A. Categorical Data Analysis; John Wiley and Sons: New York, NY, USA, 1990. [Google Scholar]
  84. Willmott, C.; Matsuura, K. Advantages of the Mean Absolute Error (MAE) over the Root Mean Square Error (RMSE) in assessing average model performance. Clim. Res. 2005, 30, 79–82. [Google Scholar] [CrossRef]
  85. Tensorflow.org. Deep Learning Library Developed by Google. Available online: https://www.tensorflow.org/ (accessed on 5 August 2021).
  86. Keras.io. The Python Deep Learning Library. Available online: https://keras.io/ (accessed on 5 August 2021).
Figure 1. Power consumption by application in Korea in 2018. (a) The power consumption by application in 2018. (b) The growth rate of the power consumption by application in 2018 compared to 2017.
Figure 1. Power consumption by application in Korea in 2018. (a) The power consumption by application in 2018. (b) The growth rate of the power consumption by application in 2018 compared to 2017.
Sustainability 13 12493 g001
Figure 2. Comparison of the power consumption per day (24 h) between Companies B and T for spring (April), summer (July), fall (October), and winter (January). (a) Company B’s power consumption for spring, summer, fall, and winter; (b) Company T’s power consumption for spring, summer, fall, and winter.
Figure 2. Comparison of the power consumption per day (24 h) between Companies B and T for spring (April), summer (July), fall (October), and winter (January). (a) Company B’s power consumption for spring, summer, fall, and winter; (b) Company T’s power consumption for spring, summer, fall, and winter.
Sustainability 13 12493 g002
Figure 3. Correlation analysis of the temperature and humidity by season. (a) Spring; (b) Summer; (c) Fall; (d) Winter.
Figure 3. Correlation analysis of the temperature and humidity by season. (a) Spring; (b) Summer; (c) Fall; (d) Winter.
Sustainability 13 12493 g003
Figure 4. Analysis of the correlation between the temperature and power consumption by season (Company T). (a) Spring; (b) Summer; (c) Fall; (d) Winter.
Figure 4. Analysis of the correlation between the temperature and power consumption by season (Company T). (a) Spring; (b) Summer; (c) Fall; (d) Winter.
Sustainability 13 12493 g004
Figure 5. Analysis of the correlation between the temperature and power consumption by season (Company B). (a) Spring; (b) Summer; (c) Fall; (d) Winter.
Figure 5. Analysis of the correlation between the temperature and power consumption by season (Company B). (a) Spring; (b) Summer; (c) Fall; (d) Winter.
Sustainability 13 12493 g005
Figure 6. Structure of DNN.
Figure 6. Structure of DNN.
Sustainability 13 12493 g006
Figure 7. Structure of the RNN.
Figure 7. Structure of the RNN.
Sustainability 13 12493 g007
Figure 8. Structure of the CNN.
Figure 8. Structure of the CNN.
Sustainability 13 12493 g008
Figure 9. Structure of the LSTM.
Figure 9. Structure of the LSTM.
Sustainability 13 12493 g009
Figure 10. Comparison of the persistence model and the deep learning performance of Companies B and T in April 2018. (a) Comparison of the persistence model and the deep learning performance (Company B); (b) Comparison of the persistence model and the deep learning performance (Company T).
Figure 10. Comparison of the persistence model and the deep learning performance of Companies B and T in April 2018. (a) Comparison of the persistence model and the deep learning performance (Company B); (b) Comparison of the persistence model and the deep learning performance (Company T).
Sustainability 13 12493 g010
Figure 11. Comparison of the power consumption patterns for the daily average by the season for Companies B and T. (a) Company B; (b) Company T.
Figure 11. Comparison of the power consumption patterns for the daily average by the season for Companies B and T. (a) Company B; (b) Company T.
Sustainability 13 12493 g011
Figure 12. Proposed flowchart.
Figure 12. Proposed flowchart.
Sustainability 13 12493 g012
Figure 13. Data partition for training and testing by four TMs. (a) TM1; (b) TM2; (c) TM3; (d) TM4.
Figure 13. Data partition for training and testing by four TMs. (a) TM1; (b) TM2; (c) TM3; (d) TM4.
Sustainability 13 12493 g013
Figure 14. Performance comparison of MAPE (%) by traditional and proposed methods for Company B (April 2018). (a) Comparison of four TMs by deep learning; (b) Deep learning comparison by four TMs.
Figure 14. Performance comparison of MAPE (%) by traditional and proposed methods for Company B (April 2018). (a) Comparison of four TMs by deep learning; (b) Deep learning comparison by four TMs.
Sustainability 13 12493 g014
Figure 15. Performance comparison of MAPE (%) by four test models and deep learning for Company B (January 2018). (a) Deep learning comparison by four TMs. (b) Comparison of four TMs by deep learning.
Figure 15. Performance comparison of MAPE (%) by four test models and deep learning for Company B (January 2018). (a) Deep learning comparison by four TMs. (b) Comparison of four TMs by deep learning.
Sustainability 13 12493 g015
Figure 16. Performance comparison of the four test models when applying deep learning (January 2018). (a) DNN performance comparison for the four TMs. (b) RNN performance comparison for the four TMs. (c) CNN performance comparison for the four TMs. (d) LSTM performance comparison for the four TMs.
Figure 16. Performance comparison of the four test models when applying deep learning (January 2018). (a) DNN performance comparison for the four TMs. (b) RNN performance comparison for the four TMs. (c) CNN performance comparison for the four TMs. (d) LSTM performance comparison for the four TMs.
Sustainability 13 12493 g016
Figure 17. Performance comparison by deep learning for TM2 (January 2018).
Figure 17. Performance comparison by deep learning for TM2 (January 2018).
Sustainability 13 12493 g017
Figure 18. Performance comparison of MAPE (%) by traditional and proposed methods for Company T (April 2018). (a) Comparison of four TMs by deep learning; (b) Deep learning comparison by four TMs.
Figure 18. Performance comparison of MAPE (%) by traditional and proposed methods for Company T (April 2018). (a) Comparison of four TMs by deep learning; (b) Deep learning comparison by four TMs.
Sustainability 13 12493 g018
Figure 19. Performance comparison of MAPE (%) by four test models and deep learning for Company T (January 2018). (a) Deep learning comparison by four TMs. (b) Comparison of four TMs by deep learning.
Figure 19. Performance comparison of MAPE (%) by four test models and deep learning for Company T (January 2018). (a) Deep learning comparison by four TMs. (b) Comparison of four TMs by deep learning.
Sustainability 13 12493 g019
Figure 20. Performance comparison of the four test models by applying deep learning (January 2018). (a) DNN performance comparison for the four TMs. (b) RNN performance comparison for the four TMs. (c) CNN performance comparison for the four TMs.
Figure 20. Performance comparison of the four test models by applying deep learning (January 2018). (a) DNN performance comparison for the four TMs. (b) RNN performance comparison for the four TMs. (c) CNN performance comparison for the four TMs.
Sustainability 13 12493 g020
Figure 21. Performance comparison by deep learning for TM2 (January 2018).
Figure 21. Performance comparison by deep learning for TM2 (January 2018).
Sustainability 13 12493 g021
Table 1. Average, standard deviation, and coefficient of variation for the power consumption of Companies B and T.
Table 1. Average, standard deviation, and coefficient of variation for the power consumption of Companies B and T.
SpringSummerFallWinter
Ave.25.7819.3327.5739.24
BStd.30.9830.1539.1338.42
CV1.161.261.090.97
Ave.124.00114.22122.77119.87
TStd.31.4424.6126.7626.15
CV0.250.210.210.21
Table 2. Training and testing option by deep learning.
Table 2. Training and testing option by deep learning.
ParameterDNNRNNCNNLSTM
Number of layers4363
Number of neurons100100100100
Number of epochs100100100100
Number of convolution filters--64-
Convolution of kernel size--1-
Learning rate0.050.050.050.05
Loss functionMSEMSEMSEMSE
OptimizerADAMADAMADAMADAM
Activation functionRELURELURELURELU
Table 3. Average, standard deviation, and coefficient of variation for the power consumption of Companies B and T.
Table 3. Average, standard deviation, and coefficient of variation for the power consumption of Companies B and T.
Season ClassBT
Metrics PMDNNRNNCNNLSTMPMDNNRNNCNNLSTM
SpringAve.31.7629.4628.6430.0029.17139.41137.23138.76137.13131.76
Std.41.5935.1434.5234.2234.7427.9726.3326.4226.3122.46
CV1.311.191.211.141.190.20.200.190.190.17
RMSE20.6520.7520.0323.0820.6619.3419.2719.0820.6220.03
MAPE43.0756.6157.9198.1253.9510.8110.6810.7611.0710.60
SummerAve.22.1229.4628.6430.0029.17114.5113.94107.71116.58112.58
Std.30.9935.1434.5234.2234.7422.6519.8119.5018.7815.90
CV1.41.191.211.141.190.20.170.180.160.14
RMSE15.0920.7520.0323.0820.6621.4318.9819.1522.8117.88
MAPE36.5956.6157.9198.1253.9513.2712.1812.6515.7611.56
FallAve.29.1329.2028.7428.3632.22127.87126.32129.51124.44124.31
Std.37.7634.3533.8031.3633.5621.3618.6919.1917.1116.43
CV1.31.181.181.111.040.170.150.150.140.13
RMSE18.7017.8217.7219.8615.9318.4717.7317.8219.6517.16
MAPE42.9762.5062.5883.0485.8210.7410.1110.6211.479.62
WinterAve.41.3639.4235.9138.8640.72121.45124.58125.99122.35115.87
Std.37.5833.0529.9831.3934.0225.7623.1723.821.8615.63
CV0.910.840.830.810.840.210.190.190.180.13
RMSE17.2716.2416.9119.0515.9318.7920.1220.3021.8919.26
MAPE22.8825.0022.9336.8128.7712.2613.8414.1214.7111.52
Table 4. Performance comparison of the average RMSE and MAPE values for four seasons by deep learning for Companies B and T.
Table 4. Performance comparison of the average RMSE and MAPE values for four seasons by deep learning for Companies B and T.
ClassCompany BCompany T
Metrics PMDNNRNNCNNLSTMPMDNNRNNCNNLSTM
RMSE17.9317.1117.0419.7416.0519.5119.0219.0821.2418.58
MAPE36.3856.1854.9793.7463.9411.7711.7012.0313.2510.82
Table 5. Proposed four models.
Table 5. Proposed four models.
VariablesModel NameModel Description
UnivariateTM1Power consumption in 2018
MultivariateTM2Power consumption for two years (2017, 2018)
TM3Power consumption in 2018, temperature (°C), humidity (%), solar radiation (MJ/m2), cloud cover
TM4Power consumption for two years (2017, 2018), temperature (°C), humidity (%), solar radiation (MJ/m2), cloud cover, precipitation, wind speed, wind direction, vapor pressure
Table 6. Comparison of the performance of the traditional and the proposed method by deep learning (Company B).
Table 6. Comparison of the performance of the traditional and the proposed method by deep learning (Company B).
Mon.TM ClassTraditional MethodProposed Method
Metrics DNNRNNCNNLSTMDNNRNNCNNLSTM
Jan.TM1RMSE16.2416.9119.0515.9315.2015.8618.1214.92
MAPE25.0022.9336.8128.7724.7522.7036.4428.49
TM2RMSE16.1016.4819.2518.2715.0515.4618.2317.36
MAPE26.2427.5740.2733.3525.9827.3039.8733.01
TM3RMSE16.4825.7018.5825.9215.4324.7817.5325.21
MAPE36.7739.7339.3231.4636.4039.3338.9331.14
TM4RMSE16.7819.7919.4419.4815.7319.1918.4618.47
MAPE39.8239.2544.0436.0039.4238.8643.6035.64
Apr.TM1RMSE20.7520.0323.0920.6618.8118.8722.3019.48
MAPE56.6157.9198.1253.9556.0557.3497.1453.42
TM2RMSE23.0221.0023.5422.5418.8219.7822.6321.33
MAPE63.7461.8192.3256.9563.1161.2091.4056.38
TM3RMSE28.3629.4726.5430.3518.0228.4225.3532.47
MAPE80.1868.1488.1071.4479.3867.4687.2270.72
TM4RMSE21.6521.7025.8626.3118.5220.8123.9325.64
MAPE87.4589.23102.13110.1786.5888.34101.11109.07
Jul.TM1RMSE13.6613.5217.0011.7112.4912.3816.1911.12
MAPE80.6476.48157.0287.2379.8475.72155.4586.36
TM2RMSE13.4714.2816.8813.1712.3013.4415.8012.47
MAPE91.4888.76162.9269.0690.5787.87161.2968.37
TM3RMSE13.1514.3713.4514.8314.5814.0013.4422.63
MAPE100.5963.42131.8760.9299.5862.78130.5560.31
TM4RMSE13.2215.9415.2918.2814.2457.9914.5617.61
MAPE126.81138.84155.12115.23125.54137.45153.57114.08
Oct.TM1RMSE17.8217.7219.8615.9316.8116.7418.7815.18
MAPE62.5062.5883.0485.8261.8861.9682.2184.97
TM2RMSE17.7217.7419.7014.4916.7216.7418.7615.71
MAPE62.2562.1794.1064.7561.6361.5593.1664.10
TM3RMSE16.9917.4818.4017.0316.1616.6317.4116.47
MAPE78.8784.95104.7095.0878.0984.10103.6594.13
TM4RMSE17.4320.6819.6724.5317.0220.3118.8524.30
MAPE86.5391.2594.88116.6085.6690.3393.93115.43
Table 7. Comparison of the performance of the traditional and the proposed method by deep learning (Company T).
Table 7. Comparison of the performance of the traditional and the proposed method by deep learning (Company T).
Mon.TM ClassTraditional MethodProposed Method
Metrics DNNRNNCNNLSTMDNNRNNCNNLSTM
Jan.TM1RMSE20.1220.3021.8919.2619.1219.3020.8318.25
MAPE13.8414.1214.7111.5213.7113.9814.5011.41
TM2RMSE19.0419.2920.4620.6918.0318.2919.4719.70
MAPE11.7412.8013.6513.0911.6212.6713.5112.97
TM3RMSE21.2320.620.2222.7820.2319.3719.3621.82
MAPE15.4612.6213.3713.8515.3112.5013.2913.74
TM4RMSE19.1925.7122.5126.1318.2024.7321.6425.20
MAPE12.3814.8714.6716.8212.2614.7514.5916.70
Apr.TM1RMSE19.2719.0820.6220.0318.2618.0719.5719.02
MAPE10.6810.7611.0910.6010.5710.6510.9410.50
TM2RMSE18.7219.2620.7819.5117.7118.2419.7218.50
MAPE10.2610.4310.9110.3010.1610.3310.7810.19
TM3RMSE19.3119.9718.6419.7718.3018.9717.5818.77
MAPE10.0810.6210.3410.639.9810.5210.2010.52
TM4RMSE19.0019.9619.1124.2017.9918.9618.0523.23
MAPE10.4811.2510.7413.3810.3711.1410.6113.27
Jul.TM1RMSE18.9819.1522.8117.8817.9818.1521.7816.90
MAPE12.1812.6515.7611.5712.0612.5415.5511.45
TM2RMSE18.3818.8521.5417.5617.3717.8420.4916.58
MAPE12.1212.7214.4812.2717.3717.8420.4916.58
TM3RMSE16.9916.1320.2717.1515.9815.1119.2216.14
MAPE11.5311.0813.9311.4211.4110.9613.7411.32
TM4RMSE17.9719.4321.6322.2916.9718.4420.6021.34
MAPE13.0613.3115.7015.3712.9213.2015.5015.25
Oct.TM1RMSE17.7317.8319.6517.1616.7416.8418.6316.17
MAPE10.1110.6211.479.6210.0210.5211.339.53
TM2RMSE17.1017.1718.9518.1016.1116.1917.9217.14
MAPE9.779.8411.4210.969.679.7611.2810.87
TM3RMSE16.4816.9118.9618.1015.4915.9317.9217.13
MAPE9.6410.1811.8411.019.5510.0911.6910.92
TM4RMSE16.4616.7419.2121.9315.4715.7518.2420.99
MAPE9.499.7811.7514.169.409.6911.6414.06
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Son, N. Comparison of the Deep Learning Performance for Short-Term Power Load Forecasting. Sustainability 2021, 13, 12493. https://doi.org/10.3390/su132212493

AMA Style

Son N. Comparison of the Deep Learning Performance for Short-Term Power Load Forecasting. Sustainability. 2021; 13(22):12493. https://doi.org/10.3390/su132212493

Chicago/Turabian Style

Son, Namrye. 2021. "Comparison of the Deep Learning Performance for Short-Term Power Load Forecasting" Sustainability 13, no. 22: 12493. https://doi.org/10.3390/su132212493

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop