Next Article in Journal
Perception of the Quality of Smart City Solutions as a Sense of Residents’ Safety
Previous Article in Journal
Machine Learning-Based Production Prediction Model and Its Application in Duvernay Formation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploratory Data Analysis Based Short-Term Electrical Load Forecasting: A Comprehensive Analysis

1
Department of Electrical and Computer Engineering, COMSATS University Islamabad, Lahore 54000, Pakistan
2
Electrical Engineering Department, University of Management and Technology, Lahore 54000, Pakistan
3
Department of Electrical Power Engineering & Mechatronics, Tallinn University of Technology, 12616 Tallinn, Estonia
*
Author to whom correspondence should be addressed.
Energies 2021, 14(17), 5510; https://doi.org/10.3390/en14175510
Submission received: 2 July 2021 / Revised: 30 August 2021 / Accepted: 31 August 2021 / Published: 3 September 2021

Abstract

:
Power system planning in numerous electric utilities merely relies on the conventional statistical methodologies, such as ARIMA for short-term electrical load forecasting, which is incapable of determining the non-linearities induced by the non-linear seasonal data, which affect the electrical load. This research work presents a comprehensive overview of modern linear and non-linear parametric modeling techniques for short-term electrical load forecasting to ensure stable and reliable power system operations by mitigating non-linearities in electrical load data. Based on the findings of exploratory data analysis, the temporal and climatic factors are identified as the potential input features in these modeling techniques. The real-time electrical load and meteorological data of the city of Lahore in Pakistan are considered to analyze the reliability of different state-of-the-art linear and non-linear parametric methodologies. Based on performance indices, such as Root Mean Square Error (RMSE), Mean Absolute Percentage Error (MAPE) and Mean Absolute Error (MAE), the qualitative and quantitative comparisons have been conferred among these scientific rationales. The experimental results reveal that the ANN–LM with a single hidden layer performs relatively better in terms of performance indices compared to OE, ARX, ARMAX, SVM, ANN–PSO, KNN, ANN–LM with two hidden layers and bootstrap aggregation models.

Graphical Abstract

1. Introduction

Short-Term Electrical Load Forecasting (STLF) is used by planning authorities to forecast energy demand, ranging from one hour to one week ahead [1,2]. Energy management has become one of the most strategic and essential cutting-edge research areas in power system engineering [3,4,5]. Security and protection of electrical power systems can be malfunctioned by irregular power flows and system congestion due to an inaccurate electrical load forecast, which may lead towards imbalanced generation planning. Therefore, electricity generation, transmission and distribution networks governed by electric utilities over the world need an accurate STLF for reliable and economical short-term operations of power systems [6]. STLF also assists in economical fuel scheduling, automatic generation control, economic dispatch, voltage stability of HVAC and HVDC lines, and avoiding highly disruptive blackouts [7]. Therefore, the inspiration behind this research work is to empower the planning authorities of electric utilities with state-of-the-art linear and non-linear parametric methodologies since linear and non-linear parametric methodologies are more suitable to handle the system dynamics and non-linearities [8].
The latest annual report of the International Energy Agency (IEA) suggests that developing counties possess the fastest electricity demand growth rate than developed countries, which have limited annual electricity demand growth rates [9]. The exponential growth is driven by the rising industrial revolution and urbanization [10,11]. Therefore, it is necessary to forecast such a diversified annual electricity demand using robust and high-performance forecasting methods for appropriate power generation planning. It may be observed that the electric utilities of developing countries still rely on conventional statistical methodologies, such as a regression analysis, Auto-Regressive Integrated Moving Average (ARIMA) and bootstrap aggregation. However, these forecasting techniques are incapable of handling system dynamics effectively. Therefore, the motivation behind this paper is to empower electric utilities, specifically those of developing countries, with the potentials and strengths of the advanced AI-based computational intelligence methodologies.
Many research studies have been presented to manifest the role of high-performance Short-Term Electrical Load Forecasting (STLF) methods for reliable power system operations. However, most of the work focuses on the electrical load datasets of developed countries, which does not give any motivation to the power companies of the developing countries. Therefore, to overcome the above-stated research gap, this paper aims to exploit the potentials and strengths of the advanced machine learning and deep learning methodologies, especially the non-linear parametric methodologies over the traditional statistical linear parametric methodologies. For this purpose, this paper also delineates a clear qualitative and quantitative analysis among eight different linear and non-linear parametric STLF methodologies, such as Output Error (OE), Auto-Regressive with Exogenous Inputs (ARX), Auto-Regressive Moving Average with Exogenous Inputs (ARMAX), Support Vector Machine (SVM), Artificial Neural Network–Levenberg–Marquardt algorithm (ANN–LM) with one and two hidden layers, Artificial Neural Network–Particle Swarm Optimization (ANN–PSO), K-Nearest Neighbor (KNN) and bootstrap aggregation. Finally, the paper recommends the ANN–LM method with one hidden layer for the real-time applications of power system operations as an approved STLF model.
The electrical load dataset of LESCO, Pakistan is used in this study that comprises the real-time historical electrical load and meteorological data of Pakistan. The dataset of LESCO is different and highly non-linear compared to any leading country of Europe and America because of frequent power outages, load shedding, failures, blackouts and power shortfall. Therefore, the STLF is a challenging task and to the best of the authors’ knowledge, any state-of-the-art statistical and machine learning algorithms have never been tested on the mentioned dataset. The selection of the input parameters must be appropriate for the accurate forecasting of different statistical and machine learning models. Therefore, the input data selection for forecasting models is obtained through exploratory data analysis, auto-correlation analysis and quantile–quantile analysis in this study. The input and output datasets are then pre-processed to convert into useable multi-variate time-series electrical load data and a novel predictor matrix for the development of the above-discussed pre-trained electrical load forecasting models. After modeling, the evaluation will be performed to estimate the accuracy of the forecasting model using key performance indicators, such as Mean Absolute Percentage Error (MAPE), Root Mean Square Error (RMSE), Mean Absolute Error (MAE), R-square value and standard deviation. Finally, the results show that the ANN–LM method with one hidden layer performs comparatively better than the other above-discussed methodologies. Considering the above-stated actualities, the main contributions of this research paper are as follows:
  • For the first time, the historical electrical load data of Pakistan have been considered for the motivation of electric utilities of developing countries to implement machine learning-based STLF methodologies for reliable power system operations.
  • To select the input features for the first time from the new unprecedented dataset, the exploratory data analysis, graphical observations and statistical techniques, such as auto-correlation analysis, quantile–quantile analysis and box-plot analysis, are implemented.
  • A comprehensive predictor matrix is developed using selected input features for linear and non-linear parametric STLF models. The predictor matrix is not mathematically complex and does not necessarily require data outside the historical patterns.
  • This study conducts a comprehensive qualitative and quantitative comparison between the traditional time-series statistical models, linear and non-linear parametric methodologies using several evaluation metrics, such as MAPE, RMSE, MSE, R-square and standard deviation. Moreover, we conducted a thorough seasonal analysis to evaluate and compare the performance of the recommended algorithms.
The rest of the paper is organized as follows: Section 2 presents a detailed literature review, Section 3 presents a comprehensive exploratory data analysis, Section 4 discusses various linear and non-linear parametric models implemented in our paper, Section 5 provides a detailed analysis and discussion on experimental results, and Section 6 is the concluding section, which concludes the paper with future directions.

2. Literature Review

The electrical load curve depends upon the massive multivariate and non-linear temporal and meteorological data, such as festivals and holidays, humidity and temperature, which make the STLF challenging and demanding [12]. Over the past two decades, extensive research work has been presented on the STLF problem to develop several forecasting methodologies. Several regression models, such as Auto-Regressive Integrated Moving Average (ARIMA) and seasonal-ARIMA (SARIMA) have been proposed in [13,14]. ARIMA and SARIMA utilize the lagged average values of STLF time-series data to convert non-stationary data into stationary ones. This seasonality can be observed by using auto-correlation (ACF) and partial auto-correlation (PACF) analysis [15]. Moreover, a review of several other statistical regression models and their variables and methods have been discussed in [16], such as single linear regression and multiple linear regression. However, the statistical regression algorithms are inadequate in capturing the temporal variations and non-linear electrical load patterns [17,18].
To overcome the above-mentioned shortcomings for STLF, Principal Component Analysis (PCA), sensitivity analysis and step-wise regression can be used to enhance the performance of statistical regression models [16]. PCA uses a dimension reduction algorithm based on the matrix decomposition technique by finding eigenvalues of the multi-variate electrical load data through correlation analysis [19]. However, the selection of the coefficients of the co-variance matrix is tedious in PCA, which may lose the key seasonal impact of temperature influences on the electrical load data [20]. In contrast, Singular Value Decomposition (SVD) is proficient in extracting both seasonal and random components and is more robust than PCA [21]. However, SVD manipulates a complex unitary matrix which is computationally expensive [22].
Machine Learning (ML) models have been introduced later on to reduce the time complexity of SVD in the STLF problem. ML algorithms have enhanced the performance of STLF by showing profound accuracy in dealing with non-linearities of the electrical load data and accurate forecasting of the peaks of electrical load as compared to the statistical regression and dimension reduction models [23,24,25]. ML methods mainly comprise Artificial Neural Networks (ANNs) which can handle the stochastic nature of the weather-sensitive loads during the prediction of electrical load forecasting [26,27,28]. The conventional ANN algorithms experience overfitting problems. For overcoming the overfitting issue, the authors in [29] proposed an improved ANN with a learning method that applies a novel search technique that abstains from overfitting. However, some major drawbacks, such as a vanishing gradient and complex hyper-parameters tuning problems due to high-dimensional input data, restrict the applications of the above-mentioned ANN models [30,31,32,33]. K-Nearest Neighbor (KNN) is also proposed as a ML algorithm, which can deal with non-linear temporal variations in electrical load data using k-most similar instances. In the KNN algorithm, when the k most similar samples are detected, the target is attained by local interpolation of the targets associated with the k detected instances [16,33]. The KNN method has proved to be imperfect while forecasting the consumption of electrical load due to insufficient hyperparameters which also results in overfitting [34]. Moreover, the curse of dimensionality occurs when KNN is subjected to high-dimensional electrical load data. The Support Vector Regression (SVR) model is another prominent ML model for STLF and was implemented as a better alternative in [35]. In [32], a new nu support vector machine (nu–SVR) based on the tuning of a newly introduced hyperparameter called nu was recommended for STLF, which generates less error than the ANN but the selection of reasonable parameters in SVR is still a challenging task [33].
The above-mentioned STLF issues can be resolved with achieving higher prediction accuracy than the old ML models, such as SVR, ANN and KNN, by using hybrid ML models. In hybrid ML models, every constituent algorithm delivers robustness and higher accuracy in STLF [36]. In [33], the weights of weighted K-Nearest Neighbor (W-KNN) are optimized using a Genetic Algorithm (GA). Moreover, the fusion of learning algorithms and clustering in KNN–ANN and KNN–SVR architectures extract prominent features, such as temperature, humidity and weekdays before the self-learning process [37]. However, the complex architectures and an unspecified number of optimal clusters limit the validity of KNN-based hybrid ML models [36]. Although, the forecasting performance of feedforward neural networks (FF–NN) can be increased through an improved learning accuracy using the Elaboration Likelihood Model (ELM) merged with the Levenberg–Marquardt (LM) and the Conditional Mutual Information-based Feature Selection (CMIFS) [38]. In [39], a hybrid BA–ELM was proposed to improve the performance of STLF model. The input weights of an extreme learning machine (ELM) and bias threshold are optimized by a Bat Algorithm (BA) to predict the electric load. However, the critical gradient and Hessian calculations for the STLF model raise the time complexity of the ELM method to a larger extent [40]. Furthermore, the Empirical Mode Decomposition (EMD)-based deep learning method has been contemplated to improve the generalization ability of the STLF model in which multivariate electrical load data are segmented into Intrinsic Mode Functions (IMFs) and then Deep Belief Networks (DBNs) are deployed in the architecture to train the extracted IMFs. Eventually, the electrical load forecasting results are acquired by summing up the weighted outputs from each DBNs by using an ensemble learning algorithm. However, the EMD-based DBN algorithm is complex and loses the original information about the electrical load and electrical load features when selecting IMFs, which eventually squashes the large forecasting errors during the training of deep learning models. Moreover, the segmentation of electrical load data into IMFs using EMD and the neural network training of the captured IMFs separately using DBNs require a high computational time [41]. Additionally, an Adaptive Network-based Fuzzy Inference System (ANFIS) enlarges the learning capacity of NNs by fusing neural architecture with high-level reasoning methodologies of fuzzy logic [42]. However, the intricate rules, a large number of antecedents and mode delays increase the complexity of the learning phase in ANFIS architectures [43].
The second group of hybrid ML algorithms incorporates stochastic global methods or metaheuristic methods, such as Particle Swarm Optimization (PSO) and GA merged with neural networks [44]. PSO and GA can be utilized to find the best input weights of an ANN in step-ahead and multi-step ahead STLF models. A GA-based Non-Linear Auto-Regressive Exogenous input Neural Network (NARX–NN) manages the feedback loop in the backpropagation path of the NN to capture unreliable weather dependencies in the input dataset [12,45]. The metaheuristic model, such as ANN, can easily converge to a local optimum but remains unsuccessfully converged to the global optimum in the diversified feature space [46].

3. Exploratory Data Analysis

Exploratory Data Analysis (EDA) is a statistical approach that examines the presence of multiple hidden features and patterns present in a dataset. Several statistical visual methodologies can be used for this purpose, such as auto-correlations plots, QQ plots, box plots, histograms and many others. EDA is quite important in developing time-series forecasting models since it exhibits and explores the relationship between input and output variables. Therefore, EDA must be performed before formal hypothesis investigation and modeling [47].

3.1. Dataset Description

Real-time accumulated datasets of electrical load consumption and climate data of Lahore, Pakistan, are used in our research work. The electrical load real-time dataset of Lahore Electric Supply Corporation (LESCO) is acquired from the Power Information Technology Company (PITC), which is a faction of the Ministry of Energy, Pakistan, and is responsible for monitoring generation, transmission and distribution through Pakistan [48]. The electrical load data is recorded in real-time at 15 min intervals and consists of 96 samples per day from 2017 to 2019 at an aggregated level (of the whole Lahore region). Moreover, the climatic dataset consists of temperature and humidity data recorded at the 15 min interval from 2017–2019 and is acquired from an online database for Lahore, Pakistan of a Russian company “Raspisaniye Pogodi Ltd.” (St. Petersburg, Russia) having operating stations over the globe [49].

3.2. Input Parameter Description

Several climatic factors, such as temperature and humidity, and temporal factors influence the electrical load demand profile of a certain area [50]. The share of domestic load is quite significant in the entire electrical load been consumed. Therefore, weekends and festival holidays show a more distinguished electrical load-consuming profile than normal working days [51]. Generally, the STLF model not only integrates historical electrical load patterns, but also the meteorological and temporal parameters as well [12]. Figure 1 shows that the EDA performed on the above-mentioned datasets identifies multiple exogenous inputs parameters, which are essential for STLF models.

3.2.1. Auto-Correlation Analysis

Auto-correlation is a statistical representation of the degree of resemblance between a given time-series and a lagged version of itself over consecutive time intervals. Figure 1a represents the auto-correlation of the electrical load consumption for the month of January-2019 over the lagged consecutive time intervals up to 1 week (96 lags/day). Here, one lag represents a lagged version of 15 min; therefore, 96 lags represent the lagged version of one complete day. The electrical load data is normalized between (0, 1). The strong correlation in these plots suggests that the present electrical load consumption value has a strong correlation with previous lagged electrical load consumption values, such as the previous hour electrical load value and the previous day same hour electrical load value.

3.2.2. Quantile–Quantile Plots

QQ plots in Figure 1b,c illustrates that there exists a relationship between the consumed electrical load and climatic data since the values of temperature and humidity seem to be directly proportional to the electrical load consumption profile. Moreover, Figure 1d–e shows the quantile–quantile (QQ) plots of the present electrical load that are plotted against its lagged values for the month of January-2019. The QQ plots show a straight line at 45 degrees, which demonstrates that there exists a strong correlation between present electrical load consumption and its highlighted lagged values and are best suited for developing our STLF model.

3.2.3. Box Plots

Weekend days, festivals and public holidays exhibit a more reduced electrical load consumption profile than normal working days due to the closure of commercial markets, offices and industries. To validate this hypothesis, a box plot analysis of the electrical load consumption profile for January-2019 is performed in Figure 1g. The box plot analysis graphically depicts groups of numerical datasets using their quartiles and validates the presence of the aforementioned temporal factors in the electrical load dataset.
The electrical load curve for the month of January-2019 is plotted in Figure 1f. Apart from the EDA performed, the plotted electrical load curve highlights the presence of seasonal patterns with the electrical load consumption profile throughout the day and the entire week. Therefore, the hour number of the day and the day number of the week are also considered in our developed STLF predictor matrix.
As per the suggestions of EDA, some potential input parameters for STLF models are highlighted and are presented in Table 1.

4. Methodology

The identification of various dynamic models can be explained using a number of models with each one relating various stochastic processes and selection measures. The forecasting methodologies used in this research project are explained below.

4.1. Auto-Regressive with Exogenous Inputs (ARX)

The dynamic processes driven by input in an uncertain environment can simply be described using the ARX models. ARX is a modified form of the Auto-Regressive (AR) model with exogenous inputs since AR models do not include inputs. ARX incorporates the stimulus signal and captures the stochastic nature of the environment as the system dynamics. The output of a process can be estimated using the sum of the previous input and output regression values as follows [52]:
y t + a 1 y t 1 + + a n a y t n a   =   b 1 u t n k   + +   b n b u t n b + n k + 1   +   e t .
The parameters n a and n b are the orders of the ARX model, and n k is the delay. A description of Equation (1) is given below:
t : the output of process at time t ;
n a : number of poles;
n b : number of zeros;
n k : number of input samples that occur before the input affects the output;
y t 1 y t n a : previous outputs on which the current output depends;
u t n k u t n k n b + 1 : previous and delayed inputs on which the current output depends;
e t : white-noise disturbance value.
The difference equation can be written in a more compact way as the following [52]:
A q y t ) = B q u ( t n k   +   e t
where q is the delay operator. The developed load forecasting model is a MISO system having multiple input parameters and the single output of electrical load, so the differential equation of the ARX model used in this paper would become [52]:
A q y t ) = B 1 q u 1 ( t n k 1   +   B 2 q u 2 t n k 2   + +   B n u q u n u t n k n u   +   e t .
The estimation error can be minimized by selecting the high order model than the actual order of the model. However, increasing the model order impacts the stability of the system. The identification method usually used in ARX is the least square method since it is the most competent polynomial estimation technique [52]. The interpretation of the ARX models can be represented in Figure 2a [52]. Here u is the input and y is the output.

4.2. Auto-Regressive Moving Average with Exogenous Inputs (ARMAX)

The equation error of complex stochastic environment having dynamic process may be defined using the ARMAX model, which is a time-series model and is developed by means of both processes, i.e., Auto-Regressive (AR) and Moving-Average (MA), having exogenous inputs incorporated. ARMAX consists of both lagged variables, i.e., dependent and independent variables, and incorporates stochastic dynamics present in its system’s structure, such as noise existence in linear dynamic systems. The model parameters are usually estimated using the Recursive Extended Least Square (RELS) method since RELS minimizes the sum of the squares of the prediction errors of a linear ARMAX model. The following Equation (4) gives a mathematical representation of the ARMAX model [53]:
y t   +   a 1 y t 1   + +   a n a y t n a   = b 1 u t n k   + +   b n b u t n k n b + 1   +   c 1 e t 1   + + c n c e t n c   +   e t .
The parameters n a and n b are the orders of the ARX model, and n k is the delay. A description of Equation (4) is given below:
y t : output at time t ;
n a : number of poles;
n b : number of zeroes plus 1;
n c : number of C coefficients;
n k : number of input samples that occur before the input affects the output;
y t 1 y t n a : previous outputs on which the current output depends;
u t n k u t n k n b + 1 : previous and delayed inputs on which the current output depends;
e t 1 e t n c : white-noise disturbance value.
The difference equation can be written in a more compact way as [53]:
A q y t ) = B q u ( t n k   +   C q e t ,
where q is the delay operator. In our research work, we are employing the MISO forecasting model, so the differential equation of the ARMAX model becomes [53]:
A q y t ) = B 1 q u 1 ( t n k 1   +   B 2 q u 2 t n k 2   + + B n u q u n u t n k n u   +   C q e t .
The ARMAX model can be represented by using Figure 2b. Here, u is the inputs of the system, e is the system’s disturbance and y is the output [53].

4.3. Output Error Model (OE)

OE is a special configuration of polynomial models and represents a conventional transfer function that relates outputs to the observed inputs with the addition of white-noise as an additive output disturbance. Therefore, disturbances need no modeling while using OE models, which gives it an advantage over other models. OE’s methodology has usually a quadratic parameters-based cost function, which solves the problem of asymptotic convergence. These parameters can be obtained by the minimization of the loss function which can be performed using various non-linear optimization techniques. The Gauss–Newton method is usually preferred. OE’s configuration can be presented as [53]:
ε OE = y y ^ = y B ^ A ^ u ,
or with delay, the above Equation (7) becomes [53]:
y t   = B q F q u t n k   +   e t
where y t is the measured output, y ^ is the transfer function models output, u is the excitation or input to the system. B A is known as the true system and B ^ / A ^ is the estimated or predicted model of the system. The output disturbance or error is e t . The OE model is created using the specified model delays, orders, and estimation options, where the order n b   n f   n k defines the number of parameters in each component of the estimated polynomial. Figure 2c represents the structure of OE based on the Equations (7) and (8) [53].

4.4. Support Vector Machine (SVM)

One of the finest models of supervised machine learning is the Support Vector Machine (SVM), which develops a set of hyperplanes in a high dimensional space for classification, outlier detection, regression and pattern detection [54]. An SVM hyperplane, based on the knowledge of the data patterns known as features, classifies the observations of one class from another. The linear SVM classifier is a type of SVM classifier that predicts a given input into two possible outputs/classes, thus making it a non-probabilistic binary linear classifier. The hypothesis of a linear SVM classifier can be described in Equation (9) [54]. The most expected label for the new input data can be predicted using that hyperplane. These features are a type of interpolations extracted from the data patterns during feature selection. The prediction is made using either linear or non-linear hyperplane as per the application.
h x i = + 1         i f   w . x + b 0 1         i f   w . x + b   < 0   .
Support vectors are the values that are close to the classification margin. The goal of SVM is to maximize the margin between the support vectors and hyperplane. If we are having s hyperplanes, with each having B i value, then we will choose the hyperplane having the largest B i value [54].
H = max i = 1 s   { h i B i } ,
In SVM, the margin size is inversely proportional to the generalization error; therefore, an accurate separation is accomplished by lowering the classifier’s generalization error using the hyperplane that has the largest widened distance to the closest training data point. A hyperplane that maximizes the margin is called an optimal hyperplane. The margins of SVM’s could be made either hard or soft, depending upon the data and the application has been used to predict its future output. The model overfitting can be avoided with the appropriate selection of hyperplane margins, so a trade-off between training error and hyperplane complexity occurs during margin selection [54].

4.5. K-Nearest Neighbour (KNN)

K-nearest neighbor algorithm is a non-parametric model that has no presumptions about the decision boundary in parametric form and can perform regression and classification on both binary and multiclass datasets. KNN searches for exemplars in the training observation dataset which most suitably corresponds to the new data point. Based on the weighted means of the nearest neighbors, the predicted outcome is allocated to the newly entered input datapoint of the most alike exemplar [55]. The value of K should be chosen wisely since it significantly affects the performance of the KNN algorithm. The optimal value of K can be chosen using several heuristic techniques, such as the hyperparameter optimization method. The Euclidean distance is used to define the closeness among the datasets and can be computed using the following Equation (11). Here A and B are the two points in the Euclidean n-space between which the distance is to be measured [55].
s i m i l a r i t y A . B = A . B A B ,
Minkowski distance is another distance metric used by cubic KNN and can be computed by the following Equation (12). Here, r parameter of the Minkowski distance metric represents the order of the norm, and A and B are the two points in space between which the distance is to be measured [55].
M i n k o w s k i A , B = i = 1 m A i B i r 1 / r

4.6. Bootstrap Aggregation (Bagged Trees)

Bootstrap aggregation is an ensemble machine learning algorithm, which bags ensemble trees for either regression or classification problems and is capable of reducing the variance of the forecasting model without increasing the bias. The trees found in the ensemble are grown on the bootstrap replica of the input data point. Several subsets of data are created from the provided training samples, which are selected randomly with substitution. Each subset group of a dataset is then used for the training of decision trees. Resultantly, an ensemble of several models is created. Since the decision tree is solely a weak learner and is quite responsive and sensitive to the training patterns, therefore a single decision tree results overfit the specific patterns in the dataset which results in model overfitting. Bagged decision trees upgrade the performance of decision trees by using many trees that are assembled in parallel structures for learning, and the prediction against new observation is obtained from the average prediction outcomes of all the decision trees [56].

4.7. Artificial Neural Network (ANN)

The ANN methodology in a data processing network is invented based on the human nervous system as depicted in Figure 3 [57]. The human brain is capable of learning instructions through experiences. These experiences and instructions are the training data provided to ANN to learn things in different scenarios. The ANN network can be tuned by learning rules and instructions, network structures, interconnection techniques and the output transfer functions. The neuron weights are optimized during the learning process to obtain the desired accuracy. ANN’s methodology has self-learning abilities with modified standards competitive to that of statistical methodologies and solves those problems which seem to be difficult for statistical analysis or humans. Neural networks are modern-day state-of-the-art methodologies that not only learn non-linear patterns in the data provided but also handle dynamic systems and extract out hidden relationships between inputs and outputs. ANNs are believed to be powerful tools for time-series modeling and prediction.
The neurons in ANNs are actually the replica of cells of the human nervous system’s neurons interconnected by nodes to form a web-like network, where each neuron contributes its part by processing information. Based on the internal weighting system, this information is continuously updated upon receiving new information from observed values of the process and is used to predict the future output. Learning rules are established in an ANN using backpropagation methodologies. The optimization methods used in our research paper are the Particle Swarm Optimization (PSO) algorithm and the Levenberg–Marquardt (LM) algorithm.

4.7.1. Particle Swarm Optimization Algorithm (PSO)

Particle Swarm Optimization Algorithm (PSO) is inspired by the activities and social behavior of fishes and birds, such as schooling, regrouping, flocking and a sudden direction-changing action by varying their velocity. A population (called a swarm) of candidate solutions, which consists of dubbed particles, is formed where these particles are moved around in a free search area having very large spaces. The velocity and position of particle movements are calculated using Equation (13) [57]. Each solution in the PSO is known as a single particle and has four features: (1) the current position x i t , (2) the best historical position p i t ˜ , which is analyzed using the objective function, (3) the best historical position p i t ^ , which is located in all particles, and (4) the current velocity v i t . The following Equation (13) describes the changes in position and velocity.
v i t + 1 = w v i t +   c 1 r 1 ( ) p i t ˜ x i t + c 2 r 2 ( ) p i t ^ x i t x i t + 1 = x i t + v i t + 1
Here, c 1 and c 2 are known as acceleration factors, r 1 ( ) and r 2 ( ) as uniformly generated random values ranging from 0 ,   1 , and w as the weight of inertia. The algorithms move towards the best-known positions in this search space and try to search the local best position of each particle and are directed by their own best-known positions found in an individual iteration as the algorithms seek towards an optimized solution. Therefore, the best positions of these particles keep on updating at each iteration. The process is recursed until a reasonable solution is found. The PSO algorithm updates the internal weights of ANN using this iterative procedure [57].

4.7.2. Levenberg–Marquardt (LM) Algorithm

Levenberg–Marquardt algorithm is a hybrid method that incorporates both the steepest descent and Gauss–Newton approaches and is quite efficient for parameter extraction. An application of a system having a large number of parameters and non-linear least square problems in the application of a least squares curve fitting can be effectively solved using the LM method. LM uses this approximation to the Hessian matrix to compute the performance of the overall ANN network as represented in Equation (14) [58].
x k + 1 = x k   J T J + μ I 1 J T e
Here, J is a Jacobian matrix that is computed through the standard backpropagation method, and it compromises the first derivatives of the ANN network errors with respect to biases and weights of outputs and inputs, and the factor μ is a scalar quantity. During this commutative iterative process, the objective is to move towards the Gauss–Newton method swiftly since the Gauss–Newton method is pretty fast and precise near the minima of error. Thus, after each successful iteration, the µ is decreased so the process may become faster near error minima. The LM algorithm updates the internal weights of the ANN network using this iterative procedure [58].

5. Experimental Results and Discussions

To evaluate the effectiveness of the non-linear parametric models, specifically NN–LM with a single hidden layer, over the traditional statistical linear parametric models, such as ARX, ARMAX and OE, this section presents a comprehensive overview of quantitative and qualitative analysis based on the evaluation metrics as discussed below.

5.1. Selection of Evaluation Metrics

The evaluation of forecasting accuracy is quite important to obtain the measure of the model’s performance. The performance of the forecasting/prediction algorithms can be determined by using various error functions. The following key performance metrics are used to evaluate the model’s performance: Mean Absolute Percentage Error (MAPE), Mean Absolute Error (MAE), Root Mean Square Error (RMSE), R-square value and standard deviation, and the relevant formulas are given in Equations (15)–(19) [57]. Here x i represents the actual output value, y i represents the predicted/forecasted output value and n represents the number of data points.
M A P E = 1 2 i = 1 n x i y i x i
M A E = Σ i = 1 n y i x i n
RMSE = Σ i = 1 N x i y i 2 n
R 2 = 1 S u m   S q u a r e   R e g r e s s i o n   E r r o r S u m   S q u a r e d   T o t a l   E r r o r
S t a n d a r d   D e v i a t i o n   S t d .   D e v . = σ = Σ x i μ 2 n

5.2. Experimental Background

First, the potential input parameters have been identified using a rigorous exploratory data analysis in Section 2. The temporal and seasonal potential input determined for our STLF model is shown in Table 1. The present electrical load is considered as the output for the training and testing of all the linear and non-linear algorithms.
The real-time electrical load data were recorded at 15 min intervals from 2017 to 2019 at an aggregated level (of the whole LESCO region). First, the data is segregated month-wise because the statistical and correlation analysis of every month was different from other months due to seasonality patterns in the yearly load curve. After the segregation, the entire data has been divided into two data sets. The training dataset consists of 2 years’ quarter-hourly electrical load data from 2017 to 2018. Whereas the remaining dataset, which consists of 1-year quarter-hourly data of 2019, is used for testing purposes. The training and testing of every month have been performed on an individual basis.

5.3. Experimental Analysis

First, the linear parametric models named ARX, ARMAX and OE have been deployed in MATLAB for the STLF problem. The predicted electrical load for every month of the year 2019 of the linear models is evaluated and compared using performance evaluation metrics. For instance, the MAPE of January for OE, ARX and ARMAX are 8.42, 6.82 and 5.12, respectively. Moreover, the other performance metrics, such as RMSE, MAE, R-square and standard deviation are also recapitulated in Table 2 and Table 3. The values reveal that ARMAX performs comparatively better than OE and ARX models for January. ARMAX mainly consists of two hyperparameters. The first parameter relates to the order of the Auto-Regressive (AR) process, which acquires the pattern of lag values of electrical load data. The second prominent parameter is the Moving Average (MA), which is used for obtaining the trends between potential input features in electrical load data. ARMAX model is suitable when there is a strong auto-correlation between the present electrical load and the lagged values of an electrical load. January has a strong auto-correlation between the present electrical load and the lagged values of an electrical load. Therefore, ARMAX would be able to capture the lagged values of electrical load and the local trends between the nominated input features comparatively better than OE and ARX. ARMAX also detected the peak load and sudden change of electrical load accurately in Figure 4b than the rest of the linear parametric models. Conversely, the ARMAX does not map the predicted and actual load accurately for the month of June as shown in Figure 5b. Similarly, the bar chart represented in Figure 6a between aforesaid statistical models suggests that for some months of the year, such as April, ARX performs better than the ARMAX. The MAPE for ARX and ARMAX for April is 2.26 and 3.05, respectively. ARX is effective when there is a weak auto-correlation between the present electrical load and the lagged values of an electrical load. The month of April has weak auto-correlation between the present electrical load and the lagged values of an electrical load as illustrated in Figure 7. Thus, ARX shows less MAPE as compared to ARMAX. Further qualitative analysis between the statistical parametric models concludes that other forecasting errors, such as RMSE, MAE, R-square and standard deviation of the statistical models as indicated in Table 2 and Table 3, are still at a higher level. The said forecasting errors suggest that there should be a further advancement in the STLF models so that the error functions can be reduced for a more superior prediction of electrical load. Moreover, Figure 4a–c and Figure 5a–c also shows that the above-mentioned linear models are incapable of forecasting electrical load peaks and abrupt changes in the load consumption, which is quite important for managing hot and cold reserves in power systems.
Next, the five non-linear parametric STLF models, such as KNN, Bagged Trees, SVM, NN–PSO and ANN–LM with one and two hidden layers have been employed for the same STLF problem. The ANN–LM models were trained on different learning rates and the number of neurons to obtain the best possible forecasting results. The quantitative analysis of the above-mentioned non-linear models is also performed for every month of the year to validate the performance of non-linear models in the presence of highly diversified non-linear factors, such as temperature and humidity. The forecasting errors are also mentioned in Table 2 and Table 3. Considering the month of January of the year 2019, the KNN achieve the MAPE values of 7.56. While the Bagged Trees, SVM, NN–PSO, ANN–LM with one dense layer and ANN–LM with two dense layers achieve the MAPE values of 4.46, 3.71, 3.95, 3.85 and 2.47, respectively. Bagged Tree significantly improves the error compared to the KNN for STLF. The KNN has only one hyperparameter denoted by “K”, which determines the number of neighbors from which similar instances have been deduced. Due to an insufficient number of parameters and lack of capability of adding regularization effect in KNN, the KNN experiences the overfitting problem which limits the forecasting performance. Therefore, the KNN may not be able to match the peaks of the predicted electrical load with the peaks of the actual electrical load as shown in Figure 4d. Therefore, KNN degrades the accuracy of the STLF model while Bagged Tree ensures better forecasting results by matching the peaks of the predicted and actual electrical load as illustrated in Figure 4e and Figure 5e. The SVM also yields fewer errors in the prediction of one-step-ahead electrical load as compared to the KNN and the Bagged Trees algorithms if the tunning hypermeters are wisely selected. Therefore, the SVM delivers the closest match between the peaks of the forecasts and the actual electrical load as depicted in Figure 4f and Figure 5f. Due to the limitations of the hyperparameters running in SVM, SVM cannot be improved further. NN–PSO, which is a metaheuristic-based NN model, also fails to remove the shortcomings of the above-discussed non-linear parametric models because they are unable to converge globally. Eventually, NN–PSO renders slightly higher MAPE than SVM for January, and the peak detection of NN–PSO is also more inadequate than SVM. At last, a sufficient improvement in the minimization of the forecasting errors has been observed when the ANN–LM model with one and two dense layers have been considered for the STLF problem. The MAPE of ANN–LM, which implements one dense layer, is 2.74, which is relatively less than the remaining above-mentioned non-linear ML models. The higher MAPE values of two hidden layers ANN–LM, as compared to MAPE values of one hidden layer ANN–LM, suggests that the former model of ANN–LM delivers higher forecasting errors due to increase in the complexity of the model. Moreover, ANN–LM with two dense layers may lose generalization capability due to how the impact of the overfitting issue increases with the increase in hidden layers [59]. However, the single hidden layer ANN–LM model is simple and does not experience overfitting. The difference in the MAPE values of the above-mentioned linear and non-linear parametric methodologies also shows the validity of a single hidden layer ANN–LM over other discussed STLF methodologies as represented in the bar chart, Figure 6b. The bar chart also suggests that a single hidden layer ANN–LM also produces fewer MAPE values for the rest of the year, which reveals the stability of ANN–LM when implemented with one hidden layer over the remaining aforesaid non-linear ML models. Therefore, ANN–LM with one hidden layer can extract both linear and non-linear relationships between the electrical load and the seasonal trends. A one hidden layer ANN–LM also generates the strongest match between the forecasted and the actual electric load, especially during the peak load and the valley load as shown in Figure 4h and Figure 5h.
We extend our analysis by comparing the forecasting performance of different linear and non-linear models in order to authenticate which algorithm can effectively perceive the real electrical load pattern in both winter and summer seasons. For this purpose, two months have been selected. The month of January generally represents the winter season. Whereas June relates to summer. In winter, the consumption of electrical load decreases in Pakistan due to the low temperature. Therefore, the efficient STLF model should find the closest match between the actual and forecasted electrical load for the months of January and June, which will ensure that the algorithm remains reliable and stable even the season undergoes different variations. As far as the best results of ANN–LM architecture are concerned, ANN–LM with a single hidden layer achieves MAPE of 2.47 and 1.81 for January and June as mentioned in Table 2 and Table 3, which is the lowest MAPE value compared to the other under-considered linear and non-linear parametric models. The illustrations in Figure 4 and Figure 5 explain that the ANN–LM with a single hidden layer forecasts the electrical load pattern accurately in both seasons as compared to the above-mentioned linear and non-linear parametric models. Similarly, the detection of peak load and the valley load is relatively finer in a single hidden layer ANN–LM than other parametric models. A single hidden layer ANN–LM enhances the efficiency of the STLF model due to the following reasons:
  • A single layer ANN–LM is different from conventional ANN which uses ANN with regularization parameters. The regularization parameter enables ANN to overcome the overfitting problem.
  • A single layer NN–LM improves the training accuracy by using a more advanced optimization algorithm termed as Levenberg–Marquardt (LM).
Considering the above-discussion and quantitative and qualitative analysis based on different forecasting errors, one hidden layer ANN–LM delivers fewer forecasting errors among all the linear and non-linear parametric methodologies. Similarly, ANN–LM with a one hidden layer forecasts the consumed electrical load accurately by detecting the non-linear temporal and seasonal variations accurately. This paper proposes the use of the ANN–LM algorithm with one hidden layer for the STLF problem, especially for those electric utilities, rather than conventional statistical methodologies.

6. Conclusions

This research paper discusses the effectiveness of eight different state-of-the-art linear and non-linear parametric methodologies, i.e., OE, ARX, ARMAX, KNN, Bagged Trees, SVM, NN–PSO, a single hidden layer ANN–LM and a two hidden layer ANN–LM, on real-time electrical load data for STLF problems. The temporal and climatic factors are also embedded as input parameters in STLF models after careful statistical correlation analysis. A quantitative and qualitative comparison based on different evaluation metrics, such as MAPE, RMSE, MAE, R-square and standard deviation, has been accomplished among the aforesaid linear and non-linear parametric modeling techniques. By assessing the MAPE values of the above-stated linear and non-linear methodologies, ANN–LM with one hidden layer generates less error for all the months of the year as compared to the errors of other foregoing linear and non-linear methodologies.
The efficacy of ANN–LM, when deployed with one hidden layer in the STLF problem, is also observed in the case of seasonal variations. Experimental results explain that the ANN–LM with one hidden layer detects the peaks of the electrical load in both the winter and summer seasons relatively better than the rest of the other linear and non-linear parametric methodologies used in this research. Therefore, a single hidden layer ANN–LM forecasts the electrical load accurately.
Finally, the explanation due to which a single hidden layer ANN–LM shows an improvement for predicting the electrical load is discussed in brief to motivate those electrical utilities which merely depend on conventional statistical methodologies. At last, the ANN–LM with one dense layer is recommended as a reliable STLF model for accurate electrical load forecasting.
In the expansion of the presented research work, we will look over the following future directions: (a) multi-step ahead STLF using neural networks to predict multiple hours or even days ahead forecast and (b) deep-learning-based STLF to achieve a much accurate prediction of load peaks.

Author Contributions

Conceptualization, U.J., K.I. and M.J.; methodology, U.J., K.I. and E.A.A.; validation, U.J., M.J. and E.A.A.; formal analysis, N.S., L.K. and O.H.; investigation, E.A.A., N.S. and M.J.; resources, E.A.A., L.K. and O.H.; data curation, U.J. and K.I.; writing—original draft preparation, U.J., K.I. and M.J.; writing—review and editing, M.J., N.S.; visualization, U.J., K.I. and M.J.; supervision, M.J., L.K. and O.H.; project administration L.K., O.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Estonian Research Council grants PSG142, PRG675, Estonian Centre of Excellence in Zero Energy and Resource Efficient Smart Buildings and Districts ZEBE, grant 2014-2020.4.01.15-0016 funded by European Regional Development Fund.

Institutional Review Board Statement

Not Applicable.

Data Availability Statement

Meteorological dataset set is open access for all users and easily available on www.rp5.ru/Weather_in_the_world (Accessed on 1 April 2020).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dudek, G. Short-term load forecasting using neural networks with pattern similarity-based error weights. Energies 2021, 14, 2334. [Google Scholar] [CrossRef]
  2. Shah, I.; Iftikhar, H.; Ali, S.; Wang, D. Short-term electricity demand forecasting using components estimation technique. Energies 2019, 12, 2532. [Google Scholar] [CrossRef] [Green Version]
  3. Kiprijanovska, I.; Stankoski, S.; Ilievski, I.; Jovanovski, S.; Gams, M.; Gjoreski, H. HousEEC: Day-ahead household electrical energy consumption forecasting using deep learning. Energies 2020, 13, 2672. [Google Scholar] [CrossRef]
  4. Jawad, M.; Qureshi, M.B.; Khan, M.U.S.; Ali, S.M.; Mehmood, A.; Khan, B.; Wang, X.; Khan, S.U. A robust optimization technique for energy cost minimization of cloud data centers. IEEE Trans. Cloud Comput. 2021, 9, 447–460. [Google Scholar] [CrossRef]
  5. Hussain, I.; Ali, S.M.; Khan, B.; Ullah, Z.; Mehmood, C.A.; Jawad, M.; Farid, U.; Haider, A. Stochastic wind energy management model within smart grid framework: A joint bi-directional Service Level Agreement (SLA) between smart grid and wind energy district prosumers. Renew. Energy 2019, 134, 1017–1033. [Google Scholar] [CrossRef]
  6. Khan, K.S.; Ali, S.M.; Ullah, Z.; Sami, I.; Khan, B.; Mehmood, C.A. Statistical energy information and analysis of Pakistan economic corridor based on strengths, availabilities, and future roadmap. IEEE Access 2020, 8, 169701–169739. [Google Scholar] [CrossRef]
  7. Wood, A.J.; Wollenberg, B.F.; Sheblé, G.B. Power Generation, Operation, and Control; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2014; Chapter 12; pp. 566–569. [Google Scholar]
  8. Jawad, M.; Rafique, A.; Khosa, I.; Ghous, I.; Akhtar, J.; Ali, S.M. Improving disturbance storm time index prediction using linear and nonlinear parametric models: A comprehensive analysis. IEEE Trans. Plasma Sci. 2019, 47, 1429–1444. [Google Scholar] [CrossRef]
  9. IEA South Asia Energy Outlook 2019. Available online: https://www.iea.org/reports/southeast-asia-energy-outlook-2019 (accessed on 25 August 2021).
  10. National Transmission and Despatch Company Limited. Power System Statistics 45th Edition. Available online: https://ntdc.gov.pk/ntdc/public/uploads/services/planning/power%20system%20statistics/Power%20System%20Statistics%2045th%20Edition.pdf (accessed on 25 August 2021).
  11. IEA India Energy Energy Outlook 2021. Available online: https://www.iea.org/reports/india-energy-outlook-2021 (accessed on 25 August 2021).
  12. Jawad, M.; Ali, S.M.; Khan, B.; Mehmood, C.A.; Farid, U.; Ullah, Z.; Usman, S.; Fayyaz, A.; Jadoon, J.; Tareen, N.; et al. Genetic algorithm-based non-linear auto-regressive with exogenous inputs neural network short-term and medium-term uncertainty modelling and prediction for electrical load and wind speed. J. Eng. 2018, 2018, 721–729. [Google Scholar] [CrossRef]
  13. Edigera, V.Ş.; Akarb, S. ARIMA forecasting of primary energy demand by fuel in Turkey. Energy Policy 2007, 35, 1701–1708. [Google Scholar] [CrossRef]
  14. Musbah, H.; El-Hawary, M. SARIMA model forecasting of short-term electrical load data augmented by fast fourier transform seasonality detection. In Proceedings of the IEEE Canadian Conference of Electrical and Computer Engineering (CCECE), Edmonton, AB, Canada, 5–8 May 2019; pp. 1–4. [Google Scholar]
  15. Dodamani, S.; Shetty, V.; Magadum, R. Short term load forecast based on time series analysis: A case study. In Proceedings of the IEEE International Conference on Technological Advancements in Power and Energy (TAP Energy), Kollam, India, 24–26 June 2015; pp. 299–303. [Google Scholar]
  16. Yildiz, B.; Bilbao, J.; Sproul, A. A review and analysis of regression and machine learning models on commercial building electricity load forecasting. Renew. Sustain. Energy Rev. 2017, 73, 1104–1122. [Google Scholar] [CrossRef]
  17. Deb, C.; Zhang, F.; Yang, J.; Lee, S.E.; Shah, K.W. A review on time series forecasting techniques for building energy consumption. Renew. Sustain. Energy Rev. 2017, 74, 902–924. [Google Scholar] [CrossRef]
  18. Tayaba, U.B.; Zia, A.; Yanga, F.; Lu, J.; Kashif, M. Short-term load forecasting for microgrid energy management system using hybrid HHO-FNN model with best-basis stationary wavelet packet transform. Energy 2020, 203, 117857. [Google Scholar] [CrossRef]
  19. Jun-long, F.; Yu, X.; Yu, F.; Yang, X.; Guo-liang, L. Rural power system load forecast based on principal component analysis. J. Northeast. 2015, 22, 67–72. [Google Scholar] [CrossRef]
  20. Bianchi, F.M.; Santis, E.D.; Rizzi, A.; Sadeghian, A. Short-term electric load forecasting using echo state networks and PCA decomposition. IEEE Access 2015, 3, 1931–1943. [Google Scholar] [CrossRef]
  21. Abu-Shikhah, N.; Elkarmi, F. Medium-term electric load forecasting using singular value decomposition. Energy 2011, 36, 4259–4271. [Google Scholar] [CrossRef] [Green Version]
  22. Arora, S.; Taylor, J.W. Short-term forecasting of anomalous load using rule-based triple seasonal methods. IEEE Trans. Power Syst. 2013, 28, 3235–3242. [Google Scholar] [CrossRef] [Green Version]
  23. Shabbir, N.; Kutt, L.; Jawad, M.; Iqbal, M.N.; Ghahfarokhi, P.S. Forecasting of energy consumption and production using recurrent neural networks. Adv. Electr. Electron. Eng. 2020, 18, 190–197. [Google Scholar] [CrossRef]
  24. Shabbir, N.; Kütt, L.; Jawad, M.; Amadiahanger, R.; Iqbal, M.N.; Rosin, A. Wind energy forecasting using recurrent neural networks. In Proceedings of the Big Data, Knowledge and Control Systems Engineering (BdKCSE), Sofia, Bulgaria, 21–22 November 2019; pp. 1–5. [Google Scholar]
  25. Ahmed, W.; Ansari, H.; Khan, B.; Ullah, Z.; Ali, S.M.; Mehmood, C.A.A. Machine learning based energy management model for smart grid and renewable energy districts. IEEE Access 2020, 8, 185059–185078. [Google Scholar] [CrossRef]
  26. Oprea, S.-V.; Bâra, A. Machine learning algorithms for short-term load forecast in residential buildings using smart meters, sensors and big data solutions. IEEE Access 2019, 7, 177874–177889. [Google Scholar] [CrossRef]
  27. Shirzadi, N.; Nizami, A.; Khazen, M.; Nik-Bakht, M. Medium-term regional electricity load forecasting through machine learning and deep learning. Designs 2021, 5, 27. [Google Scholar] [CrossRef]
  28. López, M.; Sans, C.; Valero, S.; Senabre, C. Empirical comparison of neural network and auto-regressive models in short-term load forecasting. Energies 2018, 11, 2080. [Google Scholar] [CrossRef] [Green Version]
  29. Amjady, N.; Keynia, F. A new neural network approach to short term load forecasting of electrical power systems. Energies 2011, 4, 488–503. [Google Scholar] [CrossRef]
  30. Li, W.; Shi, Q.; Sibtain, M.; Li, D.; Mbanze, D.E. A hybrid forecasting model for short-term power load based on sample entropy, two-phase decomposition and whale algorithm optimized support vector regression. IEEE Access 2020, 8, 166907–166921. [Google Scholar] [CrossRef]
  31. Mamun, A.A.; Sohel, M.; Mohammad, N.; Sunny, M.S.H.; Dipta, D.R.; Hos, E. A comprehensive review of the load forecasting techniques using single and hybrid predictive models. IEEE Access 2020, 8, 34911–134939. [Google Scholar] [CrossRef]
  32. Shi, H.; Xu, M.; Li, R. Deep learning for household load forecasting—A novel pooling deep RNN. IEEE Trans. Smart Grid 2018, 9, 5271–5280. [Google Scholar] [CrossRef]
  33. Fan, G.-F.; Guo, Y.-H.; Zheng, J.-M.; Hong, W.-C. Application of the weighted K-nearest neighbor algorithm for short-term load forecasting. Energies 2019, 12, 916. [Google Scholar] [CrossRef] [Green Version]
  34. Madrid, E.A.; Antonio, N. Short-term electricity load forecasting with machine learning. Information 2021, 12, 50. [Google Scholar] [CrossRef]
  35. Román-Portabales, A.; López-Nores, M.; Pazos-Arias, J.J. Systematic review of electricity demand forecast using ANN-based machine learning algorithms. Sensors 2021, 21, 4544. [Google Scholar] [CrossRef] [PubMed]
  36. Fallah, S.N.; Deo, R.C.; Shojafar, M.; Conti, M.; Shamshirband, S. Computational intelligence approaches for energy load forecasting in smart energy management grids: State of the art, future challenges, and research directions. Energies 2018, 11, 596. [Google Scholar] [CrossRef] [Green Version]
  37. Velasco, L.C.P.; Estoperez, N.R.; Jayson, R.J.R.; Sabijon, C.J.T.; Sayles, V.C. Day-ahead base, intermediate, and peak load forecasting using k-means and artificial neural networks. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 62–67. [Google Scholar]
  38. Li, S.; Wang, P.; Goel, L. A novel wavelet-based ensemble method for short-term load forecasting with hybrid neural networks and feature selection. IEEE Trans. Power Syst. 2016, 31, 1788–1798. [Google Scholar] [CrossRef]
  39. Sun, W.; Zhang, C. A Hybrid BA-ELM model based on factor analysis and similar-day approach for short-term load forecasting. Energies 2018, 11, 1282. [Google Scholar] [CrossRef] [Green Version]
  40. Zhang, Y.-F.; Chiang, H.-D. Enhanced ELITE-load: A novel CMPSOATT methodology constructing short-term load forecasting model for industrial applications. IEEE Trans. Ind. Inform. 2020, 16, 2325–2334. [Google Scholar] [CrossRef]
  41. Qiu, X.; Ren, Y.; Suganthan, P.N.; Amaratunga, G.A.J. Empirical mode decomposition based ensemble deep learning for load demand time series forecasting. Appl. Soft Comput. 2017, 54, 246–255. [Google Scholar] [CrossRef]
  42. Jahan, I.S.; Snasel, V.; Misak, S. Intelligent systems for power load forecasting: A study review. Energies 2020, 13, 6105. [Google Scholar] [CrossRef]
  43. Turhan, C.; Simani, S.; Zajic, I.; Akkurt, G.G. Performance analysis of data-driven and model-based control strategies applied to a thermal unit model. Energies 2017, 10, 67. [Google Scholar] [CrossRef] [Green Version]
  44. Jallal, M.A.; González-Vidal, A.; Skarmeta, A.F.; Chabaa, A.; Zerouala, A. A hybrid neuro-fuzzy inference system-based algorithm for time series forecasting applied to energy consumption prediction. Applied Energy 2020, 268, 114977. [Google Scholar] [CrossRef]
  45. Buitrago, J.; Asfour, S. Short-term forecasting of electric loads using nonlinear autoregressive artificial neural networks with exogenous vector inputs. Energies 2017, 10, 40. [Google Scholar] [CrossRef] [Green Version]
  46. Bouktif, S.; Fiaz, A.; Ouni, A.; Serhani, M.A. Optimal deep learning LSTM model for electric load forecasting using feature selection and genetic algorithm: Comparison with machine learning approaches. Energies 2018, 11, 1636. [Google Scholar] [CrossRef] [Green Version]
  47. Komorowski, M.; Marshal, D.C.; Salciccioli, l.D.; Crutain, Y. Secondary Analysis of Electronic Health Records; Springer: Berlin/Heidelberg, Germany, 2016; Chapter 15; pp. 185–203. [Google Scholar]
  48. Power Information Technology Company. Available online: http://www.pitc.com.pk (accessed on 1 April 2020).
  49. rp5.ru. Available online: www.rp5.ru/Weather_in_the_world (accessed on 1 April 2020).
  50. Rajbhandari, Y.; Marahatta, A.; Ghimire, B.; Shrestha, A.; Gachhadar, A.; Thapa, A.; Chapagain, K.; Korba, P. Impact study of temperature on the time series electricity demand of urban nepal for short-term load forecasting. Appl. Syst. Innov. 2021, 4, 43. [Google Scholar] [CrossRef]
  51. Tudose, A.M.; Picioroaga, I.I.; Sidea, D.O.; Bulac, C.; Boicea, V.A. Short-term load forecasting using convolutional neural networks in COVID-19 context: The romanian case study. Energies 2021, 14, 4046. [Google Scholar] [CrossRef]
  52. Diversi, R.; Guidorzi, R.; Soverini, U. Identification of ARX and ARARX models in the presence of input. Eur. J. Control. 2010, 16, 242–255. [Google Scholar] [CrossRef]
  53. Fung, E.H.; Wong, Y.; Ho, H.; Mignolet, M.P. Modelling and prediction of machining errors using ARMAX and NARMAX structures. Appl. Math. Model. 2003, 27, 611–627. [Google Scholar] [CrossRef]
  54. Schnyer, D.A.M. Machine Learning; Methods and Applications to Brain Disorders; Academic Press: Cambridge, MA, USA, 2020; Chapter 6; pp. 101–121. [Google Scholar]
  55. Winters-Miner, L.A.; Bolding, P. Practical Predictive Analytics and Decisioning Systems for Medicine; Informatics Accuracy and Cost-Effectiveness for Healthcare Administration and Delivery Including Medical Research; Academic Press: Cambridge, MA, USA, 2015; Chapter 15; pp. 239–259. [Google Scholar]
  56. Saeed, M.S.; Mustafa, M.W.; Sheikh, U.U.; Jumani, T.A.; Mirjat, N.H. Ensemble bagged tree based classification for reducing non-technical losses in multan electric power company of Pakistan. Electronics 2019, 8, 860. [Google Scholar] [CrossRef] [Green Version]
  57. Tarkhaneh, O.; Shen, H. Training of feedforward neural networks for data classification using hybrid particle swarm optimization, mantegna lévy flight and neighborhood search. Heliyon 2019, 5, e01275. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  58. Mathworks. Available online: https://www.mathworks.com/help/deeplearning/ref/trainlm.html (accessed on 15 April 2021).
  59. Uzair, M.; Jamil, N. Effects of hidden layers on the efficiency of neural networks. In Proceedings of the 2020 IEEE 23rd International Multitopic Conference (INMIC), Bahawalpur, Pakistan, 5–7 November 2020; pp. 1–6. [Google Scholar]
Figure 1. Exploratory Data Analysis: (a) Auto-correlation plot of electrical load consumption; (b) QQ plot of temperature vs. electrical load; (c) QQ plot of humidity vs. electrical load; (d) QQ plot of present electrical load vs. previous hour lagged load values; (e) QQ plot of present electrical load vs. previous day same hour lagged load values; (f) electrical load consumption curve of the entire month; (g) box plot of electrical load consumption.
Figure 1. Exploratory Data Analysis: (a) Auto-correlation plot of electrical load consumption; (b) QQ plot of temperature vs. electrical load; (c) QQ plot of humidity vs. electrical load; (d) QQ plot of present electrical load vs. previous hour lagged load values; (e) QQ plot of present electrical load vs. previous day same hour lagged load values; (f) electrical load consumption curve of the entire month; (g) box plot of electrical load consumption.
Energies 14 05510 g001aEnergies 14 05510 g001b
Figure 2. Statistical model representation: (a) the ARX model structure [52]; (b) the ARMAX model structure [53]; (c) the Output Error model structure [53].
Figure 2. Statistical model representation: (a) the ARX model structure [52]; (b) the ARMAX model structure [53]; (c) the Output Error model structure [53].
Energies 14 05510 g002
Figure 3. Illustration of Artificial Neural Network (ANN) architecture [57].
Figure 3. Illustration of Artificial Neural Network (ANN) architecture [57].
Energies 14 05510 g003
Figure 4. Predicted electrical load output of January-2019 using various methodologies: (a) ARX model; (b) ARMAX model; (c) OE model; (d) KNN model; (e) Bagged Trees model; (f) SVM model; (g) NN–PSO mode; (h) NN–LM model with two hidden layers; (i) NN–LM model with single hidden layer.
Figure 4. Predicted electrical load output of January-2019 using various methodologies: (a) ARX model; (b) ARMAX model; (c) OE model; (d) KNN model; (e) Bagged Trees model; (f) SVM model; (g) NN–PSO mode; (h) NN–LM model with two hidden layers; (i) NN–LM model with single hidden layer.
Energies 14 05510 g004
Figure 5. Predicted electrical load output of June-2019 using various methodologies: (a) ARX model; (b) ARMAX model; (c) OE model; (d) KNN model; (e) Bagged Trees model; (f) SVM model; (g) NN–PSO mode; (h) NN–LM model with two hidden layers; (i) NN–LM model with single hidden layer.
Figure 5. Predicted electrical load output of June-2019 using various methodologies: (a) ARX model; (b) ARMAX model; (c) OE model; (d) KNN model; (e) Bagged Trees model; (f) SVM model; (g) NN–PSO mode; (h) NN–LM model with two hidden layers; (i) NN–LM model with single hidden layer.
Energies 14 05510 g005
Figure 6. Mean Absolute Percentage Error (MAPE) between actual and forecasted electrical load of year 2019 for various models: (a) statistical models; (b) machine learning models.
Figure 6. Mean Absolute Percentage Error (MAPE) between actual and forecasted electrical load of year 2019 for various models: (a) statistical models; (b) machine learning models.
Energies 14 05510 g006
Figure 7. Auto−correlation analysis for April-2019.
Figure 7. Auto−correlation analysis for April-2019.
Energies 14 05510 g007
Table 1. Selection of potential input parameters for STLF models using EDA.
Table 1. Selection of potential input parameters for STLF models using EDA.
Meteorological FactorsTemporal FactorsHistorical Data
TemperatureHour of the DayPrevious Hour Electrical Load
HumidityDay of the WeekPrevious Day Same Hour Electrical Load
Is it a Working Day?
Table 2. Quantitative comparison of alternative models for short-term electrical load forecasting of winter season.
Table 2. Quantitative comparison of alternative models for short-term electrical load forecasting of winter season.
Period (2019)OEARXARMAXKNNBagged TreesSVMNN–PSOANN–LM
(Two Hidden Layer)
ANN–LM
(Single Hidden Layer)
Errors
January8.426.825.127.564.463.713.953.852.47MAPE
142.87120.0892.05136.1580.5165.3271.1470.1244.41MAE
173.14155.08120.25172.14105.1287.0895.4290.9457.96RMSE
0.650.720.830.650.870.910.890.900.96 R 2
4.614.184.314.894.404.874.814.914.87Std. Dev.
February5.473.122.756.333.503.423.464.972.76MAPE
92.0953.6647.92109.6358.9458.1360.2984.2347.63MAE
110.5772.1566.10143.7183.8177.1981.73123.6266.33RMSE
0.830.930.940.710.900.920.910.780.94 R 2
4.844.784.924.874.544.664.604.944.66Std. Dev.
March4.233.293.355.322.992.962.873.402.28MAPE
77.8660.5661.91100.3955.1454.8552.8762.4342.49MAE
127.8494.3395.23129.9477.4776.4877.8887.4960.54RMSE
0.750.860.860.740.910.910.910.880.94 R 2
5.164.514.704.234.394.534.574.354.53Std. Dev.
October5.612.453.275.562.612.693.374.671.92MAPE
125.0055.4272.83126.2858.4759.9175.67106.0743.30MAE
145.4986.89103.54161.3885.1085.89101.08147.0664.47RMSE
0.740.910.870.690.910.910.880.740.95 R 2
5.165.114.924.974.985.115.305.385.11Std. Dev.
November2.332.072.304.792.332.392.372.311.69MAPE
42.5238.5342.9188.1243.5444.5444.2543.0731.40MAE
121.2654.3458.60113.3059.0159.6659.6660.2743.90RMSE
0.740.950.940.770.940.940.940.930.97 R 2
4.644.304.294.364.214.254.354.264.25Std. Dev.
December3.752.713.295.212.532.532.983.671.65MAPE
63.4647.7557.8292.2244.5944.5251.9964.1529.09MAE
85.9362.6274.29122.2560.6458.1667.8384.8438.26RMSE
0.900.950.930.800.950.960.940.910.98 R 2
5.594.724.564.924.744.844.734.894.84Std. Dev.
Table 3. Quantitative comparison of alternative models for short-term electrical load forecasting of summer season.
Table 3. Quantitative comparison of alternative models for short-term electrical load forecasting of summer season.
Period (2019)OEARXARMAXKNNBagged TreesSVMNN–PSOANN–LM
(Two Hidden Layer)
ANN–LM
(Single Hidden Layer)
Errors
April4.462.263.056.493.582.982.643.262.24MAPE
105.4053.0372.20153.6979.7968.9059.4976.4152.98MAE
258.59108.92119.80206.89132.18111.2594.67103.6276.53RMSE
0.500.910.890.680.870.910.930.920.96 R 2
7.826.486.545.366.066.336.336.556.33Std. Dev.
May2.374.303.505.833.412.603.182.661.99MAPE
72.45123.04104.10180.37104.9576.8997.6482.6161.28MAE
104.58180.15160.74234.70152.99118.70143.03113.8284.12RMSE
0.930.780.830.640.850.910.870.920.95 R 2
6.606.826.586.175.956.746.176.476.74Std. Dev.
June3.712.863.826.363.052.432.443.001.81MAPE
120.3890.67122.98209.5499.1075.5379.6098.4158.77MAE
186.74143.48171.35269.87152.86130.70113.97126.4182.45RMSE
0.750.850.790.480.830.880.910.890.95 R 2
6.676.265.965.635.676.336.345.996.33Std. Dev.
July5.192.442.595.243.402.484.434.431.82MAPE
178.2977.4082.01170.63112.8477.61151.55148.2459.70MAE
217.83141.37148.72242.24177.35144.56205.33197.6095.08RMSE
0.730.890.870.670.820.880.760.780.95 R 2
9.197.457.306.256.477.396.387.237.39Std. Dev.
August2.721.751.723.622.131.702.685.911.72MAPE
90.9357.9556.84123.5972.3556.2288.81204.7657.19MAE
126.5095.9399.05165.82110.3192.61133.20253.3690.37RMSE
0.840.910.900.730.880.920.820.370.92 R 2
5.395.305.484.795.065.454.953.205.45Std. Dev.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Javed, U.; Ijaz, K.; Jawad, M.; Ansari, E.A.; Shabbir, N.; Kütt, L.; Husev, O. Exploratory Data Analysis Based Short-Term Electrical Load Forecasting: A Comprehensive Analysis. Energies 2021, 14, 5510. https://doi.org/10.3390/en14175510

AMA Style

Javed U, Ijaz K, Jawad M, Ansari EA, Shabbir N, Kütt L, Husev O. Exploratory Data Analysis Based Short-Term Electrical Load Forecasting: A Comprehensive Analysis. Energies. 2021; 14(17):5510. https://doi.org/10.3390/en14175510

Chicago/Turabian Style

Javed, Umar, Khalid Ijaz, Muhammad Jawad, Ejaz A. Ansari, Noman Shabbir, Lauri Kütt, and Oleksandr Husev. 2021. "Exploratory Data Analysis Based Short-Term Electrical Load Forecasting: A Comprehensive Analysis" Energies 14, no. 17: 5510. https://doi.org/10.3390/en14175510

APA Style

Javed, U., Ijaz, K., Jawad, M., Ansari, E. A., Shabbir, N., Kütt, L., & Husev, O. (2021). Exploratory Data Analysis Based Short-Term Electrical Load Forecasting: A Comprehensive Analysis. Energies, 14(17), 5510. https://doi.org/10.3390/en14175510

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop