Next Article in Journal
A Full-Color Holographic System Based on Taylor Rayleigh–Sommerfeld Diffraction Point Cloud Grid Algorithm
Previous Article in Journal
Design and Test of a Levelling System for a Mobile Safflower Picking Platform
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Short-Term Power Load Forecasting Based on an EPT-VMD-TCN-TPA Model

School of Information and Electrical Engineering, Hebei University of Engineering, Handan 056038, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(7), 4462; https://doi.org/10.3390/app13074462
Submission received: 1 March 2023 / Revised: 28 March 2023 / Accepted: 29 March 2023 / Published: 31 March 2023

Abstract

:
Accurate short-term load forecasting is the key to ensuring smooth and efficient power system operation and power market dispatch planning. However, the nonlinear, non-stationary, and time series nature of load sequences makes load forecasting difficult. To address these problems, this paper proposes a short-term load forecasting method (EPT-VMD-TCN-TPA) based on the hybrid decomposition of load sequences, which combines ensemble patch transform (EPT), variational modal decomposition (VMD), a temporal convolutional network (TCN), and a temporal pattern attention mechanism (TPA). In which, the trend component (Tr(t)) and the residual fluctuation component (Re(t)) of the load series are extracted using EPT, and then the Re(t) component is decomposed into intrinsic modal function components (IMFs) of different frequencies using VMD. The Tr(t) and IMFs components of the fused meteorological data are predicted separately by the TCN-TPA prediction model, and finally, the prediction results of each component are reconstructed and superimposed to obtain the final predicted value of the load. In addition, experiments after reconstructing each IMF component according to the fuzzy entropy (FE) values are discussed in this paper. To evaluate the performance of the proposed method in this paper, we used datasets from two Areas of the 9th Mathematical Modeling Contest in China. The experimental results show that the predictive precision of the EPT-VMD-TCN-TPA model outperforms other comparative models. More specifically, the experimental results of the EPT-VMD-TCN-TPA method had a MAPE of 1.25% and 1.58% on Area 1 and Area 2 test sets, respectively.

1. Introduction

As countries and regions around the world rely more and more on electricity, this requires a higher level of security, reliability, and economy in the supply of electricity [1]. For example, China’s green development goals of “carbon peaking” and “carbon neutrality” also make electricity play an important role in the national economy [2]. With the advancement of technology, the storage, generation, transmission, distribution, and consumption of electrical energy is no longer an isolated process; therefore, in the traditional power system, effective measures must be taken to maintain the dynamic balance of electrical energy supply and demand [3]. The use of accurate power load forecasting not only provides a reference for the dynamic balance between power supply and demand, but also facilitates power plants to reasonably dispatch power generation, schedule the start and stop of generating units, improve the utilization rate of power generation equipment, and reduce the cost of power generation [4]. Therefore, accurate load forecasting is of great importance to the development of power systems.
According to the length of the forecast period, power load forecasting can be divided into four types: ultra-short-term load forecasting, short-term load forecasting, medium-term load forecasting, and long-term load forecasting [5]. Different cycles of load forecasting have different roles and meanings, among which online control and real-time monitoring of power systems rely on ultra-short-term load forecasting. Short-term load forecasting can coordinate the daily start/stop and generation plans of each generating unit, while medium- and long-term load forecasting aims to provide a reference basis for the long-term planning of power systems, etc. [6]. In previous studies, numerous scholars have proposed many methods to improve the accuracy of load forecasting, most of which focus on methods for short-term load forecasting. These methods are divided into three main categories based on statistical models, artificial intelligence prediction models, and combined prediction models, as shown in Figure 1. Multiple linear regression (MLR) [7], exponential smoothing (ES) [8], Kalman filter [9], the Hidden Markov model (HMM) [10], and the Autoregressive Integrated Moving Average Model (ARIMA) [11] are commonly used in statistical models for load forecasting. In Ref. [8], Christiaanse et al. developed a forecasting model with high accuracy and simple operation based on the exponential smoothing method and analyzed the role of exponential smoothing in load forecasting. In Ref. [11], Wu et al. proposed an autoregressive differential moving average model based on long-term dependence and combined it with the cuckoo search algorithm to optimize the parameters of the prediction model. The experimental results show that the ARIMA model has good prediction results in short-term load forecasting. In Ref. [12], Duan L et al. considered the external factors and the power system itself in medium- and long-term load forecasting and used a multilevel recursive regression method for medium- and long-term load forecasting to improve the forecasting accuracy by accurately predicting the important functions of time-varying parameters and related factors. Although statistical model-based methods have the characteristics of relatively simple models and fast prediction, these models usually target linear relational models, which require high smoothness of existing data, and non-linear relationships are more difficult to be captured accurately. Due to the combination of multiple factors, the variation of the load series has high uncertainty and complexity, which makes the traditional statistical model prediction unsatisfactory. In recent years, with the improvement of computer performance, the forecasting methods of artificial intelligence models have been applied to nonlinear time series load forecasting problems.
Short-term load forecasting methods based on artificial intelligence mainly include machine learning and deep learning models. Through machine learning techniques, we can learn and mine the non-linear relationship between the independent and dependent variables, and build regression models to predict short-term loads. Some common algorithms, such as support vector regression (SVR) [13], back propagation neuron networks (BPNNs) [14], and artificial neural networks (ANNs) [15], have been widely used for load forecasting. In Ref. [13], the authors implemented a generic short-term load forecasting strategy based on support vector machines by improving the generation process in the SVR input model and selecting the hyperparameters of the SVR model based on a particle swarm optimization algorithm, which was tested on two publicly available datasets to verify its effectiveness. In Ref. [14], the authors proposed a short-term load forecasting method based on a hybrid optimization algorithm and a BPNN for the load forecasting of paper enterprises, and the experimental results showed that the forecasting results of the forecasting model using the hybrid optimization algorithm outperformed the forecasting results of the single model. Although machine learning methods can handle nonlinear data well, these methods cannot effectively extract data features of load sequences due to the complexity and time series nature of load data. As deep learning continues to evolve, it becomes the focal point of load forecasting. In recent years, convolutional neural networks (CNNs), deep belief networks (DBNs), and deep neural networks (DNNs) have achieved remarkable results in the field of short-term load forecasting [16,17,18]. Although these methods are effective in extracting features of multidimensional load data and thus improving prediction precision, they cannot capture the time series features of load sequences when used alone. A recurrent neural network (RNN) is a neural network based on gate structure control, which is able to adjust the information based on past timestamps and current input values, overcoming the problem that ANN models do not handle temporal load data well [19]. A RNN can adversely affect prediction accuracy when dealing with complex nonlinear data due to the possibility of gradient explosion and gradient disappearance, thus limiting its application. The long short-term memory (LSTM) [20] network is an improved version of the RNN, which effectively solves the problem of gradient disappearance and gradient explosion that may occur when the RNN processes complex data by providing memory function and gate structure. The model has been widely used for short-term load forecasting [21,22]. Weicong Kong et al. conducted experiments using a framework based on an LSTM recurrent neural network on a real dataset and finally achieved better prediction results [23]. However, it is difficult to select parameters for the LSTM network as the training data increases. In Ref. [24], the authors first preprocess the load data and predict the load through the established gated recursive unit (GRU) network. The GRU greatly reduces the number of network parameters by merging the forgetting and input gates of the LSTM, which greatly improves the performance of the neural network and makes it easier to achieve convergence. Considering that the LSTM and GRU can only encode the information of load sequences from front to back and cannot adequately consider the future information, some scholars have used bidirectional long short-term memory (BiLSTM) and bidirectional gated recurrent unit (BiGRU) for load prediction [25,26]. In Ref. [25], the authors design a BiLSTM forecasting model based on improved particle swarm optimization, which designs a selection mechanism taking into account trend similarity and momentary similarity, thus combining the predicted and historical similar values and improving the forecast precision. In Ref. [26], the authors proposed a CNN-BiGRU method for short-term load prediction in which high-dimensional features and time-dependent features of the load data were extracted using a CNN and BiGRU, respectively, and an attention mechanism was introduced to the BiGRU network, and the results showed that the method would be more effective than the LSTM method. The BiLSTM and BiGRU networks have both forward and backward sequential information inputs, which can fully extract information from the load data and thus may improve the accuracy of feature extraction to some extent. In order to solve the problems of CNN’s inability to capture sequence timing features and the lack of using convolutional operations to extract data features in recurrent neural networks, temporal convolutional networks (TCNs) were created. The TCN is a new architecture based on convolutional neural networks that can efficiently extract sequence timing features. The TCN has improved on the basis of the CNN for the time series problem and is able to extract feature information of nonlinear time series data, while the TCN model solves the problem of gradient explosion and gradient disappearance by initializing the weight parameters. For load prediction as a complex nonlinear optimization problem, Song et al. built a thermal load prediction model based on TCN networks to achieve the extraction of complex data features and accurate load prediction [27]. Due to the advantages of TCN networks, many scholars have used them for short-term load forecasting [28,29].
Considering the non-linear and fluctuating characteristics of power load data, many experts and scholars pre-process the power load data before forecasting. One approach is to use a suitable method to classify the time series data into deterministic and stochastic components, and then model them separately based on these two different types of components. The deterministic component takes into account the characteristics of the load series such as trend, annual, quarterly, and weekly cycles, and date similarity, while other stochastic influences will determine the stochastic component. In Ref. [30], Ismail Shah et al., considering the characteristics such as polar values of electric load data, proposed a function data modeling method that divides the sequence after polar value processing, into deterministic and random components, and uses generalized additive modeling techniques to model the deterministic component and three different models to model the random component. The experimental results show that the prediction error based on the deterministic component and the stochastic component is predicted separately with lower prediction errors than the classical comparison models. Meanwhile, in Ref. [31], Ismail Shah et al. first divided the time series (demand and price) into deterministic and stochastic components. The deterministic component is modeled using parametric and nonparametric methods, and the stochastic component is modeled using univariate autoregression (AR) and multivariate vector autoregression (VAR), and finally, the model proposed in the paper is known to have better forecasting ability by evaluating the indicators, and the effectiveness of the model proposed in the paper is further verified by statistical significance. In Ref. [32], the method based on component estimation also achieved better prediction results. Another approach is to use signal decomposition methods to decompose the load data sequence into several relatively smooth sequences, and then predict each component separately. These signal decomposition methods are mainly wavelet transforms, empirical modal decomposition (EMD), and improved methods based on EMD, etc. In Ref. [33], the authors propose a prediction method based on a combination of wavelet decomposition and second-order gray neural networks. In Ref. [34], the authors first decompose the load sequence using EMD, followed by modeling prediction using a BiLSTM network based on an attention mechanism. In Ref. [35], the authors first decompose the load sequences using an ensemble empirical modal decomposition (EEMD), followed by reconstructing enough new sequences based on the approximate entropy of each component, and finally use an extreme learning machine (ELM) to predict each sequence separately and superimpose their results. Although these decomposition methods make load forecasting more accurate to a certain extent, they also have certain problems. For example, the effect of wavelet transformation is related to the choice of wavelet basis, which is less applicable. EMD decomposition tends to lead to modal confusion and endpoint effects when there are significant step changes in the signal. The EEMD decomposition reduces the residual auxiliary noise and reconstruction errors during EMD decomposition by adding Gaussian white noise several times, thus reducing the generation of modal aliasing. In recent years, variational modal decomposition (VMD) has been widely used in the field of short-term power load forecasting [36,37]. With the VMD decomposition technique, Cai et al. divided the load sequence into more stable parts and used a predictive network model to make accurate predictions for these parts [38]. The experimental results show that the VMD decomposition technique can significantly improve the accuracy and precision of load prediction. VMD is fundamentally different from EMD in that it uses a non-recursive approach to achieve signal decomposition and can automatically adjust the number of decompositions according to the actual situation, thus effectively separating the intrinsic modal components and also effectively solving the problems of modal aliasing and endpoint effects. Recently, an ensemble patch transform (EPT) for signal decomposition and filtering has been proposed by Kim et al. [39]. This EPT method can effectively decompose complex signals by specific transformations and can extract features such as trends of non-smooth series. In the short-term load forecasting model proposed in this paper, we use a hybrid decomposition technique based on EPT and VMD to effectively decompose and extract the features of the load sequence.
In summary, a combined forecasting model based on optimization algorithms, attention mechanisms, and various forecasting methods can be found throughout, as seen in the general process of electric load forecasting shown in Figure 2. In Ref. [40], the authors used a combined approach of neural networks and particle swarm optimization algorithms to forecast load, which was tested on data from a North American electric utility, and acceptable accuracy was obtained. In Ref. [41], the authors developed a novel multi-dual forecasting system for electricity prices using multivariate and multi-input–multi-output structures. In this model, the data obtained from pre-processing are predicted by a back propagation network, a BiLSTM network, and a GRU network, and finally, the final prediction results are obtained based on the combined strategy of the salp swarm algorithm, and the experimental results show that the proposed method achieves good prediction results. In Ref. [42], the authors propose a hybrid GWO-CNN-BILSTM prediction model for the short-term load prediction of buildings. In this method, a CNN is used to extract features in load sequences, BILSTM can learn both forward and backward data features, and a GWO is used for parameter search optimization of the CNN and BILSTM networks, and finally, four different cases are analyzed to verify the effectiveness of the proposed method. A nice combinatorial model has also been proposed in Ref. [43].
Considering the volatility, nonlinearity, and time series of load series, this paper proposes a short-term load forecasting method based on the combination of EPT-VMD hybrid decomposition and the TCN-TPA network model, called EPT-VMD-TCN-TPA. Compared with other methods, the EPT method used in this paper before using VMD will extract the trend component of the load series, while the VMD method will further reduce the volatility of the remaining fluctuation components. The TCN model based on the temporal attention mechanism used in this paper not only focuses more on the temporal characteristics but also the TCN model is suitable for the prediction of large data volume. First, the trend component (Tr(t)) and the residual fluctuation component (Re(t)) of the load series are extracted using the EPT method. The residual fluctuation component (Re(t)) is then adaptively decomposed into intrinsic modal functions (IMFs) with different central frequencies using the VMD method. Finally, the Tr(t) component, as well as the IMF component, are combined with meteorological influence factors, respectively, and the TCN-TPA network is used to train and predict the electric load. We also discuss the complexity of using fuzzy entropy (FE) to calculate the components of IMFs, and the intrinsic modal function (IMF) components with similar complexity are reorganized into a new component (FE-IMF), which is called the EPT-VMD-FE-TPA-TCN load prediction model. Simulation modeling was conducted using two regional load datasets provided by the 9th National Student Electrician Mathematical Modeling Competition in China, and the results show that the proposed EPT-VMD-TCN-TPA short-term electric load forecasting model predicts better precision than other comparative models and has better applicability.
The main contributions of this paper are further elucidated as follows:
(1)
Since the load series has a certain fundamental trend, the fundamental trend dominates the direction of the load series over time. The established EPT can accurately extract the fundamental trend of the load series, thus improving the accuracy of short-term load forecasting.
(2)
The adopted VMD decomposes the residual fluctuation series into several subseries with different center frequencies, which further reduces the non-smoothness of the fluctuation series. The number of decompositions of the VMD is determined using the ratio of residual energy, which makes the decomposed subsequences smoother.
(3)
The experimental results show that the EPT-VMD hybrid decomposition method proposed in this paper is effective for load prediction.
(4)
The TCN-TPA network is used for the prediction of individual components so that the temporal characteristics of these sequences can be better extracted. The temporal pattern attention mechanism incorporated in the TCN network can better assign different weights to different variables at the same time step compared to the ordinary attention mechanism. The temporal pattern attention mechanism enhances the impact of temporal attributes of multivariate load sequences on load prediction.
(5)
In further experiments, FE was used to analyze the complexity of the IMF components and to reconstruct these components using similar entropy values. The experimental results show that the EPT-VMD-FE-TCN-TPA reconstructed model using FE has higher operational efficiency compared to the EPT-VMD-TCN-TPA prediction model, but the prediction precision is reduced.
The rest of the paper is organized as follows. Section 2 presents the relevant methods. Section 3 explores the proposed model. Section 4 provides an analysis and discussion of the experimental results. Section 5 outlines the experimental results on dataset II. Finally, Section 6 concludes the paper.

2. Related Theories and Methods

This section first introduces two data decomposition methods, ensemble patch transform (EPT) and variational modal decomposition (VMD). Then, a data sequence complexity analysis method, FE, is introduced. finally, the temporal convolutional network (TCN) and temporal pattern attention (TPA) mechanisms are introduced.

2.1. Ensemble Patch Transformation (EPT)

The ensemble patch transformation was proposed by Kim et al. [39]. The EPT method uses the multi-scale concept of computer vision scale space theory to enhance the recognition of local features in the signal data and extract the essential components of the signal data. It consists of two key components, the “patch process” and the “ensemble process”. Due to the flexibility of the EPT framework and the adaptability of the parameters, EPT can be used to discover specific patterns in signal data, such as trends, sudden changes, and periodicity. The text will use EPT to extract the trend component of the load series. Both components will be described below.

2.1.1. Patch Process

A “patch process” is a dependent patch of data at a given time t , designed to identify local features of the data. Two types of patches are considered in EPT, rectangular and oval patches. The differences between the two types of patches are not very obvious. Using an oval patch allows us to obtain the central tendency of the data, as the generated curves are smoother. However, there is a need to capture abrupt changes in the data in highly non-smooth load sequences, and in this case, a rectangular patch is more suitable than an oval patch. Therefore, a rectangular patch will be used in this paper for our proposed short-term load prediction model. Only the rectangular patch is described below.
Given a period τ , the patch P t τ ( X t ) of the load sequence X t at time t is centered at the point ( t , X t ) . The patch at this location ( t , X t ) is a closed rectangle formed by the points ( t + k , min k [ τ / 2 , τ / 2 ] { X t + k } 0.5 γ τ ) and ( t + k , max k [ τ / 2 , τ / 2 ] { X t + k } + 0.5 γ τ ) , where k [ τ / 2 , τ / 2 ] , the width of the patch is τ and the height is h t τ .
h t τ = max k [ τ / 2 , τ / 2 ] { X t + K } min k [ τ / 2 , τ / 2 ] { X t + K } + λ τ
where γ is the scale factor.
In order to make the upper and lower envelopes of the extracted load series more homogeneous and thus extract a more stable load trend, the value of γ is taken as 1 in this paper.
The upper envelope of the rectangular patch is U t τ ( X t ) , the lower envelope is L t τ ( X t ) , and the average envelope is M t τ ( X t ) :
U t τ ( X t ) = max k [ τ / 2 , τ / 2 ] { X t + k } + 0.5 γ τ
L t τ ( X t ) = max k [ τ / 2 , τ / 2 ] { X t + k } 0.5 γ τ
M t τ ( X t ) = 1 2 ( U t τ ( X t ) + L t τ ( X t ) )

2.1.2. Ensemble Process

The “ensemble process” is an “ensemble” obtained by shifting the time point t of the patch to represent the temporal transformation of the data by enhancing the temporal resolution of the data. Once the patching process is introduced, ensemble patch transformations can be defined by moving patches by a point in time t . The ensemble patch transformation is defined as follows.
For any fixed period τ , the th shifted patch at time point t is defined as P t + τ ( X t ) , [ τ / 2 , τ / 2 ] .
Thus, a collection of all possible shifted patches at time point t is defined as an ensemble patch:
E P t τ ( X t ) : = { P t + τ ( X t ) : [ τ / 2 , τ / 2 ] }
After that, we can obtain the high-frequency mode H F t τ ( X t ) and the low-frequency mode E M t τ ( X t ) :
H F t τ ( X t ) = X t E M t τ ( X t )
E M t τ ( X t ) = a v e r a g e ( M t + τ ( X t ) ) o v e r s
In the EPT framework, the period τ is chosen very flexibly, so that different temporal features in the signal can be extracted using the EPT method. In load forecasting, the load sequence has a certain daily period, monthly period, etc. Since the load dataset used in this paper is relatively dense with 96 sampling points a day sampled every 15 min, the period τ is fixed to a daily length.

2.2. Variational Modal Decomposition (VMD)

In many cases, due to the volatility of the signal data sequence, we need to decompose the signal sequence into a series of smooth subsequences in order to better perform subsequent processing. There are many commonly used decomposition methods, for example, EMD and EEMD decomposition methods, but all these methods have some limitations. In 2014, Dragomiretskiy et al. proposed a completely non-recursive variational modal decomposition method (VMD), which adaptively decomposes complex timing signals into several simple modal components [44]. The VMD algorithm is an effective method for applying the traditional Wiener filter to multiple adaptive bands, which has a broad theoretical basis and is easy to implement. VMD decomposition is a way to convert complex mathematical problems into simple mathematical expressions by building variational problems, finding the best solution, and improving the performance of the model using alternating direction multipliers to improve sampling accuracy. The calculation flow of VMD is shown in Figure 3. The specific steps are as follows.
(1)
Constructing variational problems.
Suppose the original input signal f ( t ) is decomposed into k components u . In order to ensure that the decomposed sequence is a modal component with a finite bandwidth of the central frequency, the sum of the estimated bandwidths of each modal component is required to be minimal, and the constraint is that the sum of all modal components should be equal to the original signal. The corresponding expression for the constrained variational problem is as follows:
min { u k } { w k } { k t [ δ ( t ) + j π t ] × u k ( t ) e j w k t 2 2 }
s · t · k u k = f ( t )
where f ( t ) denotes the original signal, k is the predetermined number of modal components, t is the Dirac distribution function, u k = { u 1 , u 2 , , u k } is the k modal components obtained from the adaptive decomposition of the original signal, and w k = { w 1 , w 2 , , w k } is the central frequency of each modal component.
(2)
Transformation of variational problems.
By using the quadratic penalty term α and the Lagrangian operator λ ( t ) , we can convert the variational problem with constraints into a problem with unconstrained variation by the following equation:
L ( { u k } , { ω k } , λ ) : = α k t [ ( δ ( t ) + j π t ) × u k ( t ) ] e j w k z k 2 + f ( t ) k u k ( t ) 2 2 + λ ( t ) , f ( t ) k u k ( t )
(3)
Solve the variational problem.
By using the alternating direction multiplier method (ADMM), we can solve the initial minimization problem. The Fourier isometric transform is used to obtain the results of the sequence decomposition in the frequency domain. u k n + 1 , ω k n + 1 , and λ are solved iteratively by searching the saddle points of the increasing Lagrangian function with the following equation:
u ^ k n + 1 ( ω ) = f ^ ( ω ) i k u ^ ( ω ) + λ ^ ( ω ) 2 1 + 2 α ( ω ω k ) 2
ω k n + 1 = 0 ω | u ^ ( ω ) | 2 d ω 0 | u ^ k ( ω ) | 2 d ω
λ ^ n + 1 ( ω ) = λ ^ n ( ω ) + τ ( f ^ ( ω ) k u ^ k n + 1 ( ω ) )
where u ^ k n + 1 ( ω ) is the kth modal component with the center frequency of ω at the (n + 1) th iteration. ω k n + 1 is the center frequency of the kth modal component during the (n + 1) iteration. τ is the noise tolerance of the signal.
After repeated iterations until the convergence condition is reached or the maximum number of iterations is reached and stopped, the convergence condition can be expressed as:
k u ^ k n + 1 u ^ k n 2 2 u ^ k n 2 2 < ε

2.3. Fuzzy Entropy (FE)

Fuzzy entropy was proposed by Chen et al. in 2007 and was originally intended for use in myoelectric model processing, although it has since been widely used in areas such as fault diagnosis and image processing [45]. FE is a valid measure of time series complexity, and it can reflect the complexity of the data more accurately by improving the sample entropy and approximate entropy. The FE algorithm calculates the entropy value of the time series by introducing a de-meaning operation that defines the similarity of the vectors in the phase space reconstruction formula as an exponential function. The increase in the FE value implies an increase in the complexity of the time series. The steps of its algorithm are as follows:
(1)
Knowing an N -dimensional time series [ μ ( 1 ) , μ ( 2 ) , , u ( N ) ] , let the phase space dimension be m   ( m n 2 ) and the phase space vector be:
X i = [ μ ( i ) , μ ( i + 1 ) , , μ ( i + m 1 ) ] μ 0 ( i ) , i = 1 , 2 , , N m + 1
where μ 0 ( i ) = 1 m j = 0 m 1 μ ( i + j ) denotes the spatial mean.
(2)
The maximum difference between the corresponding elements of X i and X j is defined as the distance d i , j m .
d i , j m = d [ X i , X j ] = max [ | μ ( i + p ) μ 0 ( i ) ( μ ( j + p ) μ 0 ( j ) ) | ] , p = 1 , 2 , , m 1
(3)
Introduce the affiliation function as the following equation:
A ( x ) = { 1 x = 0 exp [ ln ( 2 ) ( x r ) n ] x > 0
where r is the similar tolerance boundary width and n is the similar tolerance boundary gradient.
For the phase space, there is Equation (18).
A i , j m = exp [ ln ( 2 ) ( d i , j m r ) n ] , j = 1 , 2 , , N m + 1 , j i
(4)
Define the function.
C i m ( n , r ) = 1 N m j = 1 , j i N m + 1 A i , j m
(5)
Define the function.
ϕ i m ( r ) = ( N m + 1 ) 1 i = 1 , j i N m + 1 C i m ( n , r )
(6)
When N is a finite value, the value of FE can be expressed as Equation (21).
F E ( m , n , r ) = ln ϕ i m ( n , r ) ln ϕ i m + 1 ( n , r )
From the above, it can be seen that the size of the entropy value is related to the length of the data N and the parameters m , r , n , and r are generally taken as 0.1 0.5 σ S D ( σ S D is the standard deviation of the original sequence), and n plays a weighting role in the vector calculation of similarity and is generally taken as a smaller number (e.g., 1 or 2).

2.4. The Temporal Convolutional Network (TCN)

The TCN, as a new architecture based on convolutional neural networks (CNNs), can not only solve the problem of excessive time consumption by parallel computation but is also different from the CNN because the TCN is composed of causal convolution, dilated casual convolution, and residual block structure.

2.4.1. Causal Convolution

Causal convolution is used in the TCN network. Suppose that a given input sequence x 1 , x 2 , , x t is used to predict y 1 , y 2 , , y t . The output of the predicted value y t at the current moment in the causal convolution is only related to the input x t at the current moment and the input at the previous moment, which is also consistent with the impossibility of knowing future information in advance on the prediction problem. It can be expressed as the following:
y t = f ( x 1 , x 2 , , x t )
where f ( · ) denotes a convolution operation.

2.4.2. Dilated Casual Convolution

The historical data captured by causal convolution are only linearly related to the depth of the network, so the older the historical information traced, the more hidden layers it leads to. Thus, using a simple causal convolution does not capture a long time series. Therefore, the dilated causal convolution technique is used in the TCN to improve efficiency. The dilated causal convolution injects holes into the standard causal convolution, which allows the dilated causal convolution to obtain a larger perceptual field without changing the size of the convolution kernel and reduces the depth of the causal convolution without increasing the number of parameters. The structure of the dilated convolution is shown in Figure 4.
Assume that the one-dimensional sequence input to the temporal convolutional network is x R n and the convolutional kernel is f : { 0 , , k 1 } R . Define the null causal convolution operation s on the elements F in the sequence as follows:
F ( s ) = ( x f d ) ( s ) = i = 0 k 1 f ( i ) x s d i
where d i is the dilation factor of the i -th layer convolution, k is the size of the convolution kernel, and s d i is the direction of the sequence past time. Usually, as the number of layers increases, the dilation factor d grows exponentially by 2. The larger d is, the larger the input range is, thus increasing the perceptual field of the convolutional network, and when d = 1 , it is the standard causal convolution.

2.4.3. Residual Block

To increase the perceptual field, the TCN is implemented by increasing the convolutional kernel, increasing the expansion factor, and stacking the number of network layers, which deepens the network depth. However, a deeper network structure may cause problems, such as gradient disappearance, for which the residual block is introduced in the TCN. The relationship between the input x and output o of the residual block is shown in Equation (24), and the structure of the residual block is shown in Figure 5.
o = A c t i v a t i o n ( F 1 ( x ) + F 2 ( x ) )
where x is the input of the residual block, o is the output of the residual block, A c t i v a t i o n ( ) is the activation function, F 1 ( x ) is the output of the multilayer network, and F 2 ( x ) is the output of Conv1×1.
Each residual block contains a dilated causal convolution, Relu, WeightNorm, and Dropout, which together form the whole. The use of WeightNorm and Dropout techniques effectively prevents gradient disappearance and model overfitting. The advent of Conv1×1 greatly improves the performance of the network by allowing the network layers to learn constant mappings, thus effectively avoiding network degradation.

2.5. Temporal Pattern Attention (TPA)

The attention mechanism is inspired by an information processing mechanism derived from human vision. The core of the attention mechanism is to assign different weights to the input feature information, focus on important information, and ignore irrelevant information, so as to capture the global and local connections, optimize the prediction model, and improve the prediction accuracy. The temporal pattern attention mechanism proposed by Shun-Yao Shih et al. [46] can effectively solve the problem that the traditional attention mechanism only relies on weighting the time steps, and it can also better consider multivariate time series and thus can more accurately predict the relationship between variables. The TPA model is shown in Figure 6.
In the model in this paper, TPA firstly extracts fixed-length temporal patterns from the output information of the implicit layer of the TCN through CNN filters, determines the weights of each temporal pattern using a scoring function, and fuses the final output information according to the magnitude of the weights, as follows.
(1)
The TCN processing time series.
The original multivariate time series is processed using TCN to obtain the hidden state h i (column vector) for each time step, where each h i has dimension m and w is the length of the time window representing the selected range of historical data to obtain the hidden state matrix H = { h t w , h t w + 1 , , h t 1 } and the hidden state h t at the last moment.
(2)
CNN convolution.
k convolution kernels ( 1 × w ) are used to capture the signal patterns with variable hidden states, with the following equation:
H i , j C = l = 1 w H i , ( t w 1 + l ) × C j , T w + l
(3)
The attention weight.
Let q u e r y = h t , k e y = H C , then S c o r i n g   Function is:
f ( H i C , h t ) = ( H i C ) T W a h t
The attention weight α i is obtained by normalizing using the sigmod function, as in the following equation:
α i = s i g m o i d ( f ( H i C , h t ) )
Using the attention weight α i , a weighted summation is performed for each row H C to obtain the context vector v t , as in the following equation:
v t = i = 1 m α i H i C
(4)
Fusion.
Finally, v t and h t are fused and the output of the attention layer is obtained by linear transformation:
h t = W h h t + W v v t
y t 1 + Δ = W h h t

3. The Proposed EPT-VMD-TPA-TCN Model and the Restructured EPT-VMD-FE-TPA-TCN Model

In this paper, we propose a short-term power load forecasting model (EPT-VMD-TCN-TPA) based on the mixed decomposition of load data and TCN-TPA network forecasting. The Tr(t) component and Re(t) component of the load series are first extracted using EPT. Then, the Re(t) component is decomposed into several IMFs with different center frequencies using VMD, and the Tr(t) component and each IMF component are fused with weather and other influencing factors separately. Finally, the TCN-TPA network model is used to train and predict the data. The reorganization model EPT-VMD-FE-TPA-TCN discussed in this paper is based on the EPT-VMD-TPA-TCN model and uses the concept of fuzzy entropy (FE) to calculate the complexity of each IMF component. The IMF components of similar complexity are combined into a new component (FE-IMF), and finally, the TCN-TPA network is used to train and predict each fused weather factor FE-IMF component and Tr(t) component. The structural framework of the EPT-VMD-TPA-TCN prediction model proposed in this paper is shown in Figure 7.
The EPT-VMD-TCN-TPA prediction model proposed in this paper can be divided into three key parts: the EPT-VMD decomposition layer, the TCN-TPA network training and prediction layer, and the fully connected prediction result output layer. The EPT-VMD-FE-TCN-TPA recombination prediction model is added with the FE subsequence recombination layer after the EPT-VMD decomposition layer. This section details the key components of the experimental dataset and model.

3.1. Experimental Datasets

In this paper, the datasets of Area 1 and Area 2 provided by the 9th National Electrician Mathematical Modeling Competition in China are used as the experimental data sources. The data for each Area include load data and meteorological data between 1 January 2012 and 10 January 2015. The electrical load sampling interval was 15 min, and 96 points were taken per day in MW. The meteorological data include daily maximum temperature, minimum temperature, average temperature, rainfall, and relative humidity. The data from Area 1 are used as the main dataset for the experiment, and the dataset from Area 2 is used to verify the scalability of the prediction model proposed in this paper. In each Area, the first 80% of the data were used as the training set and the last 20% were used as the test set. The original load data series of Area 1 are shown in Figure 8. The trend of Area 1’s continuous four-week load data is shown in Figure 9. The load data statistics are shown in Table 1. The meteorological data for one week in Area 1 are described in Table 2 (only 1 August 2012–7 August 2012 are shown).
(1)
Dataset (Area 1): 1 January 2012–10 January 2015 (total 1106 days, 106,176 load data collection points).
(2)
Training set: data from 1 January 2012–2 June 2014 (884 days, 84,864 load data collection points) were used to train the model.
(3)
Test set: 3 June 2014–10 January 2015 (222 days, 21,312 load data sampling points) was used for the evaluation of the model.
Throughout the training and prediction process of the model, a sliding window training method is used, i.e., load data from the past consecutive week (168 h) and weather data are always selected to predict the future day’s load, and for each completed training, the data used for training are pushed back 24 h for the next model training.
According to Figure 8 and Figure 9, it can be seen that the power load data have obvious periodic characteristics and also show strong fluctuation. Let the original load data in the dataset be f ( t ) and the meteorological data be m ( t ) . The load forecast is determined by both the historical load data f ( t ) and meteorological data m ( t ) . Since the load data have strong volatility and periodicity, this paper will first process the load data, and then the processed data will be fused with the meteorological data. Finally, the fused data will be trained and predicted using the TCN-TPA network model. The processing of the load data will be described in detail below.

3.2. EPT-VMD Decomposition Layer

3.2.1. EPT Decomposition

From the previous analysis of the dataset, it is clear that the load data are not only non-stationary but also have a certain periodicity. If the VMD is used directly to decompose it, the trend of the load series will not be extracted. In this paper, we use EPT to process the load data f and extract the trend of the original load series, so as to obtain the trend component t and the fluctuation component R. The results of decomposing the load data series of Area 1 by using EPT are shown in Figure 10. The results of decomposing the load data series of Area 2 by using EPT are shown in Figure 11.
In Figure 10 and Figure 11, it can be seen that the Tr(t) component obtained using EPT extraction has a strong periodicity and regularity for both Area 1 and Area 2. The Re(t) component has non-smoothness and needs further processing to obtain a relatively smooth subsequence.

3.2.2. VMD

In the experiment, in order to reduce the volatility and randomness of the Re(t) component, VMD is used to decompose it so as to obtain several relatively smooth IMF components. According to Section 2.2, the number of IMFs, i.e., the value of k , must be determined before using VMD to decompose Re(t). If the value of k is too small, the decomposition result is inadequate and increases the error of prediction. If the value of k is too large, the decomposition is overdone, thus increasing the overhead of the decomposition. In this paper, we will use the method in reference [47] to determine the number k of the decomposed modal functions using the decomposed residuals (Rres), as in Equation (31):
R r e s = 1 M t = 1 M | T r ( t ) i = 1 k u i ( t ) T r ( t ) |
where Re(t) is the residual fluctuation component, u i ( t ) is the IMF subsequence obtained by VMD decomposition, the number of samples is M , and the number of IMF subsequences is k .
In this paper, when the value of Rres is less than 1% and there is no obvious downward trend, the value of k is the appropriate number of decompositions for VMD. The results of the Re ( t ) decomposition using different values of k are shown in Table 3.
As can be seen in Table 3, in Area 1, when the value of k is 11, Rres = 0.00338015 < 0.001, and when k > 11 , the decreasing trend of Rres tends to be stable. In Area 2, when the value of k is 11, Rres = 0.00198062 < 0.001, and when k > 11 , the decreasing trend of Rres tends to be stable. In this paper, the VMD algorithm has a penalty function α = 1000 , a noise tolerance τ = 0 , and a convergence tolerance ε = 1 × 10 - 6 .
When k is taken as 11, Figure 12 shows the results of the decomposition of the residual volatility component using VMD for the first month in Area 1. Figure 13 shows the results of the decomposition of the residual volatility component using VMD for the first month in Area 2.
As can be seen in Figure 12, the use of EPT and VMD for the electric load series not only extracts the trend component of the original load series but also VMD decomposes the fluctuating component series into smoother sub-series. In Area 2, Figure 13 similarly shows that the hybrid EPT-VMD decomposition method proposed in this paper is more effective.

3.3. Sub-Sequence Recombination Layer

The content in this subsection is part of a restructured model (EPT-VMD-FE-TPA-TCN) based on the proposed EPT-VMD-TPA-TCN prediction model in this paper. Since the central frequencies of each IMF component obtained by decomposing the fluctuating components using VMD are different, these IMFs will have different effects on load forecasting. Moreover, too many components from VMD decomposition will increase the time consumption of the prediction model. In this paper, we will apply fuzzy entropy to calculate the complexity of each IMF component in the EPT-VMD-FE-TPA-TCN reorganization prediction model and combine and reorganize the IMF components with similar fuzzy entropy values into a new component (FE-IMF).
In this paper, the fuzzy entropy values are calculated by taking the sequence length N as the total sample length, the phase space dimension m = 2 , the similarity tolerance limit r = 0.3 S d , and n = 2 . The fuzzy entropy values of each IMF component are shown in Figure 14.
In Figure 14, we can see that the fuzzy entropy of IMF components varies widely but there are some components with similar fuzzy entropy values, which indicates that there is some similarity among them. The fuzzy entropy of IMF1 is significantly lower than other components, while the fuzzy entropy of IMF7 is the highest. The IMF components with similar fuzzy entropy values are reorganized, as shown in Table 4.
A comparison of the original load series of the first month in Area 1 and the series after decomposition by EPT-VMD and reorganization by FE is shown in Figure 15. From the comparison in Figure 15, it can be seen that the trend component of the original load series can be advanced by EPT, the remaining fluctuation component can be decomposed into relatively stable IFM components by VMD, and the number of components is reduced by using FE to combine IMF components with similar fuzzy entropy values into FE-IMF.

3.4. TCN-TPA Network Training and Prediction Layers

Short-term power load forecasting is a multivariate time series task because short-term power load forecasting is not only related to historical load but also to variables such as meteorological factors. For this multivariate time series forecasting task, this paper will use the TCN-TPA network model for short-term power load forecasting. Firstly, the TCN network can not only control the memory length of the model by adjusting the expansion rate and changing the convolutional kernel but also the TCN network has the ability to process data in a massively parallel way, which makes the training and validation time of the prediction model shorter. Secondly, using the temporal attention mechanism not only allows assigning different weights to the input information to highlight the key influencing factors but also breaks away from the typical attention mechanism’s tendency to select time steps that are more relevant to time series prediction. The temporal attention mechanism can better capture temporal information and span multiple time steps so that temporal patterns can be extracted for each load component and focus on relevant meteorological factors to improve prediction precision.

3.5. Fully Connected Prediction Result Output Layer

The fully connected layer can obtain output information from the TCN-TPA layer to obtain prediction results for multiple components. By inverse normalizing the prediction results of each component of the fully connected layer and superimposing them, accurate load prediction results can be obtained.

3.6. Workflow of the EPT-VMD-TCN-TPA Model

In this paper, an EPT-VMD-TCN-TPA short-term electric load forecasting model is proposed. The EPT-VMD-FE-TCN-TPA forecasting model is discussed on the basis of the EPT-VMD-TCN-TPA model. The workflow of the proposed EPT-VMD-TCN-TPA model is shown in Figure 16, and each step is described as follows:
Step 1: Pre-processing of the raw data.
Step 2: The pre-processed raw data are divided into load data f ( t ) and meteorological data m ( t ) .
Step 3: Use EPT to extract the trend component (Tr(t)) and the residual fluctuation component (Re(t)) of the load data f ( t ) .
Step 4: The residual fluctuation components are decomposed into IMF components using VMD according to the predefined k values.
Step 5: Each IMF component and Tr(t) component is fused with m ( t ) separately and the fused data are normalized to eliminate the effect of different dimensional data.
Step 6: The data fused in Step 5 are divided into a training set and a test set. The training set is fed into the TCN-TPA network and the hyperparameters of the model are improved and updated by calculating the loss function MSE to learn the complex relationships between the input and output variables in each subsequence.
Step 7: Each subsequence is predicted using the trained TCN-TPA network, and the prediction result of each subsequence is derived by the fully connected layer.
Step 8: The prediction results of each sub-series are summed up by the inverse normalization process, and the final accurate prediction result of the load is obtained.
The EPT-VMD-FE-TCN-TPA prediction model is added with the recombination operation after step 4. That is, the fuzzy entropy (FE) values of each IMF component are calculated and the IMFs with similar values are reorganized to obtain each reorganization component (FE-IMF).

4. Experiments

4.1. Datasets and Experimental Environment

The dataset was analyzed and described in Section 3.1. The experiments in this paper were performed with Python 3.8 running on 64-bit Windows 10 and the open-source library vmdpy to write the VMD programs. The models were developed using the Keras-based TesorFlow framework. The neural network was implemented in Keras 2.11. The environment parameters are shown in Table 5.
In the TCN-TPA prediction model in this paper, the hyperparameters are the number of convolution kernels, convolution kernel size, dilation rate, number of iterations, number of TCN residual blocks, dropout, and batch size. The number of convolutional kernels is set to (10,20,32,64), the size of the convolutional kernels is set to (1,2,3,4,5), the expansion factor is set to (1,2,4,8), the number of TCN residual blocks is set to (1,2,3), and the batch size is set to (16,32,64,128). In this paper, we used the GridSearchCV method from the Sklearn library in Python to perform the grid search for these hyperparameters [48] using MSE as the loss function, a learning rate set to 0.001, and an Adam optimizer to optimize the parameters. The hyperparameters shown in Table 6 and Table 7 are obtained experimentally.

4.2. Data Normalization

In the EPT-VMD-TCN-TPA prediction model proposed in this paper, the original data are divided into load data f ( t ) and meteorological data m ( t ) , where the trend component Tr(t) and the residual fluctuation component Re(t) are obtained by EPT operation on the load data f ( t ) . The Re(t) component is decomposed into k IMF components by VMD, and the result is denoted as d i ( t ) , i = 1 , 2 , 3 k . (where k is the number of IMF components).
The Tr(t) and the m ( t ) are fused and denoted as F t = F ( T r , m ) .
The fusion of components d i ( t ) and m ( t ) are denoted as F i t = F ( d i , m ) (where i = 1 , 2 , 3 k , k is the number of IMF components).
The fused F t and F i t are normalized and the normalized components are input into the prediction model, and through experiments, we can then accurately predict the value of the load. The final load prediction result consists of the superposition of the Tr(t) component and each IMF component prediction result, as shown in Equation (32):
P r e a l = P F t + i = 1 n P F i t
where p r e a l is the final forecasted value of the electric load, P F t is the resultant value of the trend component forecast, and P F i t is the resultant value of the IMF component forecast.
In order to eliminate the adverse effects of numerical bias on the model performance and speed up the training of the model, this paper will use the maximum–minimum normalization method to normalize the fused data F t and F i t . The normalization formula is shown in Equation (33):
x * = x x min x max x min
where x represents the fused sample data, x * is the normalized data, x min is the minimum value in the sample data, and x max is the maximum value in the sample data.

4.3. Error Evaluation Indicators

In this paper, we evaluate the error of the prediction results by using the mean absolute percentage error (MAPE), root-mean-square error (RMSE), mean absolute error (MAE), and R-squared (R2). The smaller the values of MAPE, RMSE, and MAE, the more accurate the load prediction results are, and the larger the value of R2, the better the fit between the load prediction series curve and the actual load series curve is and the more accurate the prediction results are. This paper also performs a statistical test for all models, namely the Diebold and Mariano (DM) test [49]. The formula for the four evaluation indicators is defined as follows:
M A P E = 1 N i = 1 N | y ^ i y i | y i × 100 %
R M S E = 1 N i = 1 N ( y ^ i y i ) 2
M A E = 1 N i = 1 N | y ^ i y i |
R 2 = 1 i = 1 N ( y i y ^ i ) 2 ( y i y ¯ i ) 2
y ¯ i = 1 N i = 1 N y i
where y i is the actual value of the load, y ^ i is the predicted value of the load, and N the number of load sample points.

4.4. Ablation Experiments

In order to verify the prediction effect produced by each step of the combined model proposed in this paper, we will build the TCN model, the VMD-TCN model, the VMD-TCN-Attention model, the VMD-TCN-TPA model, the final proposed EPT-VMD-TCN-TPA model, and the EPT-VMD-FE-TCN-TPA recombination model in this paper, respectively. Among them, these models are briefly described as follows:
(1)
TCN model [27]: This model is a more advanced model in recent years, which not only can effectively extract the features of nonlinear time series data but also can effectively solve complex problems such as gradient explosion and gradient disappearance during model training.
(2)
VMD-TCN model: The load data are firstly decomposed using VMD, after which the dataset is trained and predicted using the TCN network.
(3)
VMD-TCN-Attention model: A general attention mechanism is added to the VMD-TCN model.
(4)
VMD-TCN-TPA model: Based on the VMD-TCN model, the prediction capability is improved by introducing a temporal pattern attention mechanism, which is especially suitable for handling multi-task temporal prediction.
(5)
EPT-VMD-TCN-TPA model: Our proposed final model. Before using the VMD-TCN-TPA model, the EPT method, which extracts the trend components of the load data, is first introduced.
(6)
EPT-VMD-FE-TCN-TPA model: This model adds a recombination layer to the proposed EPT-VMD-TCN-TPA model in this paper and merges the similar components of the intermediate processes according to FE values.
The prediction effects of each of the above comparison models on the test set are shown in Figure 17 (only one day is shown), and the prediction error evaluation index data and fit effect data of each model are shown in Table 8.
As can be seen in Table 8, the EPT-VMD-TCN-TPA forecasting model proposed in this paper has the smallest MAPE, RMSE, and MAE compared with other comparative models, which are 1.25%, 110.2692 MW, and 82.3180 MW, respectively. Furthermore, the EPT-VMD-TCN-TPA model has the highest R2 value, 0.9955. These evaluation indexes reflect that the EPT-VMD-TCN-TPA forecasting model proposed in this paper can achieve accurate forecasting results for short-term load forecasting. The TCN model, as the most basic model, has the worst forecasting effect and takes the shortest time. From the TCN, VMD-TCN, VMD-TCN-Attention, and VMD-TCN-TPA until the EPT-VMD-TCN-TPA model, its prediction accuracy improves step by step, which also indicates that each step in the EPT-VMD-TCN-TPA model and the combination together will be beneficial for the improvement of prediction precision. Compared with the TCN model, the MAPE, RMSE, and MAE of the EPT-VMD-TCN-TPA model are reduced by 55.99%, 55.52%, and 54.61%, respectively, which indicates that the EPT-VMD-TCN-TPA model has a high predictive fine-reading. As shown in Figure 17, the EPT-VMD-TCN-TPA model has a relatively accurate prediction effect both in the wave peak and the trough. In terms of model running efficiency, the EPT-VMD-TCN-TPA model proposed in this paper takes 443.232 s, which is the longest time among all the models. This is because the decomposition of the load using EPT and VMD makes the number of prediction components increase, which leads to the long running time of the model. In real life, the accuracy of the prediction results will be given priority, and with the continuous upgrading of hardware and the application of parallel computing, the running efficiency of the EPT-VMD-TCN-TPA model proposed in this paper will not be a problem.
As shown in Table 8, the EPT-VMD-FE-TCN-TPA recombination model, which uses the sample entropy (FE) to evaluate the complexity of the IMF, has better prediction results compared to TCN and VMD-TCN. The EPT-VMD-FE-TCN-TPA recombination model has lower prediction precision compared to the EPT-VMD-TCN-TPA model, although operational efficiency is improved due to the reduced computational size. In the next step, we will consider how to use FE more rationally to evaluate the complexity of IMF so that the EPT-VMD-FE-TCN-TPA reorganization model can be run more efficiently with guaranteed accuracy.
The autocorrelation function (ACF) and partial autocorrelation function (PACF) of the final residuals of all models in Table 8 are shown in Figure 18. In Figure 18, it can be seen that the final residuals in the EPT-VMD-TCN-TPA short-term load prediction model and the EPT-VMD-FE-TCN-TPA recombination model proposed in this paper have been whitened, which indicates the validity of these two models proposed in this paper.
In order to verify the performance of the models proposed in this paper with the ablation comparison models in Table 8, we performed DM tests for each pair of models. The results of the DM tests are listed in Table 9.
In Table 9, the DM test uses MSE as a metric. The original hypothesis H in the DM test is that the column model and the row model have the same effect. The alternative hypothesis H0 in the DM test is a column model and a row model with different effects. The experimental results are p-values outside the parentheses and DM statistics inside the parentheses. If the p-value > 0.05, the effect of the row model is the same as that of the column model in the comparison; if p < 0.05 and the DM value is positive, it means that the effect of the row model is better than that of the column model with which it is compared; and if p < 0.05 and the DM value is negative, it means that the effect of the column model is better than that of the row model with which it is compared. As can be seen in the table, the experimental effect of our proposed EPT-VMD-TCN-TPA short-term load prediction model is better than other ablation comparison experiments. The experimental results of the recombinant EPT-VMD-FE-TCN-TPA short-term load prediction model proposed in this paper are statistically inferior to the EPT-VMD-TCN-TPA model proposed in this paper. However, when combined with the results of the classical experimental comparison models in DM inspection, the recombinant model can still achieve good prediction results.

4.5. Classic Experiments

To verify the accuracy of the prediction model proposed in this paper, we compared SVR, LSTM, CNN-LSTM, LSTM-Attention, BiLSTM, and the EPT-VMD-TCN-TPA prediction model in this paper. We will illustrate that our proposed prediction models are more accurate by the evaluation metrics of the prediction results. These comparative models are related as follows:
(1)
SVR prediction model: Support vector regression prediction model.
(2)
LSTM prediction model [23]: A typical recurrent neural network with memory function and gate structure, which can effectively solve the problem of gradient disappearance and gradient explosion due to excessive sequence length in RNN models.
(3)
CNN-LSTM prediction model [50]: By combining CNN and LSTM models, it enables the model to extract features inside the data through the convolution operation of the CNN, while using LSTM models to predict changes in the time series.
(4)
LSTM-Attention prediction model [22]: This model introduces the attention mechanism based on the LSTM model. The attention mechanism assigns more weight to important features, thus strengthening the connection between the whole and the local, and improving the prediction accuracy.
(5)
BiLSTM prediction model [25]: This model is improved on the basis of the LSTM model. Both forward and backward sequence information inputs are available to fully extract the information from the load data.
(6)
EPT-VMD-TCN-TPA prediction model: Our proposed final model.
(7)
The prediction effects of the EPT-VMD-TCN-TPA model and the above comparison model on the test set are shown in Figure 19 (only one day is shown), and the prediction error evaluation index data and fit effect data for each model are shown in Table 10.
As can be seen in Table 10, our proposed EPT-VMD-TCN-TPA model has the lowest MAPE, RMSE, and MAE compared to other comparative models, with 1.25%, 110.2692 MW, and 82.3180 MW, respectively, and the largest fitting index R2 of 0.9955. It is thus clear that the EPT-VMD-TCN-TPA model can predict the load accurately. Compared with the common LSTM model, the MAPE, RMSE, and MAE of the EPT-VMD-TCN-TPA model are reduced by 64.08%, 65.60%, and 62.96%, respectively, which indicates that the EPT-VMD-TCN-TPA model has a high prediction accuracy. As can be seen in Figure 19, the EPT-VMD-TCN-TPA model has the best prediction effect and the best accuracy for the wave and trough prediction results. Of course, the running time of the EPT-VMD-TCN-TPA model is still the longest, but with the upgrade of hardware devices and parallel computing, the model’s running efficiency will be improved significantly.
The autocorrelation function (ACF) and partial autocorrelation function (PACF) for the final residuals of all the models in Table 10 are shown in Figure 20. It can be seen in Figure 20 that the classical comparative model has an autocorrelation structure, indicating that it does not predict well. In the previous Figure 18i,j, it is known that the EPT-VMD-TCN-TPA model proposed in this paper has been whitened, which further illustrates the best performance of the model in this paper.
In order to verify the performance of the models proposed in this paper with the classical comparison models in Table 10, we performed DM tests for each pair of models. The results of the DM tests are listed in Table 11.
In Table 11, this DM test table (using the MSE metric) uses the original hypothesis H, which proposes that the column and row models have the same effect. The alternative hypothesis H0 proposes that the column and row models have different effects. The experimental results are p-values outside the parentheses and DM statistics inside the parentheses. If the p-value > 0.05, the effect of the row model is the same as that of the column model in the comparison; if p < 0.05 and the DM value is positive, it means that the effect of the row model is better than that of the column model with which it is compared; and if p < 0.05 and the DM value is negative, it means that the effect of the column model is better than that of the row model with which it is compared. In the table, we can see that the experimental effect of our proposed EPT-VMD-TCN-TPA short-term load prediction model is better than other classical comparison experiments, and the experimental effect of the recombinant EPT-VMD-FE-TCN-TPA short-term load prediction model proposed in this paper is also better than other classical comparison experiments, second only to the EPT-VMD-TCN-TPA model. The prediction effects of BiLSTM, LSTM, CNN-LSTM, and LSTM-Attention are not much different. The SVR has the worst effect in the comparison experiments.
From the comparison of the ablation experiments, as well as the classical experiments, it is clear that our proposed EPT-VMD-TCN-TPA model shows greater advantages in handling complex electric load forecasting data and improving forecasting accuracy. The next section will evaluate the robustness and applicability of the EPT-VMD-TCN-TPA model using the classical comparison experiments in Area 2.

5. Experimental Results in Area 2

In the previous section, we validated our proposed short-term load forecasting method based on the data from region 1 and obtained good experimental results. However, relying on only one dataset to validate the experiments may be subject to chance, so in order to verify the accuracy as well as the generalizability of our proposed model, we conducted experiments using the dataset of Area 2 provided by the 9th National Electrician Mathematical Modeling Competition in China. One of the EPT-VMD decomposition parts is described in Section 3.2. The ablation experimental result pairs of the EPT-VMD-TCN-TPA model proposed in this paper are shown in Figure 21. The error evaluation indexes and fitting effect data of each model are shown in Table 12. The classical comparison experimental results are shown in Figure 22 and the evaluation indexes of each model are shown in Table 13.
As shown in Table 12 and Table 13 and Figure 21 and Figure 22, the MAPE, RMSE, and MAE of the EPT-VMD-TCN-TPA prediction model are 1.58%, 137.6182 MW, and 107.2617 MW, respectively. The R2 reaches 0.9923. Compared with other models, our proposed EPT-VMD-TCN-TPA prediction model has the best prediction results. Therefore, it is evident that our proposed EPT-VMD-TCN-TPA model has some stability and scalability.

6. Conclusions

Accurate and effective power load forecasting is of great significance for the safe, economical, and stable operation of power grids. However, the complexity and non-smoothness of power load data bring difficulties to power load forecasting. Therefore, this paper proposes a combined decomposition (EPT-VMD) combined with TCN-TPA network forecasting for short-term power load forecasting, i.e., the EPT-VMD-TCN-TPA forecasting model. The method first extracts the trend component and residual fluctuation component of the original load series using EPT, and the trend component has obvious periodicity, which can improve the prediction precision reading. Secondly, the residual fluctuation components are decomposed into intrinsic mode function components with different center frequencies using VMD, and the ratio of the residual energy after decomposition is used to determine the number of decompositions k of VMD, which effectively avoids over-decomposition and under-decomposition. Using a network model with TCN-TPA for training and prediction of components, the TCN-TPA network can be better applied to multivariate time series prediction. Finally, the prediction results of each component are accumulated to obtain the predicted values of the final load. We also discuss the performance of the EPT-VMD-FE-TPA-TCN prediction model using FE.
To test the performance of the proposed EPT-VMD-TCN-TPA model, experiments were conducted on real datasets from two Areas provided by the 9th National Electrician Mathematical Modeling Competition in China. The predictive precision of the models was evaluated by MAE, MAPE, RMSE, and R2, and the significance of the differences in model prediction performance was assessed using the DM test. The results show that our proposed method has better prediction accuracy than other methods. The MAPE of the EPT-VMD-TCN-TPA method proposed in this paper was 1.25% and 1.58% for the test results on two regions, respectively. Although the method proposed in this paper achieved good results, only meteorological data and load data were considered in this paper. Therefore, we suggest that more influencing factors, such as the date factor, electricity price, and social development level, should be considered in future studies. In addition, suitable optimization algorithms can be used to optimally select the model hyperparameters to produce more accurate prediction results.

Author Contributions

Conceptualization, S.Z. and Q.Z.; methodology, S.Z.; software, S.Z.; validation, S.Z. and Q.Z.; formal analysis, S.Z.; data curation, S.Z. and Q.Z.; writing—original draft preparation, S.Z.; writing—review and editing, S.Z. and Q.Z.; supervision, Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fallah, S.; Deo, R.; Shojafar, M.; Conti, M.; Shamshirband, S. Computational Intelligence Approaches for Energy Load Forecasting in Smart Energy Management Grids: State of the Art, Future Challenges, and Research Directions. Energies 2018, 11, 596. [Google Scholar] [CrossRef] [Green Version]
  2. Li, J.; Luo, Y.; Wei, S. Long-Term Electricity Consumption Forecasting Method Based on System Dynamics under the Carbon-Neutral Target. Energy 2022, 244, 122572. [Google Scholar] [CrossRef]
  3. Nti, I.K.; Teimeh, M.; Nyarko-Boateng, O.; Adekoya, A.F. Electricity Load Forecasting: A Systematic Review. J. Electr. Syst. Inf. Technol. 2020, 7, 13. [Google Scholar] [CrossRef]
  4. Ahmad, T.; Zhang, H.; Yan, B. A Review on Renewable Energy and Electricity Requirement Forecasting Models for Smart Grid and Buildings. Sustain. Cities Soc. 2020, 55, 102052. [Google Scholar] [CrossRef]
  5. Mishra, M.; Nayak, J.; Naik, B.; Abraham, A. Deep Learning in Electrical Utility Industry: A Comprehensive Review of a Decade of Research. Eng. Appl. Artif. Intell. 2020, 96, 104000. [Google Scholar] [CrossRef]
  6. Kuster, C.; Rezgui, Y.; Mourshed, M. Electrical load forecasting models: A critical systematic review. Sustain. Cities Soc. 2017, 35, 257–270. [Google Scholar] [CrossRef]
  7. Tan, F.L.; Zhang, J.; Ma, H.Z. Combined forecasting method of power load based on trend change division. J. N. China Electr. Power Univ. 2020, 47, 17–24. [Google Scholar]
  8. Christiaanse, W. Short-Term Load Forecasting Using General Exponential Smoothing. IEEE Trans. Power Appar. Syst. 1971, PAS-90, 900–911. [Google Scholar] [CrossRef]
  9. Chen, P.Y.; Fang, Y.J. Short-term load forecasting of power system for holiday point-by-point growth rate based on Kalman filtering. Eng. J. Wuhan Univ. 2020, 53, 139–144. [Google Scholar]
  10. Munkhammar, J.; van der Meer, D.; Widén, J. Very Short Term Load Forecasting of Residential Electricity Consumption Using the Markov-Chain Mixture Distribution (MCM) Model. Appl. Energy 2021, 282, 116180. [Google Scholar] [CrossRef]
  11. Wu, F.; Cattani, C.; Song, W.; Zio, E. Fractional ARIMA with an Improved Cuckoo Search Optimization for the Efficient Short-Term Power Load Forecasting. Alex. Eng. J. 2020, 59, 3111–3118. [Google Scholar] [CrossRef]
  12. Duan, L.; Niu, D.; Gu, Z. Long and medium term power load forecasting with multi-level recursive regression analysis. In Proceedings of the 2008 Second International Symposium on Intelligent Information Technology Application, Shanghai, China, 20–22 December 2008; pp. 514–518. [Google Scholar]
  13. Ceperic, E.; Ceperic, V.; Baric, A. A Strategy for Short-Term Load Forecasting by Support Vector Regression Machines. IEEE Trans. Power Syst. 2013, 28, 4356–4364. [Google Scholar] [CrossRef]
  14. Hu, Y.; Li, J.; Hong, M.; Ren, J.; Lin, R.; Liu, Y.; Liu, M.; Man, Y. Short Term Electric Load Forecasting Model and Its Verification for Process Industrial Enterprises Based on Hybrid GA-PSO-BPNN Algorithm—A Case Study of Papermaking Process. Energy 2019, 170, 1215–1227. [Google Scholar] [CrossRef]
  15. Kuhba, H.; Al-Tamemi, H.A.H. Power system short-term load forecasting using artificial neural networks. Int. J. Eng. Dev. Res. 2016, 4, 78–87. [Google Scholar]
  16. Imani, M. Electrical Load-Temperature CNN for Residential Load Forecasting. Energy 2021, 227, 120480. [Google Scholar] [CrossRef]
  17. Motepe, S.; Hasan, A.N.; Twala, B.; Stopforth, R. Power Distribution Networks Load Forecasting Using Deep Belief Networks: The South African Case. In Proceedings of the 2019 IEEE Jordan International Joint Conference on Electrical Engineering and Information Technology (JEEIT), Amman, Jordan, 9–11 April 2019; pp. 507–512. [Google Scholar]
  18. Wang, C.; Huang, S.; Wang, S.; Ma, Y.; Ma, J.; Ding, J. Short term load forecasting based on vmd-dnn. In Proceedings of the 2019 IEEE 8th International Conference on Advanced Power System Automation and Protection (APAP), Online, 21–24 October 2019; pp. 1045–1048. [Google Scholar]
  19. Zhang, B.; Wu, J.-L.; Chang, P.-C. A Multiple Time Series-Based Recurrent Neural Network for Short-Term Load Forecasting. Soft Comput. 2018, 22, 4099–4112. [Google Scholar] [CrossRef]
  20. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  21. Muzaffar, S.; Afshari, A. Short-Term Load Forecasts Using LSTM Networks. Energy Procedia 2019, 158, 2922–2927. [Google Scholar] [CrossRef]
  22. Lin, J.; Ma, J.; Zhu, J.; Cui, Y. Short-Term Load Forecasting Based on LSTM Networks Considering Attention Mechanism. Int. J. Electr. Power Energy Syst. 2022, 137, 107818. [Google Scholar] [CrossRef]
  23. Kong, W.; Dong, Z.Y.; Jia, Y.; Hill, D.J.; Xu, Y.; Zhang, Y. Short-Term Residential Load Forecasting Based on LSTM Recurrent Neural Network. IEEE Trans. Smart Grid 2019, 10, 841–851. [Google Scholar] [CrossRef]
  24. Gao, X.; Li, X.; Zhao, B.; Ji, W.; Jing, X.; He, Y. Short-Term Electricity Load Forecasting Model Based on EMD-GRU with Feature Selection. Energies 2019, 12, 1140. [Google Scholar] [CrossRef] [Green Version]
  25. Yan, L.; Zhang, H. A Variant Model Based on BiLSTM for Electricity Load Prediction. In Proceedings of the 2021 IEEE International Conference on Power, Intelligent Computing and Systems (ICPICS), Shenyang, China, 29 July 2021; pp. 404–411. [Google Scholar]
  26. Niu, D.; Yu, M.; Sun, L.; Gao, T.; Wang, K. Short-Term Multi-Energy Load Forecasting for Integrated Energy Systems Based on CNN-BiGRU Optimized by Attention Mechanism. Appl. Energy 2022, 313, 118801. [Google Scholar] [CrossRef]
  27. Song, J.; Xue, G.; Pan, X.; Ma, Y.; Li, H. Hourly Heat Load Prediction Model Based on Temporal Convolutional Neural Network. IEEE Access 2020, 8, 16726–16741. [Google Scholar] [CrossRef]
  28. Wang, Y.; Chen, J.; Chen, X.; Zeng, X.; Kong, Y.; Sun, S.; Guo, Y.; Liu, Y. Short-Term Load Forecasting for Industrial Customers Based on TCN-LightGBM. IEEE Trans. Power Syst. 2021, 36, 1984–1997. [Google Scholar] [CrossRef]
  29. Tang, X.; Chen, H.; Xiang, W.; Yang, J.; Zou, M. Short-Term Load Forecasting Using Channel and Temporal Attention Based Temporal Convolutional Network. Electr. Power Syst. Res. 2022, 205, 107761. [Google Scholar] [CrossRef]
  30. Shah, I.; Jan, F.; Ali, S. Functional Data Approach for Short-Term Electricity Demand Forecasting. Math. Probl. Eng. 2022, 2022, 6709779. [Google Scholar] [CrossRef]
  31. Shah, I.; Iftikhar, H.; Ali, S. Modeling and Forecasting Electricity Demand and Prices: A Comparison of Alternative Approaches. J. Math. 2022, 2022, 3581037. [Google Scholar] [CrossRef]
  32. Shah, I.; Iftikhar, H.; Ali, S.; Wang, D. Short-Term Electricity Demand Forecasting Using ComponentsEstimation Technique. Energies 2019, 12, 2532. [Google Scholar] [CrossRef] [Green Version]
  33. Li, B.; Zhang, J.; He, Y.; Wang, Y. Short-Term Load-Forecasting Method Based on Wavelet Decomposition with Second-Order Gray Neural Network Model Combined with ADF Test. IEEE Access 2017, 5, 16324–16331. [Google Scholar] [CrossRef]
  34. Meng, Z.; Xie, Y.; Sun, J. Short-Term Load Forecasting Using Neural Attention Model Based on EMD. Electr. Eng. 2022, 104, 1857–1866. [Google Scholar] [CrossRef]
  35. Dong, P.; Bin, X.; Jun, M.; Qian, D.; Jinjin, D.; Jinjin, Z.; Qian, Z. Short-Term Load Forecasting Based on EEMD-Approximate Entropy and ELM. In Proceedings of the 2019 IEEE Sustainable Power and Energy Conference (iSPEC), Beijing, China, 4–7 November 2019; pp. 1772–1775. [Google Scholar]
  36. Zhang, Z.; Hong, W.-C.; Li, J. Electric Load Forecasting by Hybrid Self-Recurrent Support Vector Regression Model with Variational Mode Decomposition and Improved Cuckoo Search Algorithm. IEEE Access 2020, 8, 14642–14658. [Google Scholar] [CrossRef]
  37. Yuan, F.; Che, J. An Ensemble Multi-Step M-RMLSSVR Model Based on VMD and Two-Group Strategy for Day-Ahead Short-Term Load Forecasting. Knowl. Based Syst. 2022, 252, 109440. [Google Scholar] [CrossRef]
  38. Cai, C.; Li, Y.; Su, Z.; Zhu, T.; He, Y. Short-Term Electrical Load Forecasting Based on VMD and GRU-TCN Hybrid Network. Appl. Sci. 2022, 12, 6647. [Google Scholar] [CrossRef]
  39. Kim, D.; Choi, G.; Oh, H.-S. Ensemble Patch Transformation: A Flexible Framework for Decomposition and Filtering of Signal. EURASIP J. Adv. Signal Process. 2020, 2020, 30. [Google Scholar] [CrossRef]
  40. Heydari, A.; Keynia, F.; Garcia, D.A.; De Santoli, L. Mid-Term Load Power Forecasting Considering Environment Emission Using a Hybrid Intelligent Approach. In Proceedings of the 2018 5th International Symposium on Environment-Friendly Energies and Applications (EFEA), Rome, Italy, 24–26 September 2018; pp. 1–5. [Google Scholar]
  41. Jiang, P.; Nie, Y.; Wang, J.; Huang, X. Multivariable Short-Term Electricity Price Forecasting Using Artificial Intelligence and Multi-Input Multi-Output Scheme. Energy Econ. 2023, 117, 106471. [Google Scholar] [CrossRef]
  42. Sekhar, C.; Dahiya, R. Robust Framework Based on Hybrid Deep Learning Approach for Short Term Load Forecasting of Building Electricity Demand. Energy 2023, 268, 126660. [Google Scholar] [CrossRef]
  43. Heydari, A.; Majidi Nezhad, M.; Pirshayan, E.; Astiaso Garcia, D.; Keynia, F.; De Santoli, L. Short-Term Electricity Price and Load Forecasting in Isolated Power Grids Based on Composite Neural Network and Gravitational Search Optimization Algorithm. Appl. Energy 2020, 277, 115503. [Google Scholar] [CrossRef]
  44. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2013, 62, 531–544. [Google Scholar] [CrossRef]
  45. Chen, W.; Wang, Z.; Xie, H.; Yu, W. Characterization of Surface EMG Signal Based on Fuzzy Entropy. IEEE Trans. Neural Syst. Rehabil. Eng. 2007, 15, 266–272. [Google Scholar] [CrossRef]
  46. Shih, S.-Y.; Sun, F.-K.; Lee, H. Temporal Pattern Attention for Multivariate Time Series Forecasting. Mach. Learn. 2019, 108, 1421–1441. [Google Scholar] [CrossRef] [Green Version]
  47. Liu, Y.; Yang, C.; Huang, K.; Gui, W. Non-Ferrous Metals Price Forecasting Based on Variational Mode Decomposition and LSTM Network. Knowl. Based Syst. 2020, 188, 105006. [Google Scholar] [CrossRef]
  48. Tran, T.N. Grid Search of Convolutional Neural Network model in the case of load forecasting. Arch. Electr. Eng. 2021, 2022, 3581037. [Google Scholar]
  49. Shah, I.; Iftikhar, H.; Ali, S. Modeling and Forecasting Medium-Term Electricity Consumption Using Component Estimation Technique. Forecasting 2020, 2, 163–179. [Google Scholar] [CrossRef]
  50. Rafi, S.H.; Nahid-Al-Masood; Deeba, S.R.; Hossain, E. A Short-Term Load Forecasting Method Using Integrated CNN and LSTM Network. IEEE Access 2021, 9, 32436–32448. [Google Scholar] [CrossRef]
Figure 1. Classification chart of load forecasting methods.
Figure 1. Classification chart of load forecasting methods.
Applsci 13 04462 g001
Figure 2. The general flow of power load forecasting.
Figure 2. The general flow of power load forecasting.
Applsci 13 04462 g002
Figure 3. VMD calculation process.
Figure 3. VMD calculation process.
Applsci 13 04462 g003
Figure 4. Dilated causal convolution.
Figure 4. Dilated causal convolution.
Applsci 13 04462 g004
Figure 5. Residual blocks.
Figure 5. Residual blocks.
Applsci 13 04462 g005
Figure 6. The TPA mechanism.
Figure 6. The TPA mechanism.
Applsci 13 04462 g006
Figure 7. The structural framework of the EPT-VMD-TCN-TPA model.
Figure 7. The structural framework of the EPT-VMD-TCN-TPA model.
Applsci 13 04462 g007
Figure 8. Annual load of Area 1.
Figure 8. Annual load of Area 1.
Applsci 13 04462 g008
Figure 9. Load trend chart for four consecutive weeks for Area 1.
Figure 9. Load trend chart for four consecutive weeks for Area 1.
Applsci 13 04462 g009
Figure 10. Area 1 EPT decomposition results.
Figure 10. Area 1 EPT decomposition results.
Applsci 13 04462 g010
Figure 11. Area 2 EPT decomposition results.
Figure 11. Area 2 EPT decomposition results.
Applsci 13 04462 g011
Figure 12. Area 1 VMD results.
Figure 12. Area 1 VMD results.
Applsci 13 04462 g012
Figure 13. Area 2 VMD results.
Figure 13. Area 2 VMD results.
Applsci 13 04462 g013
Figure 14. Fuzzy entropy of each IMF.
Figure 14. Fuzzy entropy of each IMF.
Applsci 13 04462 g014
Figure 15. FE reorganization results for Area 1.
Figure 15. FE reorganization results for Area 1.
Applsci 13 04462 g015
Figure 16. A flowchart of EPT-VMD-TCN-TPA model.
Figure 16. A flowchart of EPT-VMD-TCN-TPA model.
Applsci 13 04462 g016
Figure 17. Area 1 ablation experiment prediction results.
Figure 17. Area 1 ablation experiment prediction results.
Applsci 13 04462 g017
Figure 18. ACF and PACF plots of TCN (a,b), VMD-TCN (c,d), VMD-TCN-Attention (e,f), VMD-TCN-TPA (g,h), EPT-VMD-TCN-TPA (i,j), EPT-VMD-FE-TCN-TPA (k,l).
Figure 18. ACF and PACF plots of TCN (a,b), VMD-TCN (c,d), VMD-TCN-Attention (e,f), VMD-TCN-TPA (g,h), EPT-VMD-TCN-TPA (i,j), EPT-VMD-FE-TCN-TPA (k,l).
Applsci 13 04462 g018aApplsci 13 04462 g018bApplsci 13 04462 g018c
Figure 19. Area 1 classical model prediction results.
Figure 19. Area 1 classical model prediction results.
Applsci 13 04462 g019
Figure 20. ACF and PACF plots of SVR (a,b), LSTM (c,d), CNN-LSTM (e,f), LSTM-Attention (g,h), BiLSTM (i,j).
Figure 20. ACF and PACF plots of SVR (a,b), LSTM (c,d), CNN-LSTM (e,f), LSTM-Attention (g,h), BiLSTM (i,j).
Applsci 13 04462 g020aApplsci 13 04462 g020b
Figure 21. Area 2 ablation experiment prediction results.
Figure 21. Area 2 ablation experiment prediction results.
Applsci 13 04462 g021
Figure 22. Area 2 prediction results for each model.
Figure 22. Area 2 prediction results for each model.
Applsci 13 04462 g022
Table 1. Dataset load details.
Table 1. Dataset load details.
DatasetSamplesRangeNumbersMean (MW)Max (MW)Min (MW)Std. (MW)
Area 1All samples1 January 2012–10 January 2015106,1766915.3312,296.851306.082094.08
Training1 January 2012–2 June 201484,8646657.8111,446.851306.082034.47
Testing3 June 2014–10 January 201521,3127945.5312,296.852267.672010.75
Area 2All samples1 January 2012–10 January 2015106,1767357.4313,536.741986.322180.43
Training1 January 2012–2 June 201484,8647038.4412,466.101986.322054.60
Testing3 June 2014–10 January 201521,3128627.6713,536.743259.122204.07
Table 2. Data of meteorological factors from 1 August to 7 August 2012 (Area 1).
Table 2. Data of meteorological factors from 1 August to 7 August 2012 (Area 1).
DateMax. Temperature (°C)Min. Temperature (°C)Avg. Temperature (°C)Relative Humidity (avg.)Rainfall (mm)
1 August 201236.023.130.371.026.5
2 August 201235.627.831.961.00.0
3 August 201233.928.030.963.00.0
4 August 201231.327.829.575.00.0
5 August 201231.026.228.582.01.8
6 August 201230.824.726.987.012.3
7 August 201234.525.829.479.00.1
Table 3. Rres-values corresponding to different k values.
Table 3. Rres-values corresponding to different k values.
kRres-Value of Area 1Rres-Value of Area 2
20.015883810.01149027
30.010771310.01032953
40.005848490.00780275
50.005024760.00628756
60.004126820.00523697
70.003869090.00402301
80.003704310.00331197
90.003609390.00285365
100.003562380.00239487
110.003380150.00198062
120.003359770.00197231
130.003342510.00197134
140.003336110.00197128
150.003327950.00197123
Table 4. FE-IMF composition.
Table 4. FE-IMF composition.
Recombinant SequencesFE-IMF1FE-IMF2FE-IMF3FE-IMF4FE-IMF5FE-IMF6
ComponentsIMF1IMF2IMF3 + IMF10IMF4 + IMF7IMF5 + IMF6IMF8 + IMF9 + IMF11
Table 5. Experimental environment parameters.
Table 5. Experimental environment parameters.
ProcessorMemoryPythonTensorFlowKeras
Intel(R) Core (TM) i5-7200U CPU @ 2.50 GHz 2.70 GHz4 G3.82.11.02.11.0
Table 6. Model parameters (1).
Table 6. Model parameters (1).
ItemsNumber of Residual BlocksConvolution Kernel SizeNumber of Convolution KernelsDilation RateNumber of IterationsDropout
Value3320(1,2,4)1000.1
Table 7. Model parameters (2).
Table 7. Model parameters (2).
ItemsLearning RateLoss FunctionOptimizerActivation FunctionsBatch-Size
Value0.001MSEAdamReLU64
Table 8. Area 1 ablation model evaluation indicators.
Table 8. Area 1 ablation model evaluation indicators.
ModelMAPE (%)RMSE (MW)MAE (MW)R2Time (s)
TCN2.84247.9006181.35140.977434.040
VMD-TCN2.59187.9851147.27610.9870390.934
VMD-TCN-Attention1.58133.4030102.74100.9934383.140
VMD-TCN-TPA1.54122.108998.62450.9945427.891
EPT-VMD-TCN-TPA1.25110.269282.31800.9955443.232
EPT-VMD-FE-TCN-TPA1.86168.9959121.69940.9895296.137
Table 9. DM test value table.
Table 9. DM test value table.
ModelTCNVMD-TCNVMD-TCN-AttentionVMD-TCN-TPAEPT-VMD-TCN-TPAEPT-VMD-FE-TCN-TPA
TCNnan0.0112
(−2.538)
0.0007
(−3.386)
0.0003
(−3.590)
0.0002
(−3.765)
0.0110
(−2.543)
VMD-TCN0.0112
(2.538)
nan0.0877
(−1.708)
0.0472
(−1.985)
0.0260
(−2.227)
0.5255
(−0.635)
VMD-TCN-Attention0.0007
(3.386)
0.0877
(1.708)
nan0.0083
(−2.642)
0.0000
(−8.145)
0.0106
(2.555)
VMD-TCN-TPA0.0003
(3.590)
0.0472
(1.985)
0.0083
(2.642)
nan0.0066
(−2.716)
0.0021
(3.073)
EPT-VMD-TCN-TPA0.0002
(3.765)
0.0260
(2.227)
0.0000
(8.145)
0.0066
(2.716)
nan0.0001
(4.018)
EPT-VMD-FE-TCN-TPA0.0110
(2.543)
0.5255
(0.635)
0.0106
(−2.555)
0.0021
(−3.073)
0.0001
(−4.018)
nan
Table 10. Classical model evaluation indicators for Area 1.
Table 10. Classical model evaluation indicators for Area 1.
ModelMAPE (%)RMSE (MW)MAE (MW)R2Time (s)
SVR16.071309.01931022.14990.368911.666
LSTM3.48320.5932222.24770.962139.803
CNN-LSTM3.27291.7813203.86830.968645.583
LSTM-Attention3.82379.4524229.15540.947057.898
BiLSTM3.15273.6259200.78600.972493.655
EPT-VMD-TCN-TPA1.25110.269282.31800.9955443.232
Table 11. DM test value table.
Table 11. DM test value table.
ModelSVRLSTMCNN-LSTMLSTM-AttentionBiLSTMEPT-VMD-TCN-TPAEPT-VMD-FE-TCN-TPA
SVRnan0.0000
(−5.779)
0.0000
(−5.828)
0.0000
(−6.444)
0.0000
(−5.502)
0.0000
(−5.649)
0.0000
(−5.603)
LSTM0.0000
(5.779)
nan0.0010
(−3.289)
0.3682
(0.899)
0.2256
(−1.212)
0.0018
(−3.127)
0.0082
(−2.644)
CNN-LSTM0.0000
(5.828)
0.0010
(3.289)
nan0.2191
(1.229)
0.6427
(−0.464)
0.0058
(−2.759)
0.0274
(−2.205)
LSTM-Attention0.0000
(6.444)
0.3682
(−0.899)
0.2191
(−1.229)
nan0.2869
(−1.065)
0.0087
(−2.821)
0.0179
(−2.608)
BiLSTM0.0000
(5.502)
0.2256
(1.212)
0.6427
(0.464)
0.2869
(1.065)
nan0.0000
(−4.811)
0.0003
(−3.587)
EPT-VMD-TCN-TPA0.0000
(5.649)
0.0018
(3.127)
0.0058
(2.758)
0.0087
(2.821)
0.0000
(4.811)
nan0.0001
(4.018)
EPT-VMD-FE-TCN-TPA0.0000
(5.604)
0.0082
(2.644)
0.0275
(2.205)
0.0179
(2.608)
0.0003
(3.587)
0.0001
(−4.018)
nan
Table 12. Area 2 ablation model evaluation indicators.
Table 12. Area 2 ablation model evaluation indicators.
ModelMAPE (%)RMSE (MW)MAE (MW)R2Time (s)
TCN3.02279.9281207.60850.968131.759
VMD-TCN1.68147.2106116.86440.9912356.854
VMD-TCN-Attention1.63144.1170110.62460.9915379.672
VMD-TCN-TPA1.60140.2134109.34520.9919411.754
EPT-VMD-TCN-TPA1.58137.6182107.26170.9923424.006
EPT-VMD-FE-TCN-TPA2.02170.3671135.11090.9882243.455
Table 13. Area 2 evaluation indicators for each model.
Table 13. Area 2 evaluation indicators for each model.
ModelMAPE (%)RMSE (MW)MAE (MW)R2Time (s)
SVR13.111103.4697882.27210.50456.910
LSTM3.64355.7457249.56780.948532.688
CNN-LSTM3.26294.3930222.81880.964779.906
LSTM-Attention3.85383.7617265.79520.940173.906
BiLSTM2.64254.9352178.60090.9736131.753
EPT-VMD-TCN-TPA1.58137.6182107.26170.9923424.006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zan, S.; Zhang, Q. Short-Term Power Load Forecasting Based on an EPT-VMD-TCN-TPA Model. Appl. Sci. 2023, 13, 4462. https://doi.org/10.3390/app13074462

AMA Style

Zan S, Zhang Q. Short-Term Power Load Forecasting Based on an EPT-VMD-TCN-TPA Model. Applied Sciences. 2023; 13(7):4462. https://doi.org/10.3390/app13074462

Chicago/Turabian Style

Zan, Shifa, and Qiang Zhang. 2023. "Short-Term Power Load Forecasting Based on an EPT-VMD-TCN-TPA Model" Applied Sciences 13, no. 7: 4462. https://doi.org/10.3390/app13074462

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop