Next Article in Journal
Advanced Technologies for Green Hydrogen Production
Next Article in Special Issue
Ensemble Machine Learning Approaches for Prediction of Türkiye’s Energy Demand
Previous Article in Journal
Development of Manganese-Coated Graphite Electrode in a Dual-Chambered Fuel Cell for Selenite Removal and Bio-Electricity Generation from Wastewater Effluent by Bacillus cereus
Previous Article in Special Issue
Agent-Based Simulation and Micro Supply Chain of the Food–Energy–Water Nexus for Collaborating Urban Farms and the Incorporation of a Community Microgrid Based on Renewable Energy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Short-Term Load Forecasting Model Based on Crisscross Grey Wolf Optimizer and Dual-Stage Attention Mechanism

1
School of Electrical Engineering, Guangxi University, Nanning 530004, China
2
School of Traffic &Transportation, Nanning University, Nanning 530200, China
*
Author to whom correspondence should be addressed.
Energies 2023, 16(6), 2878; https://doi.org/10.3390/en16062878
Submission received: 9 February 2023 / Revised: 15 March 2023 / Accepted: 17 March 2023 / Published: 21 March 2023
(This article belongs to the Special Issue Advanced Machine Learning Applications in Modern Energy Systems)

Abstract

:
Accurate short-term load forecasting is of great significance to the safe and stable operation of power systems and the development of the power market. Most existing studies apply deep learning models to make predictions considering only one feature or temporal relationship in load time series. Therefore, to obtain an accurate and reliable prediction result, a hybrid prediction model combining a dual-stage attention mechanism (DA), crisscross grey wolf optimizer (CS-GWO) and bidirectional gated recurrent unit (BiGRU) is proposed in this paper. DA is introduced on the input side of the model to improve the sensitivity of the model to key features and information at key time points simultaneously. CS-GWO is formed by combining the horizontal and vertical crossover operators, to enhance the global search ability and the diversity of the population of GWO. Meanwhile, BiGRU is optimized by CS-GWO to accelerate the convergence of the model. Finally, a collected load dataset, four evaluation metrics and parametric and non-parametric testing manners are used to evaluate the proposed CS-GWO-DA-BiGRU short-term load prediction model. The experimental results show that the RMSE, MAE and SMAPE are reduced respectively by 3.86%, 1.37% and 0.30% of those of the second-best performing CSO-DA-BiGRU model, which demonstrates that the proposed model can better fit the load data and achieve better prediction results.

1. Introduction

Electric load forecasting plays an important role in the modernization of power system management and has become the research focus of current power enterprises [1]. It can be divided into long-term, medium-term and short-term forecasting according to its different purposes [2]. Among them, short-term load forecasting can ensure the safe and stable operation of the power system and improve social benefits [3]. Furthermore, it can accelerate the development of the power market and improve economic benefits [4]. Therefore, it is of great significance to design an efficient and accurate short-term load forecasting method.
The early short-term load methods mainly use are the exponential smoothing method [5] and hidden Markov model [6], but the ability of these methods to extract nonlinear characteristics of load is weak [7]. With the rapid increase in the installation of smart meters [8] and the development of artificial intelligence technology [9], short-term load forecasting based on big data analysis has become a current research hotspot, such as the BP neural network [10], extreme learning machine [11], support vector machine [12], etc. In addition, in order to avoid falling into local minima, some scholars use swarm intelligence optimization algorithms to optimize the artificial intelligence model. For example, Niu and Dai [13] proposed a short-term load forecasting model based on modified particle swarm optimization, in which the parameters of least squares supporting a vector machine are optimized. The experimental results show that the regression accuracy and generalization ability of the model have been improved by the proposed algorithm. To address the problem of the ease in which long short-term memory neural networks fall into local minima, a whale optimization algorithm (WOA) is used to optimize the network [14]. Li et al. [15] use grey wolf optimization (GWO) to optimize the parameters of every single kernel in an extreme learning machine to improve its forecasting ability. However, GWO quickly falls into the optima trap or fails to find the global optimal solution [16].
Moreover, the above-mentioned shallow learning method is applicable in the scenario where only the historical load is used for forecasting. If it is necessary to extract deep hidden features in massive load data, a deep learning model is needed. For example, Khan et al. [17] use a convolution neural network to extract the coupling relationship of the input features. Muzaffar and Afshari [18] use a long short-term memory (LSTM) neural network to learn the temporal correlation contained in the load time series data. In addition, a gated recurrent unit (GRU) with simpler structures is also used for short-term load prediction [19], with a high-efficiency ability of feature extraction. To extract the implicit coupling relationship between features and temporal dependency in load time series, the combination of CNN and LSTM [20], or its improved variants (e.g., CNN-GRU [21], GRU-TCN [22] and RCNN-ML-LSTM [23]) are used in load prediction. However, the above deep learning models have the problems of gradient disappearance and gradient explosion [24]. Therefore, it is of great significance to avoid these problems and accelerate the convergence speed of the model, so as to improve the accuracy of load forecasting.
The above-mentioned short-term load forecasting models based on artificial intelligence techniques do not consider the importance of input features, making important features disappear with the increase of step size [25]. Feature selection is the commonly used technique to select out the most appropriate input features in forecasting problems [26]. Kong et al. [27] utilize principal component analysis to determine the major factors affecting wind speed, reducing the dimension of relative features and improving the generation of model. Li et al. [28] develop a feature selection method to choose competitive input features. The above feature selection methods reduce the number of input features only once before prediction by simple correlation analysis, which causes the potentially important input variable be discarded [29]. To address this problem, an attention mechanism is proposed [30], with the advantage of making the model handle the dependency of long time series more easily. For example, Wang et al. [31] proposed a short-term load forecasting model based on a feature attention mechanism (FA), in which the effective characteristics of the input variables are highlighted, leading to improved prediction accuracy. In Ref. [32], a temporal attention mechanism (TA) is applied in short-term load prediction to capture the high-impact time steps of load sequence, so as to further reduce the prediction errors. However, there are few research studies that focus on combining FA with TA to propose a multi-stage attention mechanism, capturing the feature and temporal relationship in load time series.
In view of the shortcomings of existing forecasting models, this paper proposes a short-term load forecasting model (DA-CS-GWO-BiGRU) based on a dual-stage attention mechanism (DA) and crisscross grey wolf optimizer algorithm (CS-GWO). The contributions of this paper are presented as follows:
  • Combining the advantages of a feature and temporal attention mechanism, a dual-stage attention mechanism (DA) is introduced in this paper. DA is utilized at the input side of the forecasting model to comprehensively capture the correlation relationship between various variables and temporal dependency in the load time series.
  • To address the deficiency of GWO, a novel crisscross grey wolf optimizer algorithm is firstly applied in a short-term load forecasting problem. By introducing horizontal and vertical crossover operators, the global search ability and community diversity of CS-GWO are improved.
  • The proposed DA-CS-GWO-BiGRU model is verified by using the real load data set collected in a certain area. The experimental results show that the proposed model has higher forecasting accuracy than other comparison models, and has good application prospects.
The organization of the remainder of this paper is as follows. Section 2 introduces the basic principle of the deep learning model involved in this paper. Section 3 presents the methodology of the proposed DA-CS-GWO-BiGRU short-term load forecasting model. Section 4 introduces the metrics of evaluating predictions. Section 5 focuses on the details of the experiments, and the results are analyzed and discussed. Section 6 points out the limitations of and indicates the subsequent work that could follow this paper. Finally, Section 7 summarizes this paper.

2. Principle of Deep Learning Model

2.1. BiGRU Neural Network

A recurrent neural network (RNN) can only remember short-term dependencies of a time series and is often accompanied by the problem of gradient explosion or disappearance in the training process [33], leading to its limited use in practice. By modifying the calculation method of the hidden state of RNN, GRU and LSTM can effectively strengthen the long-term dependence of time series [19]. The network structure of GRU is shown in Figure 1.
It can be seen from Figure 1 that the GRU network calculates the combination degree of the current input and the previous status information by the reset gate rt. The calculation process is shown in Equation (1).
r t = σ x t W r + h t 1 U r + b r
where  x t  is the input data in t-th time step,  h t 1  is the output of the previous time step,  W r  and  U r  are the weight metrices of the reset gate,  b r  is the bias of the reset gate and  σ  is the sigmoid activation function.
In addition, the GRU network controls the retention of the previous state information  h t 1  in the current state by update gate zt, and its calculation process is shown in Equation (2).
z t = σ x t W z + h t 1 U z + b z
where  W z  and  U z  are the weight metrices of the update gate and  b z  is the bias of the update gate.
Next, GRU obtains candidate hidden states through the reset gate based on the updating mechanism of RNN [34], as shown in Equation (3).
h ˜ t = tanh x t W h + r t h t 1 U h + b h
where  W h  and  U h  are the weight metrices of candidate output and  b h  is the bias of the candidate output.
Finally, GRU obtains new hidden state  h t  by considering the previous hidden state  h t 1  and candidate hidden state  h ˜ t  as well as zt. The calculation process is shown in Equation (4).
h t = z t h t 1 + 1 z t h ˜ t 1
It can be seen from Equations (1)–(4) that GRU only considers the influence factors on the current time, and lacks consideration of future influence factors [35]. Meanwhile, a recently proposed neural network named bidirectional gated recurrent unit (BiGRU) can effectively make up for the deficiency of GRU. BiGRU can fully excavate the influence relationship hidden in the time series before and after through its unique forward and backward propagation network structure [36]. The network structure of BiGRU is shown in Figure 2, and its calculation process is shown in Equations (5)–(7).
h t = G ( x t , h t 1 )
h t = G ( x t , h t 1 )
h t = w t c t + w t c t + b t
where  h t  is the state information of forward propagation,  h t  is the state information of backward propagation,  w t  is the weight metrices of the hidden layer in forward propagation,  w t  is the weight metrices of the hidden layer in backward propagation,  b t  is the bias of the hidden layer and  G ( )  is the calculation process of GRU as shown in Equations (1)–(4).
As can be seen from Figure 2, BiGRU can exploit information both from the past and the future. Therefore, this paper uses BiGRU for timing analysis.

2.2. Attention Mechanism

By simulating human visual behavior, the attention mechanism adaptively assigns different attention weights to the input features of the model to highlight the more critical influence factors [37], helping the model predict better.
The attention mechanism is mainly composed of three parts, namely, attention weight calculation, weight normalization and intermediate semantic vector calculation. Firstly, the attention weight e of different features in the model input x or at t-th time step is calculated by using a multi-layer perceptron or neural network. Then, in order to meet the requirement that the sum of attention weights is 1, the attention weight e is normalized to find  α . Finally, the intermediate semantic vector can be obtained by considering x and  α  as shown in Equation (8).
c = α x
Therefore, this paper expects to use the attention mechanism to capture the coupling relationship between each feature and the impact of information both from the past and the future on the forecasted load value.

3. DA-CS-GWO-BiGRU Short-Term Load Forecasting Model

3.1. Mathematical Model

Set  L ( i ) = L ( 24 i 23 ) , , L ( 24 i )  as the electric load time series of the previous day,  t ( i ) = t max ( i ) , t max ( i + 1 ) , t min ( i ) , t min ( i + 1 )  as the highest and lowest temperature of the previous day and the current day,  r ( i ) = r ( i ) , r ( i + 1 )  as the rainfall of the previous day and the current day and  d ( i ) = d ( i ) , d ( i + 1 )  as the weather day type of the previous day and the current day. The short-term load forecasting problem can be regarded as using the load information of the previous day  L ( i ) , and combining its relevant characteristics  t ( i ) r ( i )  and  d ( i )  to make a prediction of the electric load values in the current day. Let the function map of the model be  F θ , and the prediction process is shown in Equation (9).
Y ^ ( i ) = F θ ( X ( i ) )
where  X ( i ) = L ( i ) , t ( i ) , r ( i ) , d ( i ) = x 1 ( i ) , x 2 ( i ) , , x T ( i ) T , and  Y ^ ( i )  represents the predicted load values of the current day.

3.2. Dual-Stage Attention Mechanism

As shown in Section 3.1, the prediction model takes historical load time series  L ( i ) , temperature  t ( i ) , rainfall  r ( i )  and weather day type  d ( i )  as inputs. According to [38], different time steps and different features of the same time step have unequal effects on the output. In order to simultaneously enhance the sensitivity of features and the temporal dimension while making short-term load predictions, this paper proposes a novel attention mechanism named dual-stage attention mechanism (DA) that combines a feature attention (FA) and a temporal attention mechanism (TA). Combining the advantages of TA and FA, DA can fully capture the relationship between variables and temporal dependence in load time series, so as to provide data support for an efficient forecasting model.

3.2.1. Feature Attention Mechanism

To highlight the more critical influence features of the input, FA is introduced. As the successor of the attention mechanism, FA also has three parts, and the calculation of the attention weight of the input features is realized through the neural network in this paper. The realization of FA is demonstrated in Figure 3 and as below.
(1) Attention weight calculation: Set  X = x 1 , x 2 , , x T N × T  as the input vector of the prediction model, where  x j = x j ( i ) i = 1 N N ( j 1 , 2 , , n f )  and N is the number of samples. The quantization process of the corresponding weight of each feature is shown in Equation (10).
e = σ ( X W e + b e )
Among them,  e = e 1 , e 2 , , e T  is the unnormalized attention weight,  W e T × T  is the trainable coefficient matrix and  b e T  is the bias.
(2) Weight normalization: In order to make the attention weight satisfy the probability distribution whose sum is 1,  e  is normalized by the softmax function, as shown in Equation (11).
α j = softmax ( e j ) = exp ( e j ) / i = 1 T exp ( e i )
(3) Intermediate semantic vector calculation: The normalized weight  α j  is multiplied by the corresponding feature vector  x j  to achieve the purpose of enhancing or reducing the expression of  x j . Finally, the adaptively optimized feature vector  X ATT  can be obtained as shown in Equation (12).
X ATT = α 1 x 1 , α 2 x 2 , , α T x T
It is worth noting that the attention weights are dynamically changed during the training process of the model, and the weights are determined only when the iterations converge.

3.2.2. Temporal Attention Mechanism

In order to capture the temporal correlation relationship between each time step in  X ATT  and the current prediction results, TA is introduced. The adaptive extraction of features at important moments is realized by integrating TA and the BiGRU network. The implementation is shown in Figure 4 and as below.
TA also has three parts, which are the same as FA.
(1) Attention weight calculation: Taking the vector  X ATT  containing the feature association relationship and the hidden state  h t 1  at the previous time step of BiGRU as the input of TA, the attention weights at t-th time step in the iterative process are quantified, as shown in Equation (13).
f t = σ ( X ATT ; h t 1 W f + b f )
where  W f ( 2 T ) × n p  is the trainable coefficient matrix,  b f n p  is the bias and  n p  is the number of hidden elements in the last layer of BiGRU.
(2) Weight normalization: In addition,  f t  is normalized using the softmax function, as shown in Equation (14).
β t j = exp ( f t j ) / i = 1 T exp ( f t i )
(3) Intermediate semantic vector calculation: In order to obtain the implicit temporal correlation relationship at t-th time step,  β t j  and  α t x t  are weighted and summed to obtain an intermediate semantic vector  d t , which contains features and temporal-related information, as shown in Equation (15).
d t = j = 1 T β t j α j x j
Once the iteration of BiGRU terminates, the final hidden state  h T  and the intermediate semantic vector  d T  are obtained. The final prediction is taken by a single-layer feedforward network utilizing  h T  and  d T , as shown in Equation (16).
Y = h T ; d T W y + b y
where  W y ( 2 T ) × n p  is the weight metrices of the feedforward network and  b y n p  is the bias.
Assuming that all parameters of the DA-BiGRU model are  θ , its loss function is shown in Equation (17). With the goal of minimizing this loss function, the final DA-BiGRU prediction model is obtained when the training is over.
J ( Y , Y , θ ) = 1 N i = 1 N ( Y ( i ) , Y ( i ) ) 2

3.3. CS-GWO Optimization Algorithm

Facts have proved that the prediction model trained with an Adam optimizer can achieve rapid convergence in the early stage, but the learning rate is too low in the later stage of training, which may affect the effective convergence of the model and cause generalization problems [39]. Swarm intelligence optimization algorithms (e.g., PSO, GWO) are used to address the aforementioned problems; however, this leads to another problem of failing to find global optimal solutions [16]. In Ref. [40], an improved GWO algorithm created by incorporating a crisscross optimization algorithm (CSO) [41] is proposed for solving the optimal power flow problem, effectively avoiding falling into the local optimum and preventing the premature convergence. Inspired by this, this paper applies a crisscross grey wolf optimizer (CS-GWO) to optimize the DA-BiGRU model in the early stage of training, so as to further accelerate the convergence and improve the generation of the short-term load prediction model. Compared with GWO, CS-GWO achieves better global search ability and group diversity by introducing horizontal crossover and vertical crossover operators of CSO.
The implementation of CS-GWO is mainly composed of five parts, which are parameter initialization, hunting, attacking prey, horizontal crossover and vertical crossover. The implementation process is described in detail as follows.

3.3.1. Parameter Initialization

Set the population of grey wolf as  Φ = θ 1 , θ 2 , , θ n M T n M × D , where  n M  is the population size and D is the population dimension. Select the individual with the best fitness value in population  Φ  as grey wolf  α , the individual with the second-best fitness value in population  Φ  as grey wolf  β , the individual with the third-best fitness value in population  Φ  as grey wolf  δ , and the rest of the population  Φ  as grey wolf  ω .

3.3.2. Hunting

Since the optimal hunting position of the wolves is unknown in the abstract search space, it is necessary to set three wolves with the strongest hunting ability to guide the hunting of the wolves. Assume that the three wolves are  α β  and  ω , respectively, and the other wolf  ω  updates its position according to the positions of the above three wolves during the iteration process, as shown in Equations (18) and (19).
θ α = θ α ( t ) A 1 C 1 θ α ( t ) θ ω ( t ) θ β = θ β ( t ) A 2 C 2 θ β ( t ) θ ω ( t ) θ δ = θ δ ( t ) A 3 C 3 θ δ ( t ) θ ω ( t )
θ ω ( t + 1 ) = θ α + θ β + θ δ 3
where  θ α ( t ) θ β ( t ) θ δ ( t )  and  θ ω ( t )  are the positions of grey wolves  α β δ  and  ω  at t-th iteration, respectively, and  A 1 A 2 A 3 C 1 C 2  and  C 3  are synergy coefficients, which are calculated as Equations (20) and (21).

3.3.3. Attack Prey

Once the prey is at rest, grey wolves stop searching and attack the prey. To simulate this process, GWO designs a synergy coefficient A, as shown in Equation (20).
A = 2 a r 1 a
where  r 1  is a random number in the range of [0, 1], and a linearly decreases from 2 to 0 during the entire iteration process.
It can be seen from Equation (20) that GWO simulates the attack process of grey wolves. When  A < 1 , the wolves attack the prey; when  A > 1 , the wolves leave the prey alone, hoping to find better prey.
In addition, in order to avoid local optimum, GWO also designs another synergy coefficient C, as shown in Equation (21).
C = 2 r 2
where  r 2  is a random number in the range of [0, 1].

3.3.4. Horizontal Crossover

In order to improve the global search ability of GWO, horizontal crossover (HC) is used to perform arithmetic crossover operations between two different individuals in all dimensions. Assuming that the i-th parent  θ i  and the j-th parent  θ j i , j 1 , 2 , , n M  perform HC operations on the d-th dimension, respectively, their offspring can be expressed as:
θ i , d H C = r 3 × θ i , d + 1 r 3 × θ j , d + C 4 × θ i , d θ j , d θ j , d H C = r 4 × θ j , d + 1 r 4 × θ i , d + C 5 × θ j , d θ i , d
where r3 and r4 are uniformly distributed random values in the range of [0, 1], and C4 and C5 are uniformly distributed random values in the range of [−1, 1]. Once the HC is over, the new population  Φ H C = θ 1 H C , θ 2 H C , , θ n M H C T n M × D  can be obtained.

3.3.5. Vertical Crossover

In order to improve the population diversity of GWO, vertical crossover (VC) is used to perform arithmetic crossover operations for all individuals between two different dimensions to generate offspring. Assuming that the  d 1 -th dimension and  d 2 -th dimension of the individual perform VC operations, the offspring can be expressed as:
θ i , d 1 V C = r × θ i , d 1 + 1 r × θ i , d 2
where r is a random value uniformly distributed in the range of [0, 1]. Once the VC is over, a new population  Φ V C = θ 1 V C , θ 2 V C , , θ n M V C T n M × D  is obtained.

3.3.6. The Detailed Implementation Steps of CS-GWO

To solve the problems of gradient disappearance and gradient explosion problems in deep neural networks [42], the CS-GWO is used to optimize weights and bias  θ  of the DA-GRU model in the early stage of training, aiming to improve the generalization performance of the model. The flow chart of CS-GWO is shown in Figure 5, and the detailed implementation steps of the CS-GWO algorithm are as follows:
(1) Initialize parameters: Set the number of grey wolf populations  n M , the maximum number of iterations T and the population dimension D (the number of weights and bias of the DA-BiGRU model), and initialize the population  Φ .
(2) Set the fitness function: Take Equation (17) as the fitness function.
(3) Determine wolves   α β  and  δ : Use Equation (17) to calculate the fitness of each individual; the individual with the best fitness value is grey wolf  α , the second-best fitness value is grey wolf  β  and the third-best fitness value is grey wolf  ω .
(4) Update position and synergy coefficient: Firstly, update the position of grey wolf  ω  according to Equations (18) and (19), and then update A and C according to Equations (20) and (21).
(5) Horizontal crossover: According to Equation (22), perform horizontal crossover on the parent population  Φ  to obtain the offspring population  Φ H C , and use Equation (17) to calculate the fitness of each individual. If the fitness of individual  θ k k 1 , 2 , , n M  in  Φ  is worse than that of individual  θ k H C  in  Φ H C , replace  θ k  with  θ k H C ; otherwise, do not replace.
(6) Vertical crossover: According to Equation (23), carry out horizontal crossover on the parent population  Φ  to obtain the offspring population  Φ V C , and use Equation (17) to calculate the fitness of each individual. If the fitness of individual  θ k  in  Φ  is worse than that of individual  θ k V C  in  Φ V C , replace  θ k  with  θ k V C ; otherwise, do not replace.
(7) Iteration termination: If the number of iterations reaches T, the position of grey wolf  α  is used as the initial weight and threshold  θ α  of the DA-GRU model; otherwise, return to step (3) and continue the iteration.

4. Evaluation Index

In order to evaluate the effectiveness of the proposed prediction model, this paper uses root mean square error (RMSE), mean absolute error (MAE), symmetric mean absolute percentage error (SMAPE) and decision coefficient (R2) to evaluate the prediction results. The definitions of the four evaluation indicators are shown in Equations (24)–(27), where RMSE, MAE and SMAPE are indicators used to describe the error between the predicted value and the real value. The smaller the value, the more accurate the prediction result. Furthermore, R2 is the indicator used to assess the linear relationship between the input and output values. The larger the value, the higher the prediction accuracy.
RMSE = 1 n t e s t n t e s t Y t e s t Y ^ t e s t 2
MAE = 1 n t e s t n t e s t Y t e s t Y ^ t e s t
SMAPE = 1 n t e s t n t e s t Y t e s t Y ^ t e s t Y t e s t + Y ^ t e s t / 2
R 2 = 1 n t e s t Y t e s t + Y ^ t e s t 2 n t e s t Y t e s t Y ¯ t e s t 2
where  Y t e s t  and  Y ^ t e s t  are the actual and predicted load values in the testing dataset, respectively,  Y ¯ t e s t  is the average value of the actual load value and  n t e s t  is the sample number of the testing dataset.

5. Experiment and Analysis

This section will verify the effectiveness of the proposed DA-CS-GWO-BiGRU short-term load forecasting model through four evaluation indicators (RMSE, MAE, SMAPE and R2) and two experiments. In addition, in order to reduce the errors caused by the experimental operation, both experiments were performed 20 times, and the average value was taken as the final experimental result. Both experiments are based on Python 3.8 and the Keras deep learning library. The core configuration of the used computer is Intel (R) Core (TM) i5-9600K 6-core processor, 3.70 GHz operating frequency, 8 GB memory capacity and Windows 10 operating system.
Particularly, the load data used in this paper is the real sample data of a region in 2018. These sample data have a total of 365 pieces, and their time resolution is 24. In order to reduce the influence of data distribution on the experimental results, the data is randomly sorted; 300 samples are selected as the training dataset, 30 samples are used as the validation dataset and 35 samples are used as the testing dataset. These datasets are depicted in Figure 6.

5.1. Parameter Settings

The attention mechanism is realized by a single-layer fully connected neural network, the number of neurons is 32 and the activation function is softmax.
The number of neurons in the hidden layer of the prediction model based on a BP neural network is 32, the activation function is ReLU, the number of neurons in the output layer is 24 and the activation function is linear. The unit number of the models based on GRU and BiGRU is 32, the number of neurons in the output layer is 24 and the activation function is linear. The BP, GRU, BiGRU, FA-BiGRU, TA-BiGRU, and DA-BiGRU models all use the Adam optimizer, and their hyperparameters β1, β2 and ε are set to 0.9, 0.999 and 1 × 10−8, respectively. In addition, MSE is used as the loss function, and the number of iterations of these models is set to 500.

5.2. Case 1: The Effectiveness of the BiGRU Model and Dual-Stage Mechanism

In order to better evaluate the effectiveness of the DA-BiGRU prediction model proposed in this paper, this section verifies the superiority of the BiGRU model and effectiveness of the dual-stage attention mechanism from the aspect of short-term load prediction. Persistence, BP, GRU, BiGRU, feature-attention-mechanism-based BiGRU (FA-BiGRU) and temporal-attention-mechanism-based BiGRU (TA-BiGRU) models are compared with the DA-BiGRU model in this case. The experimental results are shown in Table 1 and Figure 7, where Figure 7 is a comparison chart between the prediction results of different models and the real values on 29–31 December 2018.
As shown in Table 1 and Figure 7, the following conclusions can be drawn:
(1) The advantages of BiGRU model:
The prediction performance of the deep learning model is the best among the single prediction models (i.e., persistence, BP, GRU and BiGRU models), and the prediction accuracy of the BiGRU model is the highest. For example, compared with the classic baseline model persistence, the RMSE, MAE and SMAPE values of the BiGRU model are reduced by 15.46%, 14.38% and 0.942%, respectively, and the R2 value is increased by 1.67%. Compared with the shallow neural network BP model, the RMSE, MAE and SMAPE values of the BiGRU model are reduced by 10.56%, 9.40% and 0.531%, respectively, and the R2 value is increased by 1.98%. In addition, compared with the GRU model, the BiGRU model has the best RMSE, MAE, SMAPE and R2 values.
The reasons are as follows: Firstly, the machine learning model uses a large amount of historical load data for training, which can effectively capture the nonlinear relationship of load time series, and the prediction performance is improved compared with the persistence model. Secondly, the BiGRU model has a unique bidirectional propagation structure, which can link the past and future influencing factors with the current load time series so as to improve the accuracy of short-term load forecasting.
(2) The effectiveness of the dual-stage attention mechanism:
This model combined with the feature attention mechanism can automatically extract the correlation between each feature, which has the ability of reducing the prediction error of the model. Compared with the BiGRU model, the RMSE, MAE and SMAPE values of the FA-BiGRU model decreased by 4.60%, 7.92% and 7.16%, respectively, and the R2 value increased by 0.65%.
This model combined with the temporal attention mechanism realizes the adaptive extraction of features at important moments, which improves the prediction stability of the model. Compared with the BiGRU model, the RMSE, MAE and SMAPE values of the TA-BiGRU model decreased by 3.77%, 3.24% and 0.98%, respectively, and the R2 value increased by 0.11%.
In addition, compared with other comparison models in this case study, the proposed DA-BiGRU model has the best RMSE, MAE, MAPE and R2 values. This is because this model combines feature and temporal attention mechanisms, which can improve the sensitivity of the model to key features and key time steps, and finally achieve the purpose of improving prediction accuracy.

5.3. Case 2: The Effectiveness of the CS-GWO Algorithm

The popular suite of benchmark functions in validating optimization performance, i.e., CEC 2017 [43], is utilized in this subsection to conduct extensive optimization experiments. The CEC 2017 test suite has 30 functions, which can be divided into four categories: unimodal functions (F1–F3), multimodal functions (F4–F10), hybrid functions (F11–F20) and composition functions (F21–F30). The ideal optimal value of each benchmark functions is 0.
Moreover, the well-known optimization algorithms (i.e., PSO, WOA, GWO and CSO) are compared with CS-GWO to evaluate the effectiveness of the CS-GWO algorithm from various perspectives, including accuracy, the Wilcoxon signed-rank test and a paired samples t-test.

5.3.1. The Setting of the Numerical Experiments

The dimension of benchmark functions is uniformly set to 30 in this subsection. For a fair comparison, the number of iterations of the swarm intelligence optimization algorithm and the number of individuals is set to 3000 and 30, respectively. The position of PSO is set in the range of [−1, 1], and the limit of speed is set in the range of [−0.5, 0.5] [44]. For WOA [45], GWO [46] and CS-GWO, the limit of individual is set in the range of [−1, 1]. Among them, the vertical crossover probability of CSO and CS-GWO is set to 60%, and the horizontal intersection crossover is set to 100% [47]. To reduce statistical errors, all the reported results in this subsection are based on 30 independent runs.

5.3.2. The Comparison of Optimization Accuracy

The above-mentioned algorithms are evaluated using the CEC 2017 test suite, and the experimental results are shown in Table 2. The reported values in Table 2 are based on the errors between the terminated values of the optimization process and the target values of the benchmark functions. To intuitively quantify the optimization ability of the metaheuristics, mean values (Mean), minimum values (Min), maximum values (Max), standard deviation (Std) and ranks (Rank) are used to evaluate the accuracy. Mean, Min and Max reveal the optimization accuracy of the algorithm, Std reveals the optimization stability and Rank is based on the Friedman test [48] to rank the optimization performance of the algorithm from the aspect of statistics. Moreover, the minimum values of Mean in each benchmark function are shown in bold.
For the unimodal and functions (i.e., F1–F10), CS-GWO achieves the best results six times, and WOA and CSO share the remaining four best results. This reveals that WOA performs well for simple low-dimensional optimization problems. For the 10 hybrid functions (i.e., F11–F20), CS-GWO achieves the best performance seven times. Although WOA has the best values for F12 and F14, it gets the second-worst values in 4 out of 10 cases. This reveals that WOA has unstable performance in solving complex problems [49,50]. For the 10 composition functions (i.e., F21–F30), CS-GWO achieves the best performance six times, followed by CSO’s three times. The worst rank of CSO in composition functions is three with F21 and F27, indicating that CSO has the ability to escape local optimum when applied in complex optimization problems [47].
Particularly, GWO never dominates in optimization of all benchmark functions, but it ranks third overall. This reveals that GWO has stable performance in solving optimization problems but easily falls into the optima trap [51]. Comprehensively speaking, CS-GWO ranks first overall among the optimization algorithms and obtains the best performance in 19 out of 30 functions, whether from the aspect of the optimization accuracy or stability.
From the above analysis, we can conclude that CS-GWO performs the best with the 30 dimensional optimization problems among all the compared algorithms. This is because CS-GWO combines the stable optimization performance of GWO and outstanding ability in finding global optima in the problem-solving space of CSO.

5.3.3. Wilcoxon Signed-Rank Test and Paired Samples t-Test

In order to further prove the validity of the CS-GWO algorithm, the parametric and non-parametric tests, which are the paired samples t-test (PSTT) [52] and Wilcoxon signed-rank test (WSRT) [53], are adopted to evaluate the difference of optimization performance between CS-GWO and the comparison algorithms in 30 benchmark functions. The null hypothesis of PSTT and WSRT are that there is no difference between two compared samples. If the difference approximately obeyed the normal distribution, PSTT is used. When this premise is not satisfied, WSRT can be selected. The results of PSTT and WSRT are shown in Table 3 and Table 4, respectively.
Two indicators including t-value and Sig. (2-tailed) can be obtained from PSTT. If the Sig. (2-tailed) is less than 0.05, it can be concluded that the CS-GWO algorithm is better than the compared algorithm. In addition, four indicators including p-value, R+, R− and winner can be obtained from WSRT. If the p-value is less than 0.05, the null hypothesis can be rejected at 5% significance level. R+ represents a mean error of the CS-GWO algorithm that is higher than that of the compared one. R− represents a mean error of the CS-GWO algorithm that is lower than that of the compared one. Finally, winner indicates whether the CS-GWO algorithm is superior to the compared algorithm, “+” indicates that the CS-GWO algorithm is better than the compared algorithm, “−” indicates that the CS-GWO algorithm is worse than the compared algorithm and “=” indicates that the performance of the two algorithms display no obvious difference.
It can be seen that most of the Sig. (2-tailed) values in Table 3 are less than 0.05, which means that there is an obvious difference between the CS-GWO algorithm and the comparison algorithms, further indicating that the CS-GWO algorithm is superior to the involved algorithms. In Table 4, most of the combinations (p-Value, R+, R, winner) are (1.734 × 10−6, 0, 465, +), revealing that the CS-GWO algorithm outperforms the comparison algorithms. Furthermore, the results of ‘+/=/−’ is 94/14/12, indicating that the CS-GWO algorithm is better than the compared algorithm in 94 out of 120 cases.
From a statistical perspective, it can be concluded that CS-GWO dominates the other compared algorithms in optimization problems with CEC 2017 test functions.

5.4. Case 3: The Effectiveness of the CS-GWO-DA-BiGRU Model

In order to verify the effectiveness of the combination of the crisscross grey wolf optimization algorithm and the DA-BiGRU model in short-term load forecasting, PSO-DA-BiGRU, WOA-DA-BiGRU and GWO-DA-BiGRU models are compared with the combined CS-GWO-DA-BiGRU model in this case. In this subsection, the number of iterations of the swarm intelligence optimization algorithm is uniformly set to 200, and the number of individuals is set to 20.
The experimental results are shown in Table 5 and Figure 8, where Figure 8 is a comparison chart between the prediction results of different models and the real values on 29–31 December 2018.
As shown in Table 2 and Figure 8, the following conclusions can be drawn:
(1) The effectiveness of the swarm intelligence optimization algorithm:
The prediction model combined with the swarm intelligence optimization algorithm has better prediction performance than the single prediction model (i.e., DA-BiGRU). For example, compared with the DA-BiGRU model, the RMSE, MAE and SMAPE values of the PSO-DA-BiGRU and GWO-DA-BiGRU models are reduced by 1.75% and 3.99%, 2.67% and 3.21% and 1.32% and 0.59%, respectively, and the R2 values are increased by 0.21% and 0.53%, respectively. This is because the weights and the bias of the DA-BiGRU model are optimized by the swarm intelligence optimization algorithm in the initial stage of training, which can effectively avoid the problems of gradient disappearance and gradient explosion, and further improve the accuracy of load forecasting.
(2) The superiority of the CS-GWO algorithm:
Among all the comparison forecasting models, the proposed CS-GWO-DA-BiGRU short-term load forecasting model has the highest forecasting accuracy. For example, the RMSE, MAE and SMAPE are reduced by 3.86%, 1.37% and 0.30% from those of the second-best performing CSO-DA-BiGRU model, respectively. From the aspect of the R2 value, the CS-GWO-DA-BiGRU model has an increase of 0.42% compared with the CSO-DA-BiGRU model. Therefore, the CS-GWO algorithm combined with horizontal crossover and vertical crossover operators can improve the global search ability and enhance the diversity of the population, making a great contribution to improving short-term load forecasting.

6. Discussion

In this paper, a high-precision model called CS-GWO-DA-BiGRU is presented in short-term load forecasting problems. However, the proposed model still has some shortcomings that need to be improved. The limitations and future research can be summarized as follows.
(1) The CS-GWO algorithm only focuses on improving the accuracy of short-term load prediction while ignoring the prediction stability, leading to unstable prediction when extending to new data. In the future, we plan to upgrade CS-GWO to a multi-object CS-GWO algorithm to improve the accuracy and stability of short-term load prediction simultaneously.
(2) At present, the intelligent big data platform is valuable for the improvement of the prediction model. In future work, the proposed CS-GWO-DA-BiGRU prediction model will be embedded into the intelligent big data platform to construct an intelligent load forecasting system.

7. Conclusions

Short-term load prediction is essential for the stable operation and safety management of power systems. Therefore, this paper proposes a hybrid model for short-term load prediction, named CS-GWO-DA-BiGRU, which consists of a dual-stage attention mechanism, crisscross grey wolf optimization algorithm and bidirectional gated recurrent unit. The main contributions of this paper can be concluded as follows:
(1) Different from the conventional feature mechanism applied in short-term load forecasting, this paper proposes a dual-stage attention mechanism by combining feature and temporal attention mechanisms. Based on case 1, compared with FA-BiGRU, the RMSE, MAE and SMAPE values of the DA-BiGRU model are reduced by 1.79%, 0.74% and 0.70%, respectively, and the R2 value is increased by 0.21%. Therefore, DA can effectively capture the correlation relationship of input feature and temporal dependence in load time series simultaneously.
(2) By combining horizontal and vertical crossover operators, the global search ability and population diversity of GWO are enhanced. Based on the Friedman test in case 2, CS-GWO ranks first among the well-known algorithms and achieves the best results for 19 out of 30 functions in CEC 2017. In addition, CS-GWO outperforms the compared algorithms in 94 out of 120 cases based on the Wilcoxon signed-rank test. Furthermore, for the proposed CS-GWO-DA-BiGRU model in case 3, which is based on CS-GWO, the R2 value has an increase of 0.42% compared with the CSO-DA-BiGRU model and has the best forecasting performance.

Author Contributions

Conceptualization, R.G.; software, X.L.; validation, X.L.; formal analysis, R.G.; investigation, R.G. and X.L.; resources, R.G.; data curation, X.L.; writing—original draft preparation, X.L.; writing—review and editing, R.G.; visualization, X.L.; supervision, R.G.; project administration, R.G.; funding acquisition, R.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (Grant No. 61561007) and Guangxi Natural Science Foundation (Grant No. 2017GXNSFAA198168).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclatures

Abbreviations
WOAwhale optimization algorithm
GWOgrey wolf optimization
LSTMlong short-term memory
GRUgated recurrent unit
FAfeature attention mechanism
TAtemporal attention mechanism
DAdual-stage attention mechanism
CS-GWOcrisscross grey wolf optimizer algorithm
RNNrecurrent neural network
BiGRUbidirectional gated recurrent unit
CSOcrisscross optimization algorithm
HChorizontal crossover
VCvertical crossover
RMSEroot mean square error
MAEmean absolute error
SMAPEsymmetric mean absolute percentage error
R2decision coefficient
Meanmean value
Minminimum value
Maxmaximum value
Stdstandard deviation
Rankranks
PSTTpaired samples t-test
WSRTWilcoxon signed-rank test
Formula symbols
  x t input data at t-th time step
h t 1 , h t , h T hidden state at (t−1)-th, t-th and T-th time step
rtreset gate
W r U r b r weight metrices and bias of reset gate
  σ sigmoid activation function
ztupdate gate
W z , U z , b z weight metrices and bias of update gate
W h , b h weight metrices and bias of candidate output
  h ˜ t candidate hidden state
h t , h t state information of forward and backward propagation
w t , w t weight metrices of hidden layer in forward and backward propagation
  b t bias of the hidden layer
  G ( ) calculation process of GRU
eunnormalized attention weight
  α normalized attention weight
  c intermediate semantic vector
  L ( i ) electric load time series of the previous day
  t ( i ) highest and lowest temperature of the previous day and the current day
  r ( i ) rainfall of the previous day and the current day
  d ( i ) weather day type of the previous day and the current day
  F θ function map of prediction model
  X ( i ) input of prediction model
  Y ^ ( i ) predicted load values of the current day
Nnumber of samples
W e , b e weight matrix and bias in FA
  X ATT adaptively optimized feature vector
W f , b f weight matrix and bias in TA
  n p number of hidden elements in the last layer of BiGRU
d t , d T intermediate semantic vector in t-th and T-th iteration
W y , b y weight metrices and bias of the feedforward network in TA
  θ all parameters of DA-BiGRU model
  J ( ) loss function of DA-BiGRU model
  Φ population of grey wolf
  n M population size
Dpopulation dimension
θ α ( t ) , θ β ( t ) , θ δ ( t ) , θ ω ( t ) grey wolves  α β δ  and  ω  at t-th iteration
A 1 , A 2 , A 3 , C 1 C 2 , C 3 , A, Csynergy coefficients
r 1 r 2 r 3 r 4 r random number
Φ H C , Φ V C offspring population
Tmaximum number of iterations
Y t e s t , Y ^ t e s t actual and predicted load value in testing dataset
  Y ¯ t e s t average value of the actual load value
  n t e s t sample number of the testing dataset

References

  1. Vanting, N.B.; Ma, Z.; Jørgensen, B.N. A Scoping Review of Deep Neural Networks for Electric Load Forecasting. Energy Inform. 2021, 4, 49. [Google Scholar] [CrossRef]
  2. Liu, Y.; Dutta, S.; Kong, A.W.K.; Yeo, C.K. An Image Inpainting Approach to Short-Term Load Forecasting. IEEE Trans. Power Syst. 2022, 38, 177–187. [Google Scholar] [CrossRef]
  3. Li, L.; Guo, L.; Wang, J.; Peng, H. Short-Term Load Forecasting Based on Spiking Neural P Systems. Appl. Sci. 2023, 13, 792. [Google Scholar] [CrossRef]
  4. Li, S.; Kong, X.; Yue, L.; Liu, C.; Khan, M.A.; Yang, Z.; Zhang, H. Short-Term Electrical Load Forecasting Using Hybrid Model of Manta Ray Foraging Optimization and Support Vector Regression. J. Clean. Prod. 2023, 388, 135856. [Google Scholar] [CrossRef]
  5. Mayrink, V.; Hippert, H.S. A Hybrid Method Using Exponential Smoothing and Gradient Boosting for Electrical Short-Term Load Forecasting. In Proceedings of the 2016 IEEE Latin American Conference on Computational Intelligence (LA-CCI), Cartagena, Colombia, 2–4 November 2016; pp. 1–6. [Google Scholar]
  6. Wang, Y.; Kong, Y.; Tang, X.; Chen, X.; Xu, Y.; Chen, J.; Sun, S.; Guo, Y.; Chen, Y. Short-Term Industrial Load Forecasting Based on Ensemble Hidden Markov Model. IEEE Access 2020, 8, 160858–160870. [Google Scholar] [CrossRef]
  7. Guo, L.; Wu, P.; Lou, S.; Gao, J.; Liu, Y. A Multi-Feature Extraction Technique Based on Principal Component Analysis for Nonlinear Dynamic Process Monitoring. J. Process Control 2020, 85, 159–172. [Google Scholar] [CrossRef]
  8. Mehdipour Pirbazari, A.; Farmanbar, M.; Chakravorty, A.; Rong, C. Short-Term Load Forecasting Using Smart Meter Data: A Generalization Analysis. Processes 2020, 8, 484. [Google Scholar] [CrossRef]
  9. Li, B.; Hou, B.; Yu, W.; Lu, X.; Yang, C. Applications of Artificial Intelligence in Intelligent Manufacturing: A Review. Front. Inf. Technol. Electron. Eng. 2017, 18, 86–96. [Google Scholar] [CrossRef]
  10. Yu, F.; Xu, X. A Short-Term Load Forecasting Model of Natural Gas Based on Optimized Genetic Algorithm and Improved BP Neural Network. Appl. Energy 2014, 134, 102–113. [Google Scholar] [CrossRef]
  11. Li, S.; Goel, L.; Wang, P. An Ensemble Approach for Short-Term Load Forecasting by Extreme Learning Machine. Appl. Energy 2016, 170, 22–29. [Google Scholar] [CrossRef]
  12. Lu, H.; Azimi, M.; Iseley, T. Short-Term Load Forecasting of Urban Gas Using a Hybrid Model Based on Improved Fruit Fly Optimization Algorithm and Support Vector Machine. Energy Rep. 2019, 5, 666–677. [Google Scholar] [CrossRef]
  13. Niu, D.; Dai, S. A Short-Term Load Forecasting Model with a Modified Particle Swarm Optimization Algorithm and Least Squares Support Vector Machine Based on the Denoising Method of Empirical Mode Decomposition and Grey Relational Analysis. Energies 2017, 10, 408. [Google Scholar] [CrossRef] [Green Version]
  14. Shao, L.; Guo, Q.; Li, C.; Li, J.; Yan, H. Short-Term Load Forecasting Based on EEMD-WOA-LSTM Combination Model. Appl. Bionics Biomech. 2022, 2022, 2166082. [Google Scholar] [CrossRef] [PubMed]
  15. Li, T.; Qian, Z.; He, T. Short-Term Load Forecasting with Improved CEEMDAN and GWO-Based Multiple Kernel ELM. Complexity 2020, 2020, 1209547. [Google Scholar] [CrossRef] [Green Version]
  16. Image Segmentation of Leaf Spot Diseases on Maize Using Multi-Stage Cauchy-Enabled Grey Wolf Algorithm. Eng. Appl. Artif. Intell. 2022, 109, 104653. [CrossRef]
  17. Khan, B.; Khalid, R.; Javed, M.U.; Javaid, S.; Ahmed, S.; Javaid, N. Short-Term Load and Price Forecasting Based on Improved Convolutional Neural Network. In Proceedings of the 2020 3rd International Conference on Computing, Mathematics and Engineering Technologies (iCoMET), Sukkur, Pakistan, 29–30 January 2020; pp. 1–6. [Google Scholar]
  18. Muzaffar, S.; Afshari, A. Short-Term Load Forecasts Using LSTM Networks. Energy Procedia 2019, 158, 2922–2927. [Google Scholar] [CrossRef]
  19. Jung, S.; Moon, J.; Park, S.; Hwang, E. An Attention-Based Multilayer GRU Model for Multistep-Ahead Short-Term Load Forecasting. Sensors 2021, 21, 1639. [Google Scholar] [CrossRef]
  20. Alhussein, M.; Aurangzeb, K.; Haider, S.I. Hybrid CNN-LSTM Model for Short-Term Individual Household Load Forecasting. IEEE Access 2020, 8, 180544–180557. [Google Scholar] [CrossRef]
  21. Sajjad, M.; Khan, Z.A.; Ullah, A.; Hussain, T.; Ullah, W.; Lee, M.Y.; Baik, S.W. A Novel CNN-GRU-Based Hybrid Approach for Short-Term Residential Load Forecasting. IEEE Access 2020, 8, 143759–143768. [Google Scholar] [CrossRef]
  22. Cai, C.; Li, Y.; Su, Z.; Zhu, T.; He, Y. Short-Term Electrical Load Forecasting Based on VMD and GRU-TCN Hybrid Network. Appl. Sci. 2022, 12, 6647. [Google Scholar] [CrossRef]
  23. Alsharekh, M.F.; Habib, S.; Dewi, D.A.; Albattah, W.; Islam, M.; Albahli, S. Improving the Efficiency of Multistep Short-Term Electricity Load Forecasting via R-CNN with ML-LSTM. Sensors 2022, 22, 6913. [Google Scholar] [CrossRef]
  24. Chen, Q.; Zhang, W.; Zhu, K.; Zhou, D.; Dai, H.; Wu, Q. A Novel Trilinear Deep Residual Network with Self-Adaptive Dropout Method for Short-Term Load Forecasting. Expert Syst. Appl. 2021, 182, 115272. [Google Scholar] [CrossRef]
  25. Kim, S.H.; Lee, G.; Kwon, G.-Y.; Kim, D.-I.; Shin, Y.-J. Deep Learning Based on Multi-Decomposition for Short-Term Load Forecasting. Energies 2018, 11, 3433. [Google Scholar] [CrossRef] [Green Version]
  26. Bouktif, S.; Fiaz, A.; Ouni, A.; Serhani, M.A. Optimal Deep Learning Lstm Model for Electric Load Forecasting Using Feature Selection and Genetic Algorithm: Comparison with Machine Learning Approaches. Energies 2018, 11, 1636. [Google Scholar] [CrossRef] [Green Version]
  27. Kong, X.; Liu, X.; Shi, R.; Lee, K.Y. Wind Speed Prediction Using Reduced Support Vector Machines with Feature Selection. Neurocomputing 2015, 169, 449–456. [Google Scholar] [CrossRef]
  28. Li, S.; Wang, P.; Goel, L. Wind Power Forecasting Using Neural Network Ensembles with Feature Selection. IEEE Trans. Sustain. Energy 2015, 6, 1447–1456. [Google Scholar] [CrossRef]
  29. Meng, A.; Chen, S.; Ou, Z.; Ding, W.; Zhou, H.; Fan, J.; Yin, H. A Hybrid Deep Learning Architecture for Wind Power Prediction Based on Bi-Attention Mechanism and Crisscross Optimization. Energy 2022, 238, 121795. [Google Scholar] [CrossRef]
  30. Zhang, B.; Wu, J.-L.; Chang, P.-C. A Multiple Time Series-Based Recurrent Neural Network for Short-Term Load Forecasting. Soft Comput. 2018, 22, 4099–4112. [Google Scholar] [CrossRef]
  31. Wang, S.; Wang, X.; Wang, S.; Wang, D. Bi-Directional Long Short-Term Memory Method Based on Attention Mechanism and Rolling Update for Short-Term Load Forecasting. Int. J. Electr. Power Energy Syst. 2019, 109, 470–479. [Google Scholar] [CrossRef]
  32. Fazlipour, Z.; Mashhour, E.; Joorabian, M. A Deep Model for Short-Term Load Forecasting Applying a Stacked Autoencoder Based on LSTM Supported by a Multi-Stage Attention Mechanism. Appl. Energy 2022, 327, 120063. [Google Scholar] [CrossRef]
  33. Ribeiro, A.H.; Tiels, K.; Aguirre, L.A.; Schön, T. Beyond Exploding and Vanishing Gradients: Analysing RNN Training Using Attractors and Smoothness. In Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, Online, 26–28 August 2020; pp. 2370–2380. [Google Scholar]
  34. Dewangan, F.; Abdelaziz, A.Y.; Biswal, M. Load Forecasting Models in Smart Grid Using Smart Meter Information: A Review. Energies 2023, 16, 1404. [Google Scholar] [CrossRef]
  35. Zhang, D.; Kabuka, M.R. Combining Weather Condition Data to Predict Traffic Flow: A GRU-Based Deep Learning Approach. IET Intell. Transp. Syst. 2018, 12, 578–585. [Google Scholar] [CrossRef]
  36. Xuan, Y.; Si, W.; Zhu, J.; Sun, Z.; Zhao, J.; Xu, M.; Xu, S. Multi-Model Fusion Short-Term Load Forecasting Based on Random Forest Feature Selection and Hybrid Neural Network. IEEE Access 2021, 9, 69002–69009. [Google Scholar] [CrossRef]
  37. Li, C.; Liu, D.; Wang, M.; Wang, H.; Xu, S. Detection of Outliers in Time Series Power Data Based on Prediction Errors. Energies 2023, 16, 582. [Google Scholar] [CrossRef]
  38. Niu, Z.; Yu, Z.; Tang, W.; Wu, Q.; Reformat, M. Wind Power Forecasting Using Attention-Based Gated Recurrent Unit Network. Energy 2020, 196, 117081. [Google Scholar] [CrossRef]
  39. Keskar, N.S.; Socher, R. Improving Generalization Performance by Switching from Adam to Sgd. arXiv 2017, arXiv:171207628. [Google Scholar]
  40. Meng, A.; Zeng, C.; Wang, P.; Chen, D.; Zhou, T.; Zheng, X.; Yin, H. A High-Performance Crisscross Search Based Grey Wolf Optimizer for Solving Optimal Power Flow Problem. Energy 2021, 225, 120211. [Google Scholar] [CrossRef]
  41. Meng, A.; Chen, Y.; Yin, H.; Chen, S. Crisscross Optimization Algorithm and Its Application. Knowl. Based Syst. 2014, 67, 218–229. [Google Scholar] [CrossRef]
  42. Guo, Y.; Lu, W.; Li, X.; Huang, Q. Single Image Reflection Removal Based on Residual Attention Mechanism. Appl. Sci. 2023, 13, 1618. [Google Scholar] [CrossRef]
  43. Mohamed, A.W.; Hadi, A.A.; Fattouh, A.M.; Jambi, K.M. LSHADE with Semi-Parameter Adaptation Hybrid with CMA-ES for Solving CEC 2017 Benchmark Problems. In Proceedings of the 2017 IEEE Congress on Evolutionary Computation (CEC), San Sebastián, Spain, 5–8 June 2017; pp. 145–152. [Google Scholar]
  44. Zhaoyu, P.; Shengzhu, L.; Hong, Z.; Nan, Z. The Application of the Pso Based BP Network in Short-Term Load Forecasting. Phys. Procedia 2012, 24, 626–632. [Google Scholar] [CrossRef] [Green Version]
  45. Lu, Y.; Wang, G. A Load Forecasting Model Based on Support Vector Regression with Whale Optimization Algorithm. Multimed. Tools Appl. 2023, 82, 9939–9959. [Google Scholar] [CrossRef]
  46. Barman, M.; Dev Choudhury, N.B. A Similarity Based Hybrid GWO-SVM Method of Power System Load Forecasting for Regional Special Event Days in Anomalous Load Situations in Assam, India. Sustain. Cities Soc. 2020, 61, 102311. [Google Scholar] [CrossRef]
  47. Meng, A.; Li, Z.; Yin, H.; Chen, S.; Guo, Z. Accelerating Particle Swarm Optimization Using Crisscross Search. Inf. Sci. 2016, 329, 52–72. [Google Scholar] [CrossRef]
  48. Derrac, J.; García, S.; Molina, D.; Herrera, F. A Practical Tutorial on the Use of Nonparametric Statistical Tests as a Methodology for Comparing Evolutionary and Swarm Intelligence Algorithms. Swarm Evol. Comput. 2011, 1, 3–18. [Google Scholar] [CrossRef]
  49. Hashim, F.A.; Houssein, E.H.; Hussain, K.; Mabrouk, M.S.; Al-Atabany, W. Honey Badger Algorithm: New Metaheuristic Algorithm for Solving Optimization Problems. Math. Comput. Simul. 2022, 192, 84–110. [Google Scholar] [CrossRef]
  50. Abdel-Basset, M.; Mohamed, R.; Azeem, S.A.A.; Jameel, M.; Abouhawwash, M. Kepler Optimization Algorithm: A New Metaheuristic Algorithm Inspired by Kepler’s Laws of Planetary Motion. Knowl. Based Syst. 2023, 110454, in press. [Google Scholar] [CrossRef]
  51. Nadimi-Shahraki, M.H.; Zamani, H.; Fatahi, A.; Mirjalili, S. MFO-SFR: An Enhanced Moth-Flame Optimization Algorithm Using an Effective Stagnation Finding and Replacing Strategy. Mathematics 2023, 11, 862. [Google Scholar] [CrossRef]
  52. Zimmerman, D.W. Teacher’s Corner: A Note on Interpretation of the Paired-Samples t Test. J. Educ. Behav. Stat. 1997, 22, 349–360. [Google Scholar] [CrossRef]
  53. Rey, D.; Neuhäuser, M. Wilcoxon-Signed-Rank Test. In International Encyclopedia of Statistical Science; Springer: Berlin/Heidelberg, Germany, 2011; pp. 1658–1659. [Google Scholar]
Figure 1. The network structure of GRU.
Figure 1. The network structure of GRU.
Energies 16 02878 g001
Figure 2. The network structure of BiGRU.
Figure 2. The network structure of BiGRU.
Energies 16 02878 g002
Figure 3. Feature attention mechanism.
Figure 3. Feature attention mechanism.
Energies 16 02878 g003
Figure 4. Temporal attention mechanism.
Figure 4. Temporal attention mechanism.
Energies 16 02878 g004
Figure 5. The flow chart of CS-GWO.
Figure 5. The flow chart of CS-GWO.
Energies 16 02878 g005
Figure 6. Training, validation and testing datasets used in this paper: (a) training dataset, (b) validation dataset, (c) testing dataset.
Figure 6. Training, validation and testing datasets used in this paper: (a) training dataset, (b) validation dataset, (c) testing dataset.
Energies 16 02878 g006
Figure 7. Case study 1: comparison of forecasting results on 29–31 December 2018.
Figure 7. Case study 1: comparison of forecasting results on 29–31 December 2018.
Energies 16 02878 g007
Figure 8. Case study 2: comparison of forecasting results on 29–31 December 2018.
Figure 8. Case study 2: comparison of forecasting results on 29–31 December 2018.
Energies 16 02878 g008
Table 1. The experiment results of case study 1.
Table 1. The experiment results of case study 1.
Prediction ModelRMSE/MWMAE/MWSMAPER2
persistence36.67929.2584.8100.900
BP34.67127.6504.3990.911
GRU32.14125.1283.8920.924
BiGRU31.00925.0513.8680.929
FA-BiGRU29.58323.0683.5910.935
TA-BiGRU29.84024.2393.8300.930
DA-BiGRU29.05322.8973.5660.937
Table 2. Comparison of CS-GWO with well-known algorithms for CEC 2017 test functions.
Table 2. Comparison of CS-GWO with well-known algorithms for CEC 2017 test functions.
FunctionsMetricsPSOWOAGWOCSOCS-GWO
F1Mean3.169 × 10112.408 × 1033.781 × 10102.275 × 10104.875 × 103
Min1.137 × 10111.070 × 1025.121 × 1094.985 × 1091.000 × 102
Max5.806 × 10119.847 × 1038.496 × 10106.467 × 10101.771 × 104
Std9.145 × 10102.219 × 1032.197 × 10101.430 × 10105.283 × 103
Rank51432
F2Mean8.880 × 1027.076 × 1026.260 × 1026.012 × 1025.550 × 102
Min8.253 × 1026.184 × 1025.892 × 1025.705 × 1025.129 × 102
Max9.519 × 1028.124 × 1027.649 × 1026.601 × 1026.350 × 102
Std3.825 × 1014.662 × 1013.205 × 1012.046 × 1013.437 × 101
Rank54321
F3Mean1.406 × 1055.016 × 1044.718 × 1043.481 × 1043.001 × 102
Min8.673 × 1042.929 × 1043.286 × 1041.132 × 1043.000 × 102
Max3.292 × 1057.088 × 1046.894 × 1046.353 × 1043.003 × 102
Std6.473 × 1041.017 × 1049.881 × 1031.028 × 1046.049 × 10−2
Rank54321
F4Mean6.990 × 1034.687 × 1026.504 × 1025.645 × 1024.937 × 102
Min2.951 × 1034.001 × 1025.166 × 1024.833 × 1024.641 × 102
Max1.422 × 1044.911 × 1021.218 × 1037.337 × 1025.187 × 102
Std2.613 × 1031.995 × 1011.457 × 1026.179 × 1011.520 × 101
Rank51432
F5Mean8.966 × 1027.087 × 1026.214 × 1025.966 × 1025.774 × 102
Min8.469 × 1026.323 × 1025.671 × 1025.681 × 1025.202 × 102
Max9.739 × 1027.816 × 1026.944 × 1026.708 × 1026.720 × 102
Std3.825 × 1014.073 × 1012.883 × 1012.248 × 1014.910 × 101
Rank54321
F6Mean7.009 × 1026.730 × 1026.270 × 1026.196 × 1026.003 × 102
Min6.795 × 1026.482 × 1026.126 × 1026.102 × 1026.000 × 102
Max7.331 × 1027.250 × 1026.460 × 1026.366 × 1026.017 × 102
Std1.221 × 1011.364 × 1018.497 × 1006.550 × 1004.418 × 10−1
Rank54321
F7Mean1.397 × 1031.140 × 1038.841 × 1028.559 × 1028.710 × 102
Min1.218 × 1031.022 × 1038.019 × 1027.960 × 1027.698 × 102
Max1.525 × 1031.324 × 1031.076 × 1039.970 × 1028.977 × 102
Std6.925 × 1017.795 × 1016.307 × 1015.190 × 1013.069 × 101
Rank54312
F8Mean1.103 × 1039.403 × 1029.039 × 1028.979 × 1028.742 × 102
Min1.002 × 1039.114 × 1028.541 × 1028.479 × 1028.129 × 102
Max1.164 × 1039.910 × 1029.464 × 1021.032 × 1039.867 × 102
Std3.661 × 1012.327 × 1012.433 × 1013.537 × 1014.819 × 101
Rank54321
F9Mean9.729 × 1037.439 × 1032.287 × 1032.092 × 1039.002 × 102
Min5.733 × 1033.310 × 1031.410 × 1031.035 × 1039.000 × 102
Max1.324 × 1041.376 × 1044.608 × 1033.691 × 1039.029 × 102
Std1.659 × 1033.347 × 1037.803 × 1027.556 × 1025.347 × 10−1
Rank54321
F10Mean8.429 × 1036.333 × 1035.139 × 1034.483 × 1037.671 × 103
Min6.359 × 1034.002 × 1032.849 × 1033.091 × 1036.807 × 103
Max1.005 × 1049.761 × 1038.679 × 1037.903 × 1038.508 × 103
Std8.850 × 1021.699 × 1031.674 × 1031.270 × 1034.446 × 102
Rank53214
F11Mean8.717 × 1031.220 × 1032.183 × 1031.786 × 1031.159 × 103
Min4.894 × 1031.153 × 1031.389 × 1031.276 × 1031.108 × 103
Max1.418 × 1041.298 × 1034.783 × 1033.937 × 1031.217 × 103
Std2.371 × 1033.880 × 1019.575 × 1027.822 × 1023.555 × 101
Rank52431
F12Mean2.599 × 10101.764 × 1053.831 × 1085.773 × 1083.160 × 105
Min2.572 × 1091.191 × 1042.075 × 1072.392 × 1072.261 × 104
Max2.031 × 10118.030 × 1051.507 × 1093.799 × 1091.485 × 106
Std3.719 × 10101.616 × 1053.689 × 1088.210 × 1083.465 × 105
Rank51432
F13Mean1.224 × 10101.733 × 1041.497 × 1084.750 × 1071.410 × 104
Min2.027 × 1083.680 × 1033.990 × 1044.008 × 1041.416 × 103
Max2.020 × 10114.611 × 1041.498 × 1091.405 × 1094.460 × 104
Std3.616 × 10101.109 × 1044.094 × 1082.564 × 1081.138 × 104
Rank52431
F14Mean4.157 × 1061.219 × 1044.015 × 1051.794 × 1054.376 × 104
Min1.956 × 1041.746 × 1032.666 × 1042.325 × 1034.285 × 103
Max4.485 × 1071.370 × 1051.337 × 1069.272 × 1052.532 × 105
Std8.338 × 1062.417 × 1044.196 × 1052.868 × 1055.106 × 104
Rank51432
F15Mean1.424 × 1098.116 × 1034.410 × 1064.594 × 1063.994 × 103
Min2.520 × 1051.731 × 1032.667 × 1041.355 × 1041.607 × 103
Max1.273 × 10103.219 × 1043.595 × 1079.453 × 1072.100 × 104
Std2.717 × 1098.011 × 1039.902 × 1061.740 × 1074.244 × 103
Rank52431
F16Mean5.003 × 1032.990 × 1032.559 × 1032.455 × 1032.489 × 103
Min3.196 × 1032.365 × 1032.069 × 1032.033 × 1031.700 × 103
Max1.215 × 1043.871 × 1033.254 × 1033.319 × 1033.017 × 103
Std2.024 × 1033.759 × 1023.002 × 1023.224 × 1023.925 × 102
Rank54312
F17Mean5.052 × 1033.017 × 1032.460 × 1032.392 × 1032.126 × 103
Min3.917 × 1032.269 × 1032.102 × 1031.990 × 1031.612 × 103
Max1.022 × 1043.507 × 1033.213 × 1033.419 × 1033.014 × 103
Std1.165 × 1033.432 × 1022.835 × 1023.141 × 1023.478 × 102
Rank54321
F18Mean1.149 × 1087.835 × 1051.858 × 1061.463 × 1061.632 × 105
Min2.326 × 1068.942 × 1045.083 × 1048.757 × 1044.111 × 104
Max7.292 × 1082.771 × 1062.160 × 1078.673 × 1064.130 × 105
Std2.234 × 1086.988 × 1054.064 × 1061.773 × 1068.817 × 104
Rank54321
F19Mean1.516 × 1099.880 × 1031.345 × 1072.830 × 1067.321 × 103
Min8.062 × 1061.979 × 1033.550 × 1046.985 × 1031.991 × 103
Max2.216 × 10104.385 × 1043.380 × 1081.373 × 1073.052 × 104
Std4.402 × 1099.797 × 1036.135 × 1073.282 × 1067.050 × 103
Rank52431
F20Mean9.231 × 1036.463 × 1036.165 × 1035.747 × 1034.678 × 103
Min5.555 × 1032.300 × 1034.389 × 1032.475 × 1032.300 × 103
Max1.129 × 1049.968 × 1031.015 × 1049.513 × 1039.283 × 103
Std1.071 × 1031.958 × 1031.440 × 1031.854 × 1033.017 × 103
Rank54321
F21Mean2.697 × 1032.520 × 1032.408 × 1032.388 × 1032.403 × 103
Min2.583 × 1032.411 × 1032.373 × 1032.341 × 1032.319 × 103
Max2.838 × 1032.650 × 1032.523 × 1032.436 × 1032.469 × 103
Std5.997 × 1016.193 × 1012.773 × 1012.040 × 1013.968 × 101
Rank54213
F22Mean9.247 × 1036.023 × 1036.002 × 1036.264 × 1033.598 × 103
Min6.356 × 1032.300 × 1032.731 × 1032.672 × 1032.300 × 103
Max1.083 × 1041.204 × 1041.041 × 1049.926 × 1039.689 × 103
Std1.019 × 1032.548 × 1031.493 × 1032.164 × 1032.650 × 103
Rank54231
F23Mean3.363 × 1033.479 × 1032.792 × 1032.757 × 1032.708 × 103
Min3.098 × 1033.095 × 1032.726 × 1032.701 × 1032.674 × 103
Max3.869 × 1033.882 × 1032.946 × 1032.895 × 1032.767 × 103
Std1.838 × 1021.772 × 1025.363 × 1013.990 × 1012.849 × 101
Rank45321
F24Mean3.533 × 1033.531 × 1032.993 × 1032.947 × 1032.970 × 103
Min3.281 × 1033.233 × 1032.899 × 1032.881 × 1032.865 × 103
Max3.856 × 1033.882 × 1033.095 × 1033.092 × 1033.011 × 103
Std1.543 × 1021.353 × 1025.745 × 1015.943 × 1013.716 × 101
Rank45312
F25Mean4.003 × 1032.915 × 1033.005 × 1032.960 × 1032.888 × 103
Min3.581 × 1032.884 × 1032.930 × 1032.906 × 1032.883 × 103
Max5.042 × 1032.948 × 1033.220 × 1033.044 × 1032.910 × 103
Std3.285 × 1022.514 × 1017.924 × 1013.339 × 1014.344 × 100
Rank52431
F26Mean1.005 × 1046.454 × 1034.605 × 1034.554 × 1034.105 × 103
Min7.873 × 1032.800 × 1034.114 × 1034.051 × 1033.698 × 103
Max1.259 × 1041.021 × 1045.272 × 1035.723 × 1034.894 × 103
Std1.211 × 1032.659 × 1033.458 × 1023.358 × 1022.572 × 102
Rank54231
F27Mean3.833 × 1034.218 × 1033.200 × 1033.200 × 1033.211 × 103
Min3.410 × 1033.644 × 1033.200 × 1033.200 × 1033.201 × 103
Max5.619 × 1034.903 × 1033.200 × 1033.200 × 1033.221 × 103
Std4.177 × 1023.348 × 1022.205 × 10−43.106 × 10−45.428 × 100
Rank45213
F28Mean5.399 × 1033.157 × 1033.315 × 1033.317 × 1033.210 × 103
Min4.195 × 1033.100 × 1033.296 × 1033.296 × 1033.100 × 103
Max6.761 × 1033.265 × 1033.474 × 1033.465 × 1033.267 × 103
Std6.843 × 1026.460 × 1014.611 × 1014.549 × 1013.247 × 101
Rank51342
F29Mean3.507 × 1033.498 × 1032.987 × 1032.963 × 1032.953 × 103
Min3.197 × 1033.312 × 1032.878 × 1032.877 × 1032.860 × 103
Max3.771 × 1033.695 × 1033.111 × 1033.074 × 1032.992 × 103
Std1.371 × 1029.597 × 1016.437 × 1016.320 × 1013.351 × 101
Rank45321
F30Mean4.622 × 1091.995 × 1041.353 × 1073.037 × 1078.942 × 103
Min6.778 × 1077.843 × 1033.555 × 1041.607 × 1045.375 × 103
Max7.022 × 10104.219 × 1042.789 × 1083.440 × 1081.593 × 104
Std1.505 × 10107.370 × 1035.074 × 1077.612 × 1073.023 × 103
Rank52341
Mean rank4.8833.1833.1172.3161.500
Final rank54321
Table 3. The paired samples t-test results.
Table 3. The paired samples t-test results.
FunctionsCS-GWO–PSOCS-GWO–WOACS-GWO–GWOCS-GWO–CSO
t-ValueSig. (2-Tailed)t-ValueSig. (2-Tailed)t-ValueSig. (2-Tailed)t-ValueSig. (2-Tailed)
F1−1.898 × 1016.748 × 10−182.413 × 1002.236 × 10−2−8.712 × 1001.367 × 10−9−9.427 × 1002.472 × 10−10
F2−3.355 × 1019.349 × 10−25−1.603 × 1016.012 × 10−16−7.118 × 1007.828 × 10−8−7.771 × 1001.434 × 10−8
F3−3.733 × 1014.521 × 10−26−1.447 × 1018.498 × 10−15−2.535 × 1001.688 × 10−2−4.732 × 1005.343 × 10−5
F4−1.361 × 1014.016 × 10−145.103 × 1001.907 × 10−5−6.128 × 1001.122 × 10−6−5.836 × 1002.499 × 10−6
F5−2.913 × 1015.005 × 10−23−1.207 × 1017.880 × 10−13−2.016 × 1005.320 × 10−2−4.119 × 1002.892 × 10−4
F6−4.482 × 1012.469 × 10−28−2.925 × 1014.471 × 10−23−1.638 × 1013.381 × 10−16−1.711 × 1011.078 × 10−16
F7−4.000 × 1016.363 × 10−27−1.717 × 1019.852 × 10−17−1.211 × 1002.358 × 10−1−1.954 × 1006.035 × 10−2
F8−1.977 × 1012.240 × 10−18−6.441 × 1004.791 × 10−7−2.143 × 1004.061 × 10−2−3.119 × 1004.073 × 10−3
F9−2.915 × 1014.917 × 10−23−1.070 × 1011.390 × 10−11−8.638 × 1001.639 × 10−9−9.730 × 1001.224 × 10−10
F10−4.268 × 1001.926 × 10−43.982 × 1004.199 × 10−41.263 × 1012.564 × 10−137.920 × 1009.815 × 10−9
F11−1.744 × 1016.491 × 10−17−6.545 × 1003.613 × 10−7−4.355 × 1001.517 × 10−4−5.824 × 1002.584 × 10−6
F12−3.827 × 1006.381 × 10−42.057 × 1004.877 × 10−2−3.849 × 1006.013 × 10−4−5.683 × 1003.811 × 10−6
F13−1.855 × 1007.384 × 10−2−1.045 × 1003.045 × 10−1−1.014 × 1003.188 × 10−1−2.002 × 1005.470 × 10−2
F14−2.698 × 1001.150 × 10−22.856 × 1007.859 × 10−3−2.508 × 1001.800 × 10−2−4.504 × 1001.004 × 10−4
F15−2.870 × 1007.579 × 10−3−2.481 × 1001.915 × 10−2−1.445 × 1001.592 × 10−1−2.437 × 1002.118 × 10−2
F16−6.619 × 1002.964 × 10−7−5.076 × 1002.056 × 10−5−3.658 × 10−17.172 × 10−1−1.120 × 1002.720 × 10−1
F17−1.250 × 1013.371 × 10−13−9.608 × 1001.622 × 10−10−2.962 × 1006.043 × 10−3−4.194 × 1002.354 × 10−4
F18−1.801 × 1012.752 × 10−17−1.777 × 1013.930 × 10−17−4.058 × 1003.412 × 10−4−4.287 × 1001.828 × 10−4
F19−1.886 × 1006.940 × 10−2−1.182 × 1002.468 × 10−1−4.709 × 1005.685 × 10−5−1.200 × 1002.399 × 10−1
F20−7.822 × 1001.258 × 10−8−2.627 × 1001.363 × 10−2−1.677 × 1001.042 × 10−1−2.327 × 1002.715 × 10−2
F21−2.256 × 1016.096 × 10−20−8.681 × 1001.476 × 10−91.939 × 1006.226 × 10−2−6.014 × 10−15.523 × 10−1
F22−1.092 × 1018.593 × 10−12−3.382 × 1002.076 × 10−3−4.182 × 1002.434 × 10−4−4.385 × 1001.395 × 10−4
F23−1.887 × 1017.868 × 10−18−2.395 × 1011.174 × 10−20−5.113 × 1001.853 × 10−5−7.663 × 1001.895 × 10−8
F24−1.967 × 1012.579 × 10−18−2.128 × 1013.048 × 10−192.003 × 1005.460 × 10−2−2.219 × 1003.446 × 10−2
F25−1.859 × 1011.176 × 10−17−5.979 × 1001.688 × 10−6−1.154 × 1012.322 × 10−12−8.043 × 1007.194 × 10−9
F26−2.662 × 1016.231 × 10−22−4.748 × 1005.111 × 10−5−6.544 × 1003.618 × 10−7−6.443 × 1004.764 × 10−7
F27−8.182 × 1005.069 × 10−9−1.640 × 1013.285 × 10−161.129 × 1013.900 × 10−121.129 × 1013.902 × 10−12
F28−1.764 × 1014.810 × 10−174.099 × 1003.058 × 10−4−1.060 × 1011.737 × 10−11−9.310 × 1003.254 × 10−10
F29−2.047 × 1018.731 × 10−19−2.666 × 1016.021 × 10−22−6.827 × 10−15.002 × 10−1−2.368 × 1002.476 × 10−2
F30−1.682 × 1001.033 × 10−1−7.369 × 1004.050 × 10−8−2.184 × 1003.718 × 10−2−1.460 × 1001.550 × 10−1
Table 4. Wilcoxon signed-rank test results of CEC 2017.
Table 4. Wilcoxon signed-rank test results of CEC 2017.
FunctionsCS-GWO vs. PSOCS-GWO vs. WOA
p-ValueR+R−Winnerp-ValueR+R−Winner
F11.734 × 10−60465+6.564 × 10−2322143=
F21.734 × 10−60465+1.734 × 10−60465+
F31.734 × 10−60465+1.734 × 10−60465+
F41.734 × 10−60465+2.163 × 10−543926
F51.734 × 10−60465+1.921 × 10−61464+
F61.734 × 10−60465+1.734 × 10−60465+
F71.734 × 10−60465+1.734 × 10−60465+
F81.734 × 10−60465+2.843 × 10−529436+
F91.734 × 10−60465+1.734 × 10−60465+
F105.287 × 10−464401+9.627 × 10−439372
F111.734 × 10−60465+1.238 × 10−520445+
F121.734 × 10−60465+8.972 × 10−2315150+
F131.734 × 10−60465+1.589 × 10−1164301=
F143.182 × 10−66459+1.150 × 10−442045
F151.734 × 10−60465+3.609 × 10−391374+
F161.734 × 10−60465+5.307 × 10−536429+
F171.734 × 10−60465+2.879 × 10−65460+
F181.734 × 10−60465+1.734 × 10−60465+
F191.734 × 10−60465+1.470 × 10−1162303=
F205.216 × 10−611454+1.480 × 10−2114351+
F211.734 × 10−60465+1.734 × 10−60465+
F222.353 × 10−63462+3.379 × 10−390375+
F231.734 × 10−60465+1.734 × 10−60465+
F241.734 × 10−60465+1.734 × 10−60465+
F251.734 × 10−60465+1.359 × 10−447418+
F261.734 × 10−60465+2.613 × 10−455410+
F271.734 × 10−60465+1.734 × 10−60465+
F281.734 × 10−60465+6.639 × 10−439867
F291.734 × 10−60465+1.734 × 10−60465+
F301.734 × 10−60465+4.286 × 10−69456+
+/=/− 30/0/0 23/3/4
FunctionsCS-GWO vs. GWOCS-GWO vs. CSO
p-ValueR+Rwinnerp-ValueR+Rwinner
F11.734 × 10−60465+1.734 × 10−60465+
F23.182 × 10−66459+1.127 × 10−519446+
F31.359 × 10−447418+3.327 × 10−2129336+
F41.921 × 10−61464+3.882 × 10−68457+
F54.196 × 10−461404+7.190 × 10−2145320=
F61.734 × 10−60465+1.734 × 10−60465+
F71.470 × 10−1162303=4.779 × 10−1198267=
F88.217 × 10−3104361+3.872 × 10−2132333+
F91.734 × 10−60465+1.734 × 10−60465+
F101.238 × 10−5445202.879 × 10−64605
F111.734 × 10−60465+1.734 × 10−60465+
F121.734 × 10−60465+1.734 × 10−60465+
F131.734 × 10−60465+1.734 × 10−60465+
F146.892 × 10−539426+7.865 × 10−2147318=
F151.734 × 10−60465+1.734 × 10−60465+
F163.709 × 10−1189276=7.813 × 10−1219246=
F173.065 × 10−457408+1.319 × 10−2112353+
F186.156 × 10−466399+3.589 × 10−459406+
F191.734 × 10−60465+1.734 × 10−60465+
F202.564 × 10−2124341+1.986 × 10−1170295=
F217.971 × 10−1245220=4.950 × 10−2328137
F226.156 × 10−466399+1.484 × 10−378387+
F232.879 × 10−65460+3.405 × 10−531434+
F245.984 × 10−2141324=3.872 × 10−2333132
F251.734 × 10−60465+1.734 × 10−60465+
F263.112 × 10−530435+4.286 × 10−69456+
F271.734 × 10−646501.734 × 10−64650
F281.734 × 10−60465+1.734 × 10−60465+
F297.190 × 10−2145320=8.130 × 10−1221244=
F301.734 × 10−646501.734 × 10−64650
+/=/− 22/5/3+/=/− 19/6/5
Table 5. The experiment results of case study 2.
Table 5. The experiment results of case study 2.
Prediction ModelRMSE/MWMAE/MWSMAPER2
DA-BiGRU29.05322.8973.5660.937
PSO-DA-BiGRU28.54622.2853.5190.939
WOA-DA-BiGRU28.20922.2213.4710.941
GWO-DA-BiGRU27.89522.1623.5450.942
CSO-DA-BiGRU27.19421.2553.3470.945
CS-GWO-DA-BiGRU26.14420.9633.3370.949
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gong, R.; Li, X. A Short-Term Load Forecasting Model Based on Crisscross Grey Wolf Optimizer and Dual-Stage Attention Mechanism. Energies 2023, 16, 2878. https://doi.org/10.3390/en16062878

AMA Style

Gong R, Li X. A Short-Term Load Forecasting Model Based on Crisscross Grey Wolf Optimizer and Dual-Stage Attention Mechanism. Energies. 2023; 16(6):2878. https://doi.org/10.3390/en16062878

Chicago/Turabian Style

Gong, Renxi, and Xianglong Li. 2023. "A Short-Term Load Forecasting Model Based on Crisscross Grey Wolf Optimizer and Dual-Stage Attention Mechanism" Energies 16, no. 6: 2878. https://doi.org/10.3390/en16062878

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop