Next Article in Journal
Climate Characterization and Energy Efficiency in Container Housing: Analysis and Implications for Container House Design in European Locations
Previous Article in Journal
Correction: Hussain et al. The Causal Connection of Natural Resources and Globalization with Energy Consumption in Top Asian Countries: Evidence from a Nonparametric Causality-in-Quantile Approach. Energies 2020, 13, 2273
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multiscale Hybrid Wind Power Prediction Model Based on Least Squares Support Vector Regression–Regularized Extreme Learning Machine–Multi-Head Attention–Bidirectional Gated Recurrent Unit and Data Decomposition

School of Mechanical Engineering, Shanghai Dianji University, No. 300, Shuihua Road, Pudong New Area District, Shanghai 201306, China
*
Author to whom correspondence should be addressed.
Energies 2024, 17(12), 2923; https://doi.org/10.3390/en17122923
Submission received: 10 April 2024 / Revised: 11 June 2024 / Accepted: 12 June 2024 / Published: 14 June 2024
(This article belongs to the Topic Advances in Wind Energy Technology)

Abstract

:
Ensuring the accuracy of wind power prediction is paramount for the reliable and stable operation of power systems. This study introduces a novel approach aimed at enhancing the precision of wind power prediction through the development of a multiscale hybrid model. This model integrates advanced methodologies including Improved Intrinsic Mode Function with Ensemble Empirical Mode Decomposition with Adaptive Noise (ICEEMDAN), permutation entropy (PE), Least Squares Support Vector Regression (LSSVR), Regularized Extreme Learning Machine (RELM), multi-head attention (MHA), and Bidirectional Gated Recurrent Unit (BiGRU). Firstly, the ICEEMDAN technique is employed to decompose the non-stationary raw wind power data into multiple relatively stable sub-modes, while concurrently utilizing PE to assess the complexity of each sub-mode. Secondly, the dataset is reconstituted into three distinct components as follows: high-frequency, mid-frequency, and low-frequency, to alleviate data complexity. Following this, the LSSVR, RELM, and MHA-BiGRU models are individually applied to predict the high-, mid-, and low-frequency components, respectively. Thirdly, the parameters of the low-frequency prediction model are optimized utilizing the Dung Beetle Optimizer (DBO) algorithm. Ultimately, the predicted results of each component are aggregated to derive the final prediction. The empirical findings illustrate the exceptional predictive performance of the multiscale hybrid model incorporating LSSVR, RELM, and MHA-BiGRU. In comparison with other benchmark models, the proposed model exhibits a reduction in Root Mean Squared Error (RMSE) values of over 10%, conclusively affirming its superior predictive accuracy.

1. Introduction

Wind energy is an environmentally friendly and economically viable form of renewable energy. According to the Global Wind Energy Council (GWEC)’s Global Wind Report 2023, global onshore wind power capacity is projected to exceed 100 GW for the first time by 2024 [1]. However, wind energy data exhibit significant randomness and non-stationarity, which has a substantial impact on the stable operation of power systems. Therefore, accurate prediction of wind power is crucial [2,3,4].
Currently, wind energy forecasting methods are mainly divided into physical models, statistical models, and artificial intelligence models [5]. Physical models primarily consider various sources of geographical information, essentially based on Numerical Weather Prediction (NWP) [6] and high-precision wind farm simulation strategies. This approach has been widely applied in countries such as Spain, Denmark, and Germany, with examples including the SOWIE model developed by Eurowind in Germany [7] and a wind sequence correction algorithm based on NWP proposed by Wang et al. [8]. These algorithms utilize large amounts of data to calculate accurate and reliable wind power predictions. However, because of the large data scale, physical models suffer from slow computational speed and low efficiency and are affected by adverse wind farm conditions, making reliable data collection difficult. Statistical methods, on the other hand, do not consider external conditions such as geography or electrical factors. Their core principle uses the relationships between historical wind power data for prediction in order to improve prediction efficiency [9]. Classical models such as Moving Average (MA) [10] and Autoregressive Integrated Moving Average (ARIMA) [11] are based on modeling linear relationships among data. However, when facing complex patterns, these methods suffer from issues of low prediction accuracy and poor performance.
In comparison with the aforementioned approaches, artificial intelligence models have exhibited increasingly remarkable performance in the domain of wind energy prediction. Methods such as Support Vector Regression (SVR) [12], Extreme Learning Machine (ELM) [13], and Gated Recurrent Unit (GRU) [14] have yielded significant accomplishments in wind energy prediction research. However, because of the considerable prediction errors commonly associated with individual models, hybrid prediction models have gained widespread adoption in recent years. Presently, hybrid models primarily encompass the following three facets: data preprocessing, optimization algorithm tuning, and the prediction of single or combined models. Wang et al. [15] optimized the input weights of ELM using genetic algorithms, whereas Zhai, Ma, and Tan [16,17] utilized the Artificial Fish Swarm Algorithm and Salp Swarm Algorithm to optimize the initial input weights and thresholds of ELM. Their outcomes indicate that these models manifest high prediction accuracy. In recent years, the wind power prediction domain has begun embracing deep learning models such as Long Short-Term Memory (LSTM) [18,19], Temporal Convolutional Neural Network (TCN), and Bidirectional Gated Recurrent Unit (BiGRU) [20]. These models and their derivatives have emerged as principal tools in this sphere. Scholars like W. Wang proposed a prediction methodology based on the fusion of TCN and Light Gradient Boosting Machine (LightGBM) [21]. Researchers such as Chi [22] integrated the attention mechanism into the BiGRU-TCN hybrid model and employed wavelet denoising (WT) processed raw data for prediction. Experimental results corroborate the robust predictive capability of this model. Presently, researchers generally favor hybrid models grounded in intelligent algorithms. Zhang et al. [23] proposed a sparse search algorithm (SSA) to optimize the TCN-BiGRU model and employed the Variational Mode Decomposition (VMD) algorithm to decompose data, thereby mitigating the non-stationarity of wind power data. Ablation experiments demonstrated that this model achieved heightened prediction accuracy compared with scenarios where the SSA algorithm was not employed for parameter optimization. The research by the aforementioned scholars underscores that hybrid models, predicated on algorithmic parameter optimization, can further enhance model prediction accuracy.
Wind power data are inherently characterized by randomness and non-stationarity, necessitating data preprocessing to mitigate prediction errors effectively. To address this, scholars have proposed methodologies grounded in signal decomposition for model formulation. For instance, Gao et al. [24] introduced a composite model combining Empirical Mode Decomposition (EMD) with GRU for prediction tasks. However, the EMD algorithm encounters notable challenges such as mode mixing when confronted with gapped signals. In response, scholars have advocated for the incorporation of uniformly distributed white noise into EMD, manifesting as Ensemble Empirical Mode Decomposition (EEMD) and Complementary Ensemble Empirical Mode Decomposition (CEEMD). This technique has found widespread adoption within the prediction domain. For instance, Torres et al. [25] argued that inadequate decomposition processing frequencies may lead to the persistence of white noise’s influence and the emergence of pseudo-mode phenomena. Consequently, they proposed Complementary Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN), effectively mitigating residual white noise’s impact. Addressing residual noise and pseudo-mode concerns further, Colominas et al. [26] introduced Improved Complementary Ensemble Empirical Mode Decomposition with Adaptive Noise (ICEEMDAN), demonstrating heightened reconstruction accuracy for components and enhanced suitability for nonlinear signal analysis.
In summary, several limitations persist within the realm of wind power prediction. Firstly, prevalent approaches tend to employ single-scale prediction models for all decomposed sub-modes, overlooking the distinctive characteristics of sub-modes across varying frequencies. Secondly, contemporary artificial intelligence prediction methods often encounter challenges in parameter optimization, resulting in high trial and error costs. Lastly, some conventional optimization algorithms exhibit insufficient optimization capabilities and sluggish convergence speeds. Therefore, this paper advocates for a multi-faceted approach by merging data decomposition techniques with multiple models and harnessing the Differential Bees Optimization algorithm (DBO) [27]. This strategy rectifies the existing wind power model shortcomings by addressing deficiencies in sub-mode prediction methods, parameter optimization, and scale singularity.
The main contributions of this paper are as follows:
(1)
This paper proposed a multiscale wind power prediction hybrid model combining data decomposition, LSSVR, RELM, and MHA-BiGRU.
(2)
This paper introduced the ICEEMDAN and PE methods to process the original wind energy series. These methods can effectively address mode mixing and residual noise, thereby better handling the nonlinear and non-stationary characteristics of wind power series.
(3)
This paper introduced the multi-head attention mechanism in the prediction of low-frequency signals, utilizing its strong ability to capture inter-data correlations. Combined with the BiGRU model, this mechanism avoids information loss.
(4)
This paper introduced the DBO optimization algorithm to optimize four parameters including the learning rate, the number of BiGRU neurons, the number of heads in multi-head attention, and the number of filters and regularization parameters. This addresses the limitations and arbitrariness of manual tuning when the MHA-BiGRU model has too many parameters.
(5)
This paper considered the characteristics of information in different frequency bands, used different applicable models, and summed up the results to achieve multiscale hybrid prediction. This overcomes the limitation of insufficient prediction accuracy of a single model.

2. Methods

2.1. Intrinsic Combined Ensemble Empirical Mode Decomposition with Adaptive Noise

Empirical Mode Decomposition (EMD), proposed by Huang et al. in 1998, is a signal processing method suitable for nonlinear and non-stationary processes.
To overcome the limitations of EMD, Colominas et al. further improved CEEMDAN by introducing the Iterative Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (ICEEMDAN) method. The core of this signal decomposition method lies in selecting the Kth IMF component of the white noise decomposed by EMD as the auxiliary noise. Through multiple iterations of noise addition and decomposition, ICEEMDAN comprehensively addresses the randomness and non-stationarity of data, thereby enhancing the stability and reliability of the final decomposition results and reducing the residual noise generated during reconstruction. The computational process is as follows:
  • Based on the original wind power signal s, construct a new sequence s i by adding i groups of white noise ω i to s, resulting in the first group of residues R 1 .
    s i = s + α 0 E 1 ( ω i ) ,   R 1 = M ( s i )
    In Equation (1), E k ( ) represents the k-th mode component generated by EMD decomposition, M ( ) represents the local mean of the signal generated by the EMD algorithm, and represents the overall mean.
  • Calculate the first mode component I IMF 1 = s R 1 iteratively to obtain the k-th group of residues R k and mode component I IMFk .
    R k = M ( R k 1 + α k 1 E k ( ω i ) )
    I IMFk = R k 1 R k
    In Equation (2), α k can be expressed as α k = ε 0 std ( λ ) / std ( E 1 ( ω i ) ) , λ = s ε 0 std ( λ ) , λ = R k , k = 1 , 2 , , K .
  • Repeat step 2 until the calculation is complete to obtain all wind power sequence mode components and the final residue.

2.2. Permutation Entroy

The PE (permutation entropy) algorithm, introduced by Bandt et al., is a method for characterizing the complexity of time series. Its core principle lies in assessing the irregularity in a time series through the examination of permutation patterns within its subsequences. A higher entropy value signifies greater complexity within the time series, whereas a lower value indicates a higher degree of regularity. Applied to the model proposed in this paper, the algorithmic formula is as follows:
  • Consider a wind power generation sequence X = x ( i ) , i = 1 , 2 , , k , where i represent the number one of the wind power sequence.
  • Perform phase space reconstruction on the time series, resulting in a reconstruction matrix Z with a given dimension m and time delay τ .
    Z T = z ( 1 )   z ( 2 )     Z ( k ( m 1 ) τ )
    In Equation (4), z ( j ) = x ( j ) , x ( j + τ ) , , x ( j + ( m 1 ) τ ) .
  • Sort the elements of z ( j ) in ascending order and record the sequence of elements in each row of the reconstruction matrix. Calculate the probability of occurrence for each element sequence to obtain p 1 , p 2 , p q .
  • Define the permutation entropy of wind power sequence X as:
    H m = k = 1 q P k ln P k
    In Equation (5), P k represents the probability of each element size relationship permutation in the reconstruction matrix Z, m is the given dimension, k is the number of subsequences, and q is the total number of elements.

2.3. Bidirectional Gated Recurrent Unit

GRU (Gated Recurrent Unit) is an enhanced version of the Long Short-Term Memory (LSTM) network within the domain of recurrent neural networks. It is tailored to capture long-term dependencies within sequential data while boasting fewer parameters than LSTM, thereby mitigating computational costs. The core principle involves amalgamating the forget gate and input gate into a unified update gate. Through the management of information flow and state updates, it effectively reduces the parameter count and computational overhead. The model’s architectural depiction is presented in Figure 1, and the computational formulas are as follows:
z t = σ W z h t 1 , x t
r t = σ W r h t 1 , x t
h ^ t = tanh W r t h t 1 , x t
h t = 1 z t η t 1 + z k h ^ t
In Equations (6)–(9), z t represents the update gate and r t , x t , h t , and h ^ t represent the reset gate, current time step input information, new cell vector, and hidden state vector, respectively. W z , W r , and W represent the weight matrices.
Because of the strong temporal characteristics inherent in wind power load data, information corresponding to both the previous time step (t − 1) and the current time step (t + 1) significantly impact the prediction results at time t during model training. Consequently, the GRU model fails to fully exploit the inherent information within wind power sequences. In contrast, the Bidirectional Recurrent Neural Network (BiGRU) model addresses this limitation by utilizing both past and future data to enhance prediction accuracy. This effectively overcomes the drawback of low data information utilization observed in the GRU network. Comprising two GRU models, BiGRU possesses the capability to capture bidirectional dependencies within sequential data, thereby enabling it to adapt to more complex sequence patterns. The network structure is visually depicted in Figure 2.

2.4. Multi-Head Attention

The attention mechanism (AM) is a computational method for the efficient allocation of information resources, prioritizing more crucial tasks to effectively mitigate information overload. In AM, input information is represented by key vectors (keys) and value vectors (values), while target information is represented by query vectors (queries). The weights of the value vectors are determined based on the similarity between the query vector and the key vector.
Following this, the final attention values are computed by aggregating the weighted value vectors. The fundamental formula is as follows:
S a t t = W × V
W = f u n c Q , K
In Equations (10) and (11), S a t t is the attention value, Q represents the query vector, K represents the key vector in the key–value pairs, V represents the value vector in the key–value pairs, W represents the weight corresponding to V, and f u n c is the weight transformation function.
The multi-head attention mechanism originates from the Transformer [20] model. Its core principle involves mapping query, key, and value vectors to multiple spaces through distinct linear transformations, followed by calculating the scaled dot-product attention. The computational formula is as follows:
Attention Q , K , V = s o f t m a x Q K T d k V
In Equation (12), d k represents the dimensionality of the keys, which is used to scale the dot product to prevent the issue of gradient vanishing.
Subsequently, vectors Q, K, and V of dimensionality d are transformed into single vectors of dimensionality d / n using different weight matrices W i Q , K i , and Q i . Here, n denotes the number of parallel layers or heads, and these individual vectors are input into corresponding parallel attention layers. Finally, the outputs of each layer are concatenated and fused together using a linear layer to amalgamate all head output results.
As depicted in the multi-head attention section in Figure 3, in our model, we harness the powerful capability of the multi-head attention mechanism to capture diverse temporal scale features in time series, thereby predicting wind power sequences. The mathematical computation formula is as follows:
MHA Q , K , V = Concat head 1 , head 2 , , head n W O
Q i = Q W i Q K i = K W i K i = 1 , 2 , , n V i = V W i V
In Equations (13) and (14), W i Q d × d k , W i K d × d k , W i V d × d v , W O n d v × d , and d k = d v = d / n .
The multi-head attention architecture facilitates leveraging the complexity in input sequences and capturing long-range dependencies, thereby enhancing prediction accuracy. In this study, to mitigate overfitting, the model training employs Mean Squared Error (MSE) as the loss function and utilizes the Adam optimizer for parameter updates.

2.5. MHA-Bidirectional Gated Recurrent Unit

Although BiGRU performs effectively in handling wind power sequences, it lacks the ability to parallelize data processing, resulting in information overload and reduced computational efficiency when dealing with large datasets. To address this limitation, this study combines BiGRU with the multi-head attention mechanism (MHA), as depicted in Figure 3. This integrated model architecture effectively resolves the aforementioned issue. The core process involves utilizing the data trained through the BiGRU hidden layers as input for the MHA network. The decomposed Q, K, and V obtained are then fed into each head for attention value computation. Subsequently, the different results outputted from each channel are feature-weighted and concatenated through a connection layer to form the sequence.

2.6. DBO-MHA-Bidirectional Gated Recurrent Unit

The Dung Beetle Optimizer (DBO) optimization algorithm, introduces a novel swarm intelligence optimization approach. Its primary principle involves simulating five distinct behaviors observed in dung beetles, including rolling, dancing, foraging, stealing, and reproducing, to address optimization problems. Leveraging the DBO algorithm, the MHA-BiGRU model undergoes optimization of parameters such as the learning rate, the number of BiGRU neurons, the number of attention heads, the filter count, and regularization parameters. This optimization ensures the model’s convergence, with the objective of minimizing the loss function. The optimization formula is as follows:
f = min ( t r a i n _ l o s s ( e _ s ,   h _ s ,   n _ l ,   l _ r ,   e p o c h ) )
In Equation (15), e _ s , h _ s , n _ l , l _ r , and epoch represent the number of attention heads, the filter count, the MHA-BiGRU model layers, the learning rate, and the regularization parameter, respectively. train_loss denotes the loss function during the training process.

2.7. Least Squares Support Vector Regression

Least Squares Support Vector Regression (LSSVR), proposed by Vapnik in the early 1990s, is a statistical learning method known for its fast training speed, good generalization performance, and strong ability to fit nonlinear functions.
It particularly excels in handling high-frequency signals, as its core algorithm transforms the solution of a convex quadratic optimization problem into solving a system of linear equations. Consequently, LSSVR requires fewer parameters to train compared with SVR, resulting in faster training speed.
Suppose there is a training set x i , y i i = 1 n , where x i R d represents the input and y i R represents the output. The model calculation formula is as follows:
f ( x ) = w T ϕ ( x ) + b = i = 1 n α i κ ( x i , x ) + b
In Equation (16), w = i = 1 n α i ϕ ( x i ) , w represents the weight vector, ϕ ( ) denotes the non-linear mapping from the input space to the high-dimensional feature space, α i stands for the Lagrange multiplier, κ ( x i , x ) is the kernel function, and b is the bias term.

2.8. Regularized Extreme Learning Machine

Extreme Learning Machine (ELM) [28] is a machine learning algorithm proposed by Professor Huang from Nanyang Technological University in 2004. Its distinguishing feature is a single hidden layer feedback neural network. It can be transformed into solving the generalized inverse problem of the M-P matrix by simply adding a least squares minimum norm problem. Consequently, ELM has fewer model parameters and boasts a fast training speed. It demonstrates excellent capability in handling medium-frequency sequence information.
Consider a wind power sequence set ( X i , t i ) , where the model input is denoted by X i = [ x i 1 , x i 2 , , x i n ] T and the model output by t i = [ t i 1 , t i 2 , , t i n ] T . Then, an ELM network with L hidden layer nodes can be defined as:
i = 1 L β i × g ( ω i x j + b i ) = o j   ,   j = 1 , 2 , , N
In Equation (17), g ( ) represents the activation function, ω i denotes the connection weights between the output layer and the hidden layer, β i signifies the output weights between the hidden layer and the output layer, b i stands for the bias of the i-th hidden unit, and o j represents the network output. The calculation formula used to minimize the output error is as follows: min β H β T F , where H denotes the hidden layer output and T represents the expected output.
To improve the model’s generalization performance, a regularization parameter λ is introduced to solve β , effectively addressing numerical instability issues when computing the pseudo-inverse of H. The computation process of the RELM model, utilizing a regularized least squares method to solve β , is mathematically expressed as follows: min β H β T F + ( 1 / λ ) β F .

2.9. Composition of the Proposed Model

Drawing from the aforementioned methodologies, this study introduces a multiscale hybrid wind power prediction model that integrates ICEEMDAN signal decomposition, permutation entropy (PE) reconstruction, and LSSVR-RELM-MHA-BiGRU. The model parameters are optimized using the DBO optimization algorithm. The overall model workflow, as depicted in Figure 4, is further elucidated with detailed step-by-step explanations as follows:
Step 1: ICEEMDAN decomposes the original wind power data into multiple Intrinsic Mode Function (IMF) components and a residual R. Using permutation entropy (PE), all IMF components are reconstructed to reduce computational complexity. Subsequently, the reconstructed components are categorized into high-frequency, medium-frequency, and low-frequency components based on their PE values.
Step 2: The DBO optimization algorithm is applied to optimize the hyperparameters of the MHA-BiGRU model. The optimized model is then used to predict the IMF low-frequency component after reconstruction.
Step 3: The high-frequency, medium-frequency, and low-frequency components are separately fed into the LSSVR, RELM, and MHA-BiGRU models, respectively. Predictions are obtained for each component.
Step 4: The predictions for the high-, medium-, and low-frequency IMF components are aggregated to obtain the final prediction result.

3. Research Study

3.1. Data Description

The wind power dataset in this paper is sourced from a wind farm operated by Elia in Belgium. To validate the accuracy of the model, four sets of power data from different seasons were selected as the original data for the model. The wind power data were recorded every 15 min. Details of the data are provided in Figure 5 and Table 1. During the data collection process, issues such as missing data and data errors are inevitable, which can significantly affect the accuracy of model predictions. Therefore, this paper adopts the method of removing zeros and mean interpolation to preprocess the data. Each month contains 2880 data points, and each set of data is divided into training and testing sets at a ratio of 7:3.
In Table 1 and Figure 5, it is visually evident that the wind power data exhibit significant randomness and non-stationarity. Because of the large numerical values in the wind power dataset, all data are normalized through scaling to facilitate observing the data characteristics.
Subsequently, the ICEEMDAN method is employed to decompose the wind power sequence into signals (as shown in Figure 6). During decomposition, 50 instances of white noise are added, with a standard deviation (Ntsd) set to 0.2 and a maximum allowable iteration of 100. ICEEMDAN decomposes the original sequence into multiple Intrinsic Mode Function (IMF) sub-sequences. Taking the wind power sequence of March as an example, the decomposition results are illustrated in Figure 6.
As shown in Figure 6, the data are decomposed into multiple IMF components. According to the workflow diagram, in this paper, it is necessary to predict each reconstructed sub-component and then aggregate the prediction results to obtain the final forecast. However, during this process, the computational complexity significantly increases, leading to substantial errors in the prediction results. Therefore, the permutation entropy (PE) method is employed to calculate the entropy value of each sub-component separately, re-quantifying the complexity of each sub-component.
Simultaneously, all sub-components are reconstructed into three new sub-components, including high-frequency, medium-frequency, and low-frequency sub-components, thus reducing the sequence complexity. The PE values of each sub-component are shown in Table 2.
Based on the entropy values, IMF1–IMF3 are selected as high-frequency components, IMF4–IMF9 as medium-frequency components, and IMF10 as the low-frequency component. The newly reconstructed sub-components are illustrated in Figure 7.

3.2. Performance Metrics

To better assess the performance of the model predictions, this study employs three different error evaluation metrics including Root Mean Squared Error (RMSE), Mean Absolute Error (MAE), and Mean Absolute Percentage Error (MAPE). Mean Absolute Percentage Error (MAPE) is used to measure forecast accuracy; Mean Absolute Error (MAE) reflects the actual situation of errors; and the RMSE value reflects the sensitivity to the abnormal value of the wind power sequence.
e RMSE = i = 1 n f x i a x i 2 n
e MAE = 1 n i = 1 n f x i a x i
e MAPE = 1 n i = 1 n f x i a x i a x × 100 %
The above equation represents the calculation of the Mean Absolute Percentage Error (MAPE) metric, where a x i denotes the actual observed value of the i-th training sample, f x i represents the predicted value of the i-th sample, and n represents the total number of samples.

4. Comparative Results

Experiment and Results Analysis
This study employed nine models to forecast wind power sequences for four different seasonal months, as outlined in Table 3. A comparative analysis was conducted between the proposed ICEEMDAN-LSSVR-RELM-DBO-MHA-BiGRU model and other models including LSSVR, RELM, BiGRU, MHA-BiGRU, ICEEMDAN-LSSVR, ICEEMDAN-RELM, ICEEMDAN-MHA-BiGRU, and ICEEMDAN-LSSVR-RELM-MHA-BiGRU. The evaluation was performed using three different metrics including MAE, RMSE, and MAPE. The results were visually presented using bar charts, radar charts, and stacked bar charts, as depicted in Figure 8 and summarized in Table 4, providing a comprehensive comparison of the predictive performance of the models.
Upon careful examination of Table 4 and Figure 8, the following conclusions can be drawn:
  • The comparison of the LSSVR, RELM, and BiGRU models suggests that multiscale hybrid models outperform single-scale models in wind power sequence prediction.
  • The evaluation of the BiGRU and MHA-BiGRU models reveals a notable decrease of 18.72% in RMSE and 7.08% in MAE for the May forecast. Additionally, for the March forecasts, the RMSE values decrease by approximately 10%, indicating that the incorporation of multi-head attention mechanisms enhances predictive accuracy.
  • The inclusion of decomposition algorithms generally enhances predictive performance in wind power prediction compared with single models. For example, in the August metrics, the LSSVR, RELM, and BiGRU models incorporating the ICEEMDAN decomposition algorithm exhibit reductions of 25.57%, 25.29%, and 27.63% in RMSE values, respectively, along with approximately 20% decreases in the other metrics. This underscores the effectiveness of models incorporating decomposition algorithms in improving prediction accuracy.
  • The comparison between Model 8 and the single-scale models combined with decomposition algorithms indicates that hybrid algorithms generally exhibit superior predictive performance. For instance, in the radar chart of MAPE values in Figure 7, the proposed model consistently exhibits the lowest values along its axes. Furthermore, in terms of RMSE values for March, Model 8 demonstrates a 57.20% reduction compared with Model 5. This underscores the advantage of multiscale hybrid algorithms in achieving smaller errors and higher predictive accuracy.
  • The evaluation of Model 8 against the proposed model highlights improvements in predictive performance with the introduction of DBO optimization algorithms. For instance, in the November forecast, there are reductions of 25.59%, 46.40%, and 28.96% in the RMSE, MAE, and MAPE values, respectively. This indicates that the incorporation of DBO optimization algorithms enhances predictive performance, rendering the model proposed in this study more suitable for wind power sequence prediction.
As shown in Figure 6, the data are decomposed into multiple IMF components. According to the workflow diagram, in this paper, it is necessary to predict each reconstructed sub-component and then aggregate the prediction results to obtain the final forecast. However, during this process, the computational complexity significantly increases, leading to substantial errors in the prediction results. Therefore, the permutation entropy (PE) method is employed to calculate the entropy value of each sub-component separately, re-quantifying the complexity in each sub-component.
Simultaneously, all sub-components are reconstructed into three new sub-components, including high-frequency, medium-frequency, and low-frequency sub-components, thus reducing the sequence complexity. The PE values of each sub-component are shown in Table 2. Based on the entropy values, IMF1–IMF3 are selected as high-frequency components, IMF4–IMF9 as medium-frequency components, and IMF10 as the low-frequency component. The newly reconstructed sub-components are illustrated in Figure 7.
To further validate the accuracy of the proposed model, Figure 9 illustrates a comparison of forecasts between the ICEEMDAN-LSSVR-RELM-DBO-MHA-BiGRU model and other models for four months. Figure 10 displays the relative error plots for the BiGRU, MHA-BiGRU, ICEEMDAN-MHA-BiGRU, and ICEEMDAN-LSSVR-RELM-MHA-BiGRU models and the proposed model. Figure 11 presents linear regression plots with 95% confidence intervals for the true values and predictions of these five models.
Upon closer examination of the magnified portions in Figure 9, it is evident that models incorporating decomposition algorithms and multiscale hybrid models exhibit the highest degree of fitting. This underscores the effectiveness of the proposed model. In Figure 10, it can be observed that the proposed method demonstrates the smallest fluctuation range in relative errors. Moreover, in Figure 11, the linear regression confidence band for the proposed model is the narrowest. Particularly, after incorporating attention mechanisms and decomposition algorithms, the scatter plots become more concentrated, and the confidence band notably narrows.
Based on the aforementioned analysis, it can be concluded that the LSSVR-RELM-MHA-BiGRU and data decomposition model proposed in this study exhibits effectiveness and applicability and yields satisfactory prediction results in wind power forecasting.

5. Discussion

This paper aims to improve the accuracy of wind power prediction. Based on the comparison of prediction errors among various models and the proposed model, as well as the verification of methods such as error bands and confidence intervals, it can be concluded that the proposed method indeed improves the accuracy of wind power prediction. These results are based solely on historical data prediction, similar to the methods used to handle historical data in the literature [19,20,21,22,23,24,25], but the accuracy in this paper may be relatively higher. This paper combines machine learning, deep learning, and signal processing methods, thus contributing a new method of artificial intelligence hybrid prediction of wind power. Building on the research of previous scholars, the method introduced in this paper, which incorporates a multi-head attention mechanism and data-denoising technology, may provide some reference for subsequent researchers.
The multiscale hybrid wind power prediction model proposed in this study advances the capability of single-step prediction in wind power forecasting, offering valuable insights for the rational planning of power systems. However, there are avenues for improving prediction methods, which include the following:
  • This study relies solely on historical wind power data without considering additional factors such as geographical conditions and turbine statuses, which can significantly influence wind power prediction accuracy. Thus, future research could explore integrating multiple factors to enable multi-step prediction.
  • The dataset used in this study is limited to a single wind farm, which may restrict the model’s ability to generalize across different environments. Future endeavors should aim to validate the proposed model using data from multiple wind farms to enhance its robustness and applicability.

6. Conclusions

This paper proposes a multiscale hybrid model incorporating a multi-head attention mechanism and data decomposition technique and optimizes the parameters of the low-frequency signal segment of the model using the beetle optimization algorithm. By introducing the data decomposition technique, the most crucial information of the wind power series is extracted, thus eliminating the redundancy in feature selection required by conventional algorithms. On this basis, a multi-head attention mechanism is introduced to extract features of the low-frequency signal close to the time step, further addressing the issue of information loss after data decomposition, thereby limiting the impact of random fluctuations in wind power. The feature correlation results calculated by the sample entropy method can be directly applied to input variables. It avoids the data clustering used in conventional methods. The beetle algorithm is employed to improve prediction accuracy, avoiding the influence of artificially determined network hyperparameters.
The simulation results on datasets demonstrate that the proposed method achieves higher prediction accuracy compared with previous algorithms. This approach provides a new avenue for researchers to improve the prediction accuracy of wind power series by introducing methods such as the multi-head attention mechanism.
However, only single-step prediction is considered, and future work can focus on improving the method for multi-step prediction. Additionally, since most neural networks contain hyperparameters, this approach is theoretically applicable to other neural networks and can be further validated in future studies.

Author Contributions

Writing—original draft, S.Z.; writing—review and editing, Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Shanghai High Peak Plateau Science Project (A1-5101-24-004-05).

Data Availability Statement

The data presented in this study are available on request from the first author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

AbbreviationAcademic name
LSSVRLeast Squares Support Vector Regression
RELMRegularized Extreme Learning Machine
BiGRUBidirectional Gated Recurrent Unit
MHAmulti-head attention
ICEEMDANImproved Complementary Ensemble Empirical Mode Decomposition with Adaptive Noise
DBODung Beetle Optimizer optimization algorithm
PEpermutation entropy
LSTMLong Short-Term Memory

References

  1. Global Wind Energy Council. 2022 Global Wind Report; Global Wind Energy Council (GWEC): Brussels, Belgium, 2023; pp. 6–7. [Google Scholar]
  2. Yuan, X.; Chen, C.; Yuan, Y.; Huang, Y.; Tan, Q. Short-term wind power prediction based on LSSVM–GSA model. Energy Convers. 2015, 101, 393–401. [Google Scholar] [CrossRef]
  3. Hu, S.; Xiang, Y.; Zhang, H.; Xie, S.; Li, J.; Gu, C.; Sun, W.; Liu, J. Hybrid forecasting method for wind power integrating spatial correlation and corrected numerical weather prediction. Appl. Energy 2021, 293, 116951. [Google Scholar] [CrossRef]
  4. Ye, L.; Dai, B.; Li, Z.; Pei, M.; Zhao, Y.; Lu, P. An ensemble method for short-term wind power prediction considering error correction strategy. Appl. Energy 2022, 322, 119475. [Google Scholar] [CrossRef]
  5. Wang, J.; Hu, J.; Ma, K.; Zhang, Y. A self-adaptive hybrid approach for wind speed forecasting. Renew. Energy 2015, 78, 374–385. [Google Scholar] [CrossRef]
  6. Xiong, B.; Meng, X.; Xiong, G.; Ma, H.; Lou, L.; Wang, Z. Multi-branch wind power prediction based on optimized variational mode decomposition. Energy 2022, 8, 11181–11191. [Google Scholar]
  7. Costa, A.; Crespo, A.; Navarro, J.; Lizcano, G.; Madsen, H.; Feitosa, E. A review on the young history of the wind power short-term prediction. Renew. Sustain. Energy 2008, 12, 1725–1744. [Google Scholar] [CrossRef]
  8. Wang, H.; Han, S.; Liu, Y.; Yan, J.; Li, L. Sequence transfer correction algorithm for numerical weather prediction wind speed and its application in a wind power forecasting system. Appl. Energy 2019, 237, 1–10. [Google Scholar] [CrossRef]
  9. Liang, T.; Zhao, Q.; Lv, Q.; Sun, H. A novel wind speed prediction strategy based on Bi-LSTM, MOOFADA and transfer learning for centralized control centers. Energy 2021, 230, 120904. [Google Scholar] [CrossRef]
  10. Erdem, E.; Shi, J. ARMA based approaches for forecasting the tuple of wind speed and direction. Energy 2011, 88, 1405–1414. [Google Scholar] [CrossRef]
  11. Valipour, M.; Banihabib, M.; Behbahani, S. Comparison of the ARMA, ARIMA, and the autoregressive artificial neural network models in forecasting the monthly inflow of Dez dam reservoir. J. Hydrol 2013, 476, 433–441. [Google Scholar] [CrossRef]
  12. Awad, M.; Khanna, R. Support vector regression. In Efficient Learning Machines; Springer: Berlin/Heidelberg, Germany, 2015; pp. 67–80. [Google Scholar]
  13. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the Conference Extreme Learning Machine: A New Learning Scheme of Feedforward Neural Networks, Budapest, Hungary, 25–29 July 2004; Volume 2, pp. 985–990. [Google Scholar]
  14. Li, C.; Tang, G.; Xue, X.; Saeed, A.; Hu, X. Short-term wind speed interval prediction based on ensemble GRU Model. IEEE Trans. Sustain. Energy 2020, 11, 1370–1380. [Google Scholar] [CrossRef]
  15. Wang, X.; Wang, C.; Li, Q. Short-term wind power prediction using GA-ELM. Open Electr. Electron. Eng. J. 2017, 11, 48–56. [Google Scholar] [CrossRef]
  16. Zhai, X.; Ma, L. Medium and long-term wind power prediction based on artificial fish swarm algorithm combined with extreme learning machine. Int. Core J. Eng. 2019, 5, 265–272. [Google Scholar]
  17. Tan, L.; Han, J.; Zhang, H. Ultra-short-term wind power prediction by salp swarm algorithm-based optimizing extreme learning machine. IEEE Access 2020, 8, 44470–44484. [Google Scholar] [CrossRef]
  18. Bouktif, S.; Fiaz, A.; Ouni, A.; Serhani, M.A. Multi-Sequence LSTM-RNN Deep Learning and Metaheuristics for Electric Load Forecasting. Energies 2020, 13, 391. [Google Scholar] [CrossRef]
  19. Gundu, V.; Simon, S.P. PSO–LSTM for short term forecast of heterogeneous time series electricity price signals. J. Ambient. Intell. Hum. Comput. 2021, 12, 2375–2385. [Google Scholar] [CrossRef]
  20. Meng, Y.; Chang, C.; Huo, J.; Zhang, Y.; Mohammed Al-Neshmi, H.M.; Xu, J.; Xie, T. Research on Ultra-Short-Term Prediction Model ofWind Power Based on Attention Mechanism and CNN-BiGRU Combined. Front. Energy Res. 2022, 10, 920835. [Google Scholar] [CrossRef]
  21. Wang, Y.; Chen, J.; Chen, X.; Zeng, X.; Kong, Y.; Sun, S.; Guo, Y.; Liu, Y. Short-term load forecasting for industrial customers based on TCN-LightGBM. IEEE Trans. Power Syst. 2021, 36, 1984–1997. [Google Scholar] [CrossRef]
  22. Chi, D.; Yang, C. Wind power prediction based on WT-BiGRU-attention-TCN model. Front. Energy Res. 2023, 11, 1156007. [Google Scholar] [CrossRef]
  23. Zhang, Y.; Zhang, L.; Sun, D.; Jin, K.; Gu, Y. Short-Term Wind Power Forecasting Based on VMD and a Hybrid SSA-TCN-BiGRU Network. Appl. Sci. 2023, 13, 9888. [Google Scholar] [CrossRef]
  24. Gao, X.; Li, X.; Zhao, B.; Ji, W.; Jing, X.; He, Y. Short-term Electricity Load Forecasting Model based on EMD-GRU with Feature Selection. Energies 2019, 12, 1140. [Google Scholar] [CrossRef]
  25. Colominas, M.A.; Schlotthauer, G.; Torres, M.E.; Flandrin, P. Noise-assisted EMD methods in action. Adv. Adapt. Data Anal. 2012, 4, 1250025. [Google Scholar] [CrossRef]
  26. Colominas, M.A.; Schlotthauer, G.; Torres, M.E. Improved complete ensemble EMD:A suitable tool for biomedical signal processing. Biomed. Signal Process. Control 2014, 14, 19–29. [Google Scholar] [CrossRef]
  27. Xue, J.; Shen, B. Dung beetle optimizer: A new meta-heuristic algorithm for global optimization. J. Supercomput. 2003, 79, 7305–7336. [Google Scholar] [CrossRef]
  28. Xiong, J.; Peng, T.; Tao, Z.; Zhang, C.; Song, S.; Nazir, M.S. A dual-scale deep learning model based on ELM-BiLSTM and improved reptile search algorithm for wind power prediction. Energy 2022, 266, 126419. [Google Scholar] [CrossRef]
Figure 1. Structure of the GRU cell.
Figure 1. Structure of the GRU cell.
Energies 17 02923 g001
Figure 2. Structure of the BiGRU network.
Figure 2. Structure of the BiGRU network.
Energies 17 02923 g002
Figure 3. Structure of the MHA-BiGRU network.
Figure 3. Structure of the MHA-BiGRU network.
Energies 17 02923 g003
Figure 4. Flowchart of the proposed ICEEMDAN-LSSVR-RELM-DBO-MHA-BiGRU model.
Figure 4. Flowchart of the proposed ICEEMDAN-LSSVR-RELM-DBO-MHA-BiGRU model.
Energies 17 02923 g004
Figure 5. Training data and test data for four months.
Figure 5. Training data and test data for four months.
Energies 17 02923 g005
Figure 6. Decomposition diagram of the wind power signal in March.
Figure 6. Decomposition diagram of the wind power signal in March.
Energies 17 02923 g006
Figure 7. Permutation entropy polymerization of the wind power signal in March.
Figure 7. Permutation entropy polymerization of the wind power signal in March.
Energies 17 02923 g007
Figure 8. A comparison of RMSE, MAE, and MAPE values for each model.
Figure 8. A comparison of RMSE, MAE, and MAPE values for each model.
Energies 17 02923 g008
Figure 9. Comparison of four months of wind power forecast model results.
Figure 9. Comparison of four months of wind power forecast model results.
Energies 17 02923 g009
Figure 10. Relative error of wind prediction results for all models.
Figure 10. Relative error of wind prediction results for all models.
Energies 17 02923 g010
Figure 11. Linear regression analysis of predicted values and real values.
Figure 11. Linear regression analysis of predicted values and real values.
Energies 17 02923 g011
Table 1. Basic information of the four monthly wind power datasets.
Table 1. Basic information of the four monthly wind power datasets.
MonthsDatasetData LengthMaxMinMeanStd Dev
MarchAll (MWh)28803780.861.43920.67911.52
Training (MWh)20163780.8616.511146.53976.85
Testing (MWh)8642271.291.43393.69382.49
MayAll (MWh)28803483.900.21862.15918.57
Training (MWh)20163455.840.21784.92869.87
Testing (MWh)8643483.902.181042.351000.30
AugustAll (MWh)28802751.172.53580.44482.30
Training (MWh)20162523.702.53543.60457.46
Testing (MWh)8642751.177.70666.39525.85
NovemberAll (MWh)28804206.606.312082.611361.31
Training (MWh)20164206.6018.082229.311345.26
Testing (MWh)8643810.656.311740.321336.83
Table 2. Permutation entropy of each sample.
Table 2. Permutation entropy of each sample.
ComponentPE
IMF 10.9936
IMF 20.8908
IMF 30.7174
IMF 40.5798
IMF 50.4913
IMF 60.4424
IMF 70.4135
IMF 80.3999
IMF 90.3911
IMF 100.0451
Table 3. Code name for each model.
Table 3. Code name for each model.
NameModel
Model 1LSSVR
Model 2RELM
Model 3BiGRU
Model 4MHA-BiGRU
Model 5ICEEMDAN-LSSVR
Model 6ICEEMDAN-RELM
Model 7ICEEMDAN-MHA-BiGRU
Model 8ICEEMDAN-LSSVR-RELM-MHA-BiGRU
ProposedICEEMDAN-LSSVR-RELM-DBO-MHA-BiGRU
Table 4. Statistical measures of wind power prediction.
Table 4. Statistical measures of wind power prediction.
DatasetModelsRMSEMAEMAPE
MarchModel 10.75630.80211.4089
Model 20.50390.66710.8476
Model 30.46570.43320.6599
Model 40.36050.36241.3988
Model 50.41050.37660.6821
Model 60.47100.56911.2042
Model 70.39330.29680.6470
Model 80.20530.20120.3535
Proposed0.17570.11330.2297
MayModel 10.68520.57111.4820
Model 20.57430.61130.7851
Model 30.56700.50530.5627
Model 40.37980.41820.8501
Model 50.39830.33981.2908
Model 60.34220.22430.3348
Model 70.28620.23311.2788
Model 80.15280.16620.2924
Proposed0.13540.11780.2102
AugustModel 10.62030.77061.675
Model 20.51040.59631.6358
Model 30.49150.47840.85306
Model 40.46880.37961.1223
Model 50.46170.35540.5117
Model 60.38130.30480.4701
Model 70.35570.32910.7367
Model 80.22850.20930.3467
Proposed0.16610.12430.1069
NovemberModel 10.70210.60531.2463
Model 20.61200.58251.2038
Model 30.58110.53510.9602
Model 40.46550.45281.1745
Model 50.38420.34130.8221
Model 60.35220.32271.1934
Model 70.31060.30670.6332
Model 80.19030.20260.3374
Proposed0.14160.10860.2397
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, Y.; Zhang, S. A Multiscale Hybrid Wind Power Prediction Model Based on Least Squares Support Vector Regression–Regularized Extreme Learning Machine–Multi-Head Attention–Bidirectional Gated Recurrent Unit and Data Decomposition. Energies 2024, 17, 2923. https://doi.org/10.3390/en17122923

AMA Style

Sun Y, Zhang S. A Multiscale Hybrid Wind Power Prediction Model Based on Least Squares Support Vector Regression–Regularized Extreme Learning Machine–Multi-Head Attention–Bidirectional Gated Recurrent Unit and Data Decomposition. Energies. 2024; 17(12):2923. https://doi.org/10.3390/en17122923

Chicago/Turabian Style

Sun, Yuan, and Shiyang Zhang. 2024. "A Multiscale Hybrid Wind Power Prediction Model Based on Least Squares Support Vector Regression–Regularized Extreme Learning Machine–Multi-Head Attention–Bidirectional Gated Recurrent Unit and Data Decomposition" Energies 17, no. 12: 2923. https://doi.org/10.3390/en17122923

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop