Next Article in Journal
A Method of Selecting the Number of Coordinates of a Vector Criterion in a Polyoptimal Decision-Making Process
Next Article in Special Issue
Stochastic Extreme Wind Speed Modeling and Bayes Estimation under the Inverse Rayleigh Distribution
Previous Article in Journal
Regenerative Organic Rankine Cycle as Bottoming Cycle of an Industrial Gas Engine: Traditional and Advanced Exergetic Analysis
Previous Article in Special Issue
Flexible Power System Defense Strategies in an Isolated Microgrid System with High Renewable Power Generation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Short-Term Wind Speed Prediction Based on Principal Component Analysis and LSTM

1
School of Mechanical and Automotive Engineering, Shanghai University of Engineering Science, Shanghai 201620, China
2
Department of Electrical and Computer Engineering, Kansas State University, Manhattan, KS 66506, USA
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(13), 4416; https://doi.org/10.3390/app10134416
Submission received: 30 May 2020 / Revised: 23 June 2020 / Accepted: 25 June 2020 / Published: 27 June 2020
(This article belongs to the Special Issue Wind Power Systems: Design, Operation, and Control)

Abstract

:
An accurate prediction of wind speed is crucial for the economic and resilient operation of power systems with a high penetration level of wind power. Meteorological information such as temperature, humidity, air pressure, and wind level has a significant influence on wind speed, which makes it difficult to predict wind speed accurately. This paper proposes a wind speed prediction method through an effective combination of principal component analysis (PCA) and long short-term memory (LSTM) network. Firstly, PCA is employed to reduce the dimensions of the original multidimensional meteorological data which affect the wind speed. Further, differential evolution (DE) algorithm is presented to optimize the learning rate, number of hidden layer nodes, and batch size of the LSTM network. Finally, the reduced feature data from PCA and the wind speed data are merged together as an input to the LSTM network for wind speed prediction. In order to show the merits of the proposed method, several prevailing prediction methods, such as Gaussian process regression (GPR), support vector regression (SVR), recurrent neural network (RNN), and other forecasting techniques, are introduced for comparative purposes. Numerical results show that the proposed method performs best in prediction accuracy.

1. Introduction

As one of the clean and renewable energy sources, wind power has developed rapidly all over the world during the last decade. In 2018, the global installed wind power capacity was 592 GW, which is expected to increase to 800 GW by the end of 2021 [1]. Wind speed is the most important factor affecting wind power generation [2]. Variability and uncertainty stemming from noncontrollable and nonadjustable wind speed bring tremendous difficulties to large-scale wind power integration and operation in power systems. A more accurate wind speed prediction can help reduce the negative impact of wind power integration and improve the efficacy and stability of power system operations [3,4,5].
To date, there have been three noteworthy approaches to wind speed prediction. The first approach is with respect to physical methods such as numerical weather prediction (NWP) models [6,7], which primarily use mathematical models of the atmosphere and oceans to obtain the estimates of wind speed forecasts. NWP is typically a physical model with the advantages of high precision and strong basis. However, there exist a variety of challenges facing NWP, such as the difficulty of collecting meteorological data, the requirement of large-scale computing resources, and so on.
The second approach is statistical methods, which make use of historical and measured data to establish input and output function models. These methods include the Kalman filter [8], autoregressive integrated moving average model (ARIMA), support vector regression (SVR), Gaussian regression, and grey prediction method [9,10,11,12]. Cassola et al. [13] verified that the Kalman filter can significantly improve the output results of the model by adjusting the time step and prediction range of the filter, especially in a short-term prediction. Aasim et al. [14] proposed an ARIMA model based on repeated wavelet transform and compared them on different time scales to show its superiority.
Finally, the third approach is to employ machine learning methods to establish a model between inputs and outputs, which is similar to the structure of synapse connection based on neural networks. Machine learning has been successfully applied to many domains such as cloud services [15,16,17], context classification [18], and wind power or wind speed forecasting [19,20,21,22]. Different machine learning methods have an apparent influence on forecasting performance. The nonlinear can lead to poor forecasting results when an unsuitable machine learning method is utilized. The back-propagation (BP) neural network is a typical artificial neural network. Zhang et al. [23] proposed a wind speed prediction method based on Lorentz perturbation to optimize the weights and thresholds of the BP neural network by means of an improved particle swarm optimization (PSO) algorithm. Liu et al. [24] proposed a wind speed prediction method based on singular spectrum analysis, convolutional neural network method, gated recurrent unit method, and support vector regression (SSA-CNNGRU-SVR). Deep learning has been widely applied to wind speed prediction recently. Ma et al. used the negative correlation neural network model optimized by the PSO algorithm to predict wind speed [25]. Pei et al. [26] adopted wavelet transform to process the original data of wind speed, and proposed a new cell update long short-term memory (LSTM) network to predict the future wind speed. Zhang et al. [27] presented an LSTM model with shared weights to forecast wind speed. Khosravi et al. [28] used a method of combined multilayer feed-forward neural network (MLFFNN), fuzzy inference system (FIS), and SVR to predict wind speed. An approach of combined multilayer perceptron neural network and extreme learning machine was presented to predict short-term wind speed [29]. A combined deep learning neural network based on convolutional neural network (CNN) and radial basis neural network was presented to predict wind power generation for the next day [30]. The above approaches discussed a variety of methods to show the improvement in modeling techniques for wind speed or power prediction. Nevertheless, these approaches ignore the fact that the wind speed is highly influenced by meteorological characteristics such as temperature, air pressure, wind speed, and humidity. Therefore, it is not sufficient and realistic to formulate prediction models only using the historical data of wind speed. It has been demonstrated that using meteorological features for wind speed prediction can improve its prediction accuracy [31,32,33]. A nonlinear autoregressive wind speed prediction model was proposed in [31], which employed meteorological time series as inputs. Fan et al. [32] put forward a wind farm prediction model based on dynamic Bayesian clustering of meteorological features and support vector regression. Factors such as wind direction, temperature, humidity, and altitude were used as model inputs. Zhang et al. [33] established a causal relationship between wind speed and meteorological factors to propose an LSTM network based on neighborhood gates structure; however, the hyperparameters of the LSTM networks are chosen empirically.
Wind speed prediction considering meteorological features usually involves a huge amount of high-dimensional data with massive redundant information, which leads to inaccurate prediction [34,35]. It is crucial to eliminate the redundant information in the features while reducing the complexity of prediction models. A recent study presented an effective principal component analysis (PCA)-based decision tree machine learning classification technique [36]. Following this framework, this paper proposes a hybrid wind speed prediction method based on PCA and LSTM networks. Specifically, a PCA algorithm is employed to process the original data for reducing the dimensions and redundant information. A differential evolution (DE) algorithm is presented to determine hyperparameters for LSTM networks. The main components processed by PCA and historical wind speed data are used as the input of LSTM. The main contributions of this work in comparison with the existing literature are as follows:
  • A wind speed prediction algorithm considering meteorological features based on PCA and LSTM networks is presented. DE as a hyperparameter selection method is also included in the proposed method.
  • The PCA preprocessing method can effectively reduce the dimensions and retain the features in the data, which lays an important foundation for more accurate prediction with improved computational efficiency.
  • The proposed method is validated on three different cases considering real-world data, and experimental results show that the proposed method outperforms other popular forecasting methods.
The rest of this paper is organized as follows. The proposed hybrid prediction approach is described in Section 2. The experimental design and numerical validation are shown in Section 3. Conclusions and future work are provided in Section 4.

2. Methodology

In this section, the underlying theories and developed method are described, including the PCA algorithm, LSTM networks, and the hyperparameter selection based on DE, as well as the prediction framework proposed in this paper.

2.1. PCA Algorithm

In many cases, there is a correlated relationship among variables, which makes the problem under study very complicated. When there is a certain correlation between two variables, it can be explained that the two variables reflecting the information of this problem have a certain overlap. PCA is devised to delete redundant (closely related) variables and to establish as few new variables as possible, so that these new variables are irrelevant [37].
In the PCA algorithm, orthogonal transformation is used to convert a set of variables that may have linear correlation into a set of linearly uncorrelated variables. The converted variables are used as the principal components. Take a data matrix of M × N as shown in Equation (1), for example; it means that there are N samples, and each sample has M features.
T = [ d 11 d 12 d 1 N d 21 d 22 d 2 N d M 1 d M 2 d M N ]
The calculation process of the PCA algorithm to reduce the dimensions of T is as follows:
(1)
Calculate the covariance matrix of T by Equation (2):
Q = n = 1 N ( T n T ¯ ) ( T n T ¯ ) N 1
where Q is the covariance matrix; T n represents the nth sample vector; T ¯ denotes the mean value of the nth sample; and ( T n T ¯ ) is the transposed matrix.
(2)
Calculate the eigenvalues and eigenvectors of Q by Equation (3):
Q = m = 1 M λ m v m v m = V Λ V
where v m is the mth eigenvector of Q and λ m is the mth eigenvalue of Q . V and Λ are composed of eigenvectors and eigenvalues of Q , respectively.
(3)
Arrange the eigenvalues from large to small and then calculate the contribution rate of each feature and cumulative contribution rate of all features by Equations (4) and (5) as follows:
p l = λ l / l = 1 M λ l
p ¯ = l = 1 M p l
where p l is the contribution rate of the lth component, λ l is the lth eigenvalue arranged from large to small. p ¯ represents the cumulative contribution rate.
(4)
According to (3), select I (I ≤ M) components which contain the most information of T from M components. The eigenvectors corresponding to the selected components constitute the transformation matrix U. The reduced dimension matrix Z is obtained by multiplying the original data matrix T and the transformation matrix U as described in Equation (6)
Z = T U = T ( V 1 , V 2 , , V I )
where Z is the reduced dimension matrix m. U represents the transformation matrix.
In this paper, up to 11 meteorological characteristics affecting wind speed are collected, including air temperature, air pressure, humidity, and so forth. The PCA algorithm above is employed to reduce the dimensions of the meteorological data. The data after dimensionality reduction can effectively keep the original meteorological information as much as possible.

2.2. LSTM and Hyperparameter Selection

2.2.1. LSTM

As a special kind of recurrent neural network (RNN), the LSTM neural network was first proposed by Hochreiter and Schmidhuber [38]. According to the LSTM structure in Figure 1, the current state of a cell will be affected by the previous cell state, which reflects the recurrent characteristics of LSTM. Based on RNN, the candidate cell, forget gate, input gate, and output gate are added to the hidden layer of LSTM. LSTM with such a structure does not cause the gradient to disappear or explode and can learn the information contained in time series data more effectively [39]. An LSTM unit composed of a candidate cell, an input gate, an output gate, and a forget gate is shown in Figure 2. The input gate controls the extent of the values which flow into the cell. The forget gate controls the extent of the values which remain in the cell. The output gate and the value in the cell determine the output of an LSTM unit [39].
The calculation process of LSTM in Figure 2 is as follows:
(1)
Calculate the inputs for three gate units and the candidate cell by Equations (7)–(10):
n e t i , t = W i [ x t , h t 1 ] + b i
n e t f , t = W f [ x t , h t 1 ] + b f
n e t o , t = W o [ x t , h t 1 ] + b o
n e t c ˜ , t = W c ˜ [ x t , h t 1 ] + b c ˜
where x t represents the inputs at time t , and h t 1 represents the cell output at time t 1 . n e t i , t n e t f , t n e t o , t n e t c ˜ , t are input gate, forget gate, output gate, and the candidate cell, respectively. W i , W f , W o , W c ˜ are the input gate, forget gate, output gate, and weight matrix of the candidate cell, respectively. b i , b f , b o , b c ˜ are the bias of input gate, forget gate, output gate, and candidate cell, respectively.
(2)
Calculate the three gate units by Equations (11)–(15):
i t = σ ( n e t i , t )
f t = σ ( n e t f , t )
C ˜ t = tan h ( n e t c ˜ , t )
o t = σ ( n e t o , t )
C t = f t C t 1 + i t C ˜ t
where i t , f t , C t , o t , c ˜ t represent the input gate at time t, forget gate, unit state, output gate, and candidate cell vector unit output, respectively. σ ( ) stands for sigmoid activation function and is expressed by Equation (16):
σ ( x ) = 1 1 + e x
tanh(·) stands for tanh activation function and is expressed by Equation (17):
tan ( x ) = e x e x e x + e x
(3)
Calculate the output by Equation (18):
h t = o t tan h ( C t )
where h t is the unit output at time t .

2.2.2. Selection of Hyperparameters

Many parameters for LSTM can affect its accuracy and performance. The selected hyperparameters are a learning rate, the number of units of the hidden layer, and the number of batch size. If the selected learning rate is too small, the convergence will be too slow; otherwise, the cost function will oscillate. The number of units of the hidden layer will influence the effect of fitting. For the number of batch size, if this number is too small, then the training data will be extremely difficult to converge, which will lead to underfitting. If the number is too large, then the required memory will increase significantly. For example, when the number of units of the hidden layer is specified within (1, 100) and the range of the number of batch size is specified within (1, 500), a total of 50,000 combinations will be generated. Thus, to overcome the computational burden, a simple yet reliable algorithm should be utilized to select the optimal combination of parameters for balancing predictive performance and computational efficiency. In this paper, the hyperparameters of LSTM are determined by means of the DE algorithm. DE is a heuristic random search algorithm based on group differences [40]. The objective function of the optimization problem to select hyperparameters is the root mean square error (RMSE), which represents the sum of the squared deviations of the predicted value and the true value calculated by Equation (19):
R M S E = 1 s i = 1 s ( y ˜ i y i ) 2
where y ˜ s is the sth predicted value, y s is the sth true value.
The process of the DE algorithm to select LSTM hyperparameters is shown below:
(1)
Initialization: Initialize the following parameters: length of individual D, number of iterations G, population size NP, crossover rate, and scaling or mutation factor. The population is randomly generated by Equation (20):
X ω , k = X ω , k L + r a n d ( X ω , k U X ω , k L )
where ω = 1, 2, …, NP; k = 1, 2, …D; X ω , k L and X ω , k U are the upper and lower bounds of the k-th dimension, respectively.
(2)
Mutation: Mutation operator is used to generate the mutation vector (Hi) for the individual of the population using Equation (21).
H i ( g + 1 ) = X r 1 ( g ) + F ( X r 2 ( g ) X r 3 ( g ) )
where x r 1 ( g ) , x r 2 ( g ) , x r 3 ( g ) are individuals randomly selected from the population, and r 1 r 2 r 3 , F is the scaling factor and g represents the g-th generation.
(3)
Crossover: Crossover operation is to randomly select individuals using Equation (22)
U ω , k ( g + 1 ) = { H ω , k ( g + 1 ) i f   r a n d ( 0 , 1 ) C R X ω , k ( g ) otherwise
where U ω , k is the new individual generated in the crossover operation, C R is the crossover rate.
(4)
Selection: In the selection operator as shown in Equation (23), for minimization problems, if the fitness value f ( U ω , k ( g + 1 ) ) of the trial vector U ω , k ( g + 1 ) is less than or equal to the fitness value f ( X ω , k ( g ) ) of the target vector X ω , k ( g ) , the trial vector will replace the target vector and enter the population. Otherwise, the target vector is still retained.
X ω , k ( g + 1 ) = { U ω , k ( g + 1 ) i f   f ( U ω , k ( g + 1 ) f ( X ω , k ( g ) ) X ω , k ( g )
The best individual is the three hyperparameters of the LSTM prediction network structure. In this paper, the best set of hyperparameters is selected by comparing the RMSE corresponding to different hyperparameter values generated by training.

2.3. Proposed Prediction Framework

The framework of the proposed prediction algorithm, shown in Figure 3, includes three parts. They are Part A (Data processing), Part B (Hyperparameters optimization), and Part C (Forecasting).
In Part A, 11 types of meteorological data (feature data in Part A), such as temperature, humidity, air pressure, and wind level, are dimensionally reduced by PCA. The data processed by the PCA algorithm and historical wind speed data together form the input of Part B. Part B uses the hyperparameters selected by DE to obtain a new LSTM model. Part C first divides the processed data in part A into a training set and a test set, and then applies the LSTM model in part B to forecast wind speed.

2.4. Evaluation Metric of Prediction

In order to evaluate the performance of the proposed model, four different indicators, including RMSE (see Equation (19)), mean absolute error (MAE), mean absolute percentage error (MAPE), and coefficient of determination (denoted by R 2 ), are adopted as evaluation metrics.
(1) MAE
MAE is the average of the absolute error and can reflect the error between the predicted value and the actual value well. The smaller the MAE is, the higher the accuracy the prediction achieves. The formula is as follows.
M A E = 1 s i = 1 n | y ˜ i y i |
(2) MAPE
MAPE represents the ratio between the error and the true value. The smaller the MAPE is, the closer the predicted value is to the true value. The formula is as follows.
M A P E = 1 s i = 1 s | y ˜ i y i | y i
(3) Coefficient of determination
The coefficient of determination ( R 2 ) is the square of the correlation between predicted values and true values. The range of R 2 is specified within [0,1]. The closer R 2 is to 1, the more perfect fit the prediction model has. Therefore, R 2 can be used as an important indicator. The formula is as follows.
R 2 = i = 1 s ( y ˜ i Y ¯ ) 2 i = 1 s ( y i Y ¯ ) 2
where Y ¯ is the mean of the observed value.

3. Case Study

In this section, the description of meteorological data is first introduced. Then, the experimental design and parameter settings are described. Finally, results and analysis are shown to validate the performance of the proposed method.

3.1. Data Description

The meteorological data came from Fuyun Meteorological Station (N 46.59°, E 89.31°) located at Xinjiang province in China [33]. The time span was one month (from 00:00 on the 15th of July, 2018 to 23:00 on the 14th of August). The total length of the sample was 744 with an hourly resolution. The basic information of meteorological factors is listed in Table 1. Hourly wind speed data shown in Figure 4 were divided into a training set (the first 624 points) and a test set (the last 120 points).

3.2. Experimental Design and Parameter Settings

We first used the PCA algorithm to calculate the contribution and cumulative contribution rate of each component as shown in Table 2. It can be seen that as the number of features increased, the correlation became more and more obvious, which means that there was no need to measure all the features. The cumulative contribution rate of the first five principal components was 0.9653. Therefore, these five principal components and wind speed were selected to form the input to the prediction models.
To validate the prediction performance of the proposed PCA-LSTM method, BPNN-, SVR-, GPR-, and RNN-based methods were selected for comparison. The following three cases were studied:
Case I: LSTM model optimized by DE algorithm and comparisons with BPNN, GPR, RNN, and SVR only using historical wind speed data
Case II: Feature-LSTM model optimized by DE algorithm and comparisons with Feature-BPNN, Feature-GPR, Feature-RNN, and Feature-SVR using all meteorological factors related to wind speed data
Case III: the proposed model and comparisons with PCA-BPNN, PCA-GPR, PCA-RNN, and PCA-SVR using meteorological factors processed by the PCA algorithm.
On the basis of several trials and similar works [40,41], the parameters of the DE algorithm were set as follows: population size NP = 10, number of iterations G = 20, scaling factor F = 0.6, and crossover factor CR = 0.8. For the LSTM network, the range of the learning rate was specified within [0, 1], the range of the number of hidden layer units was within [1, 100], and the range of the batch size was specified within [1, 100]. Table 3 shows parameter settings of LSTM determined by the DE algorithm and other models. Furthermore, the Adam algorithm was adopted to make calculation of the LSTM network more efficient [40]. For SVR and GPR, the results of each run were the same, hence, they ran once. For network-based methods, such as BPNN, LSTM, and RNN, the results obtained by each run were different; these models ran independently for 10 times, and the average was recorded as the final result.

3.3. Prediction Result Analysis

Case I was to demonstrate the effectiveness of the LSTM model optimized by the DE algorithm compared with the BPNN, GPR, RNN, and SVR models. The forecasting indicators obtained from the above models are shown in Table 4, where the best results for each model are in bold. For the LSTM model, the average values of RMSE, MAE, MAPE, and R 2 were 0.3327, 0.2598, 0.1004, and 0.9655, respectively. Compared with other models, the LSTM model optimized by DE performed the best in RMSE, MAE, and R 2 . In order to more intuitively compare the above five models, the forecasting wind speed and R 2 for each model are shown in Figure 5. The coefficient of determination of the LSTM model was the largest. So, the fitting effect was the best and the prediction was the most accurate. To summarize, the LSTM model optimized by the DE algorithm provided better performance than the other four traditional models.
In Case II, we used meteorological characteristics related to wind speed as input of the five forecasting models. Table 5 and Figure 6 show the performance metrics of forecasting results achieved by the above five different models. RMSE, MAPE, and MAE of the Feature-LSTM model were 0.1745, 0.0488, and 0.1212, respectively. R 2 of the Feature-LSTM was 0.9749, which was slightly smaller than that of Feature-BPNN. Overall, the comprehensive prediction performance of the Feature-LSTM was the best among all the above models. Compared with Case 1, all four indicators improved significantly.
The prediction results based on the PCA and LSTM methods of Case III are shown in Figure 7. The figure shows the true values and the predicted values of the five different models. Table 6 is the evaluation index value of each prediction model. Through comparing PCA-LSTM with the PCA-BPNN, PCA-GPR, PCA-RNN, and PCA-SVR models, it can be clearly observed that combined methods had an apparent influence on forecasting performance. From Table 6 and Figure 7, the proposed model outperformed the other four competitors for short-term wind speed forecasting with the smallest mean value of RMSE as 0.1474, MAPE as 0.0382, and MAE as 0.1015, as well as the highest mean value of R 2 as 0.9989, making it the best among all prediction models. Therefore, compared with Case I and Case II, the four indicators of the proposed PCA-LSTM method in this article were the best. The PCA-LSTM model achieved superiority in the 15 forecasting models in all cases.

3.4. Discussion

In the above experiments, the BPNN, GPR, LSTM, RNN, and SVR methods were selected in Case I to predict wind speed using only historical wind speed data. Among them, the comprehensive performance of LSTM was the best, which reflects the strong fitting ability of the LSTM model to the nonlinear problem. By comparing Case I with Case II, the use of meteorological characteristics related to wind speed during wind speed prediction can improve the prediction results. In the comparison between Case II and Case III, the PCA algorithm was used to reduce the dimension of meteorological characteristics related to wind speed, and the prediction results were further improved in all prediction models.

4. Conclusions

In this paper, a hybrid PCA and LSTM prediction method is presented. PCA is used to process original meteorological data. LSTM is optimized by the DE algorithm to obtain the best prediction model. Combining PCA and LSTM shows great advantages. The proposed method is applied to predict wind speed and the results prove that the method has strong predictive ability for time series data. Based on the analyses of Cases I–III, the proposed model not only requires less data than other models, but also largely improves the accuracy of forecasting results.
In our future work, hybrid methods using different deep learning models will be considered for time series prediction. In addition, we will improve the PCA method to make it more applicable and efficient.

Author Contributions

D.G. and H.Z. proposed the algorithm. D.G. implemented the experiments and wrote the manuscript. H.Z. supervised the study and revised the manuscript. H.W. co-supervised the study. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China under Grant No. 61703268, and Open Project Program of Shanghai Key Laboratory of Intelligent Manufacturing and Robotics under Grant No. zk1703.

Acknowledgments

The authors would like to thank the anonymous reviewers and the editor for their valuable comments which are extremely helpful for improving the quality of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dyrholm, M.; Rebollo, A.; Backwell, B.; Aufderheide, B.; Ohlenforst, K. Global Wind Report 2018; Global Wind Energy Council (GWEC): Brussels, Belgium, 2019. [Google Scholar]
  2. Yang, X.Y.; Zhang, Y.F.; Yang, Y.W.; Lv, W. Deterministic and Probabilistic Wind Power Forecasting Based on Bi-Level Convolutional Neural Network and Particle Swarm Optimization. Appl. Sci. 2019, 9, 1794. [Google Scholar] [CrossRef] [Green Version]
  3. Zhang, S.H.; Liu, Y.W.; Wang, J.Z.; Wang, C. Research on Combined Model Based on multi-objective optimization and application in wind speed forecast. Appl. Sci. 2019, 9, 423. [Google Scholar] [CrossRef] [Green Version]
  4. Jaesung, J.; Robert, P.B. Current status and future advances for wind speed and power forecasting. Renew. Sustain. Energy Rev. 2014, 31, 762–777. [Google Scholar]
  5. Wang, Q.; Wu, H.; Florita, A.; Martinez, C.B.; Hodge, B.M. The value of improved wind power forecasting: Grid flexibility quantification, ramp capability analysis, and impacts of electricity market operation timescales. Appl. Energy 2016, 184, 696–713. [Google Scholar] [CrossRef] [Green Version]
  6. Chen, N.Y.; Zheng, Q.; Nabney, L.T.; Meng, X.F. Wind power forecasting using Gaussian Processes and Numerical weather prediction. IEEE Trans. Power Syst. 2014, 29, 656–665. [Google Scholar] [CrossRef] [Green Version]
  7. Piotrowski, P.; Baczyński, D.; Kopyt, M.; Szafranek, K.; Helt, P.; Gulczyński, T. Analysis of forecasted meteorological data (NWP) for efficient spatial forecasting of wind power generation. Electr. Power Syst. Res. 2019, 175, 105891. [Google Scholar] [CrossRef]
  8. Marta, P.; Pilar, P.; José, R.P. Automatic tuning of Kalman filters by maximum likelihood methods for wind energy forecasting. Appl. Energy 2013, 108, 349–362. [Google Scholar]
  9. Erdem, E.; Shi, J. ARMA based approaches for forecasting the tuple of wind speed and direction. Appl. Energy 2011, 88, 1405–1414. [Google Scholar] [CrossRef]
  10. Zhou, J.Y.; Shi, J.; Li, G. Fine tuning support vector machines for short-term wind speed forecasting. Energ. Convers. Manag. 2011, 52, 1990–1998. [Google Scholar] [CrossRef]
  11. Zhang, C.; Wang, H.K.; Zhao, X.; Liu, T.H.; Zhang, K.J. A Gaussian process regression based hybrid approach for short-term wind speed prediction. Energy Convers. Manag. 2016, 126, 1084–1092. [Google Scholar] [CrossRef]
  12. Xia, J.; Wu, W.Q.; Huang, B.L.; Li, W.P. Application of a new information priority accumulated grey model with time power to predict short-term wind turbine capacity. J. Clean. Prod. 2020, 244, 118573. [Google Scholar] [CrossRef] [Green Version]
  13. Cassiola, F.; Burlando, M. Wind speed and wind energy forecast through Kalman filtering of Numerical Weather Prediction model output. Appl. Energy 2012, 99, 154–166. [Google Scholar] [CrossRef]
  14. Singh, S.N.; Mohapatra, A. Repeated wavelet transform based ARIMA model for very short-term wind speed. Renew. Energy 2019, 136, 758–768. [Google Scholar]
  15. Rjoub, G.; Bentahar, J.; Wahab, O.A.; Bataineh, A. Deep Smart Scheduling: A Deep Learning Approach for Automated Big Data Scheduling Over the Cloud. In Proceedings of the 2019 7th International Conference on Future Internet of Things and Cloud (FiCloud), Istanbul, Turkey, 26–28 August 2019; pp. 189–196. [Google Scholar]
  16. Wahab, O.A.; Bentahar, J.; Otrok, H.; Mourad, A. Resource-Aware Detection and Defense System Against Multi-Type Attacks in the Cloud: Repeated Bayesian Stackelberg Game. IEEE Trans. Dependable Secur. Comput. 2019, 2907946. [Google Scholar] [CrossRef]
  17. Wahab, O.A.; Cohen, R.; Bentahar, J.; Otrok, H.; Mourad, A.; Rjoub, G. An endorsement-based trust bootstrapping approach for newcomer cloud services. Inf. Sci. 2020, 527, 159–175. [Google Scholar] [CrossRef]
  18. Sarker, I.H.; Kayes, A.S.M.; Watters, P. Effectiveness Analysis of Machine Learning Classification Models for Predicting Personalized Context-Aware Smartphone Usage. J. Big Data 2019, 6, 57. [Google Scholar] [CrossRef]
  19. Zameer, A.; Khan, A.; Javed, S.G. Machine Learning based short term wind power prediction using a hybrid learning model. Comput. Electr. Eng. 2015, 45, 122–133. [Google Scholar]
  20. Heinermann, J.; Kramer, O. Machine learning ensembles for wind power prediction. Renew. Energy 2016, 89, 671–679. [Google Scholar] [CrossRef]
  21. Treiber, N.A.; Heinermann, J.; Kramer, O. Wind power prediction with Machine Learning. Comput. Sustain. 2016, 645, 13–29. [Google Scholar]
  22. Hu, Y.L.; Chen, L. A nonlinear hybrid wind speed forecasting model using LSTM network, hysteretic ELM and Differential Evolution algorithm. Energy Convers. Manag. 2018, 173, 123–142. [Google Scholar] [CrossRef]
  23. Zhang, Y.G.; Chen, B.; Zhao, Y.; Pan, G. Wind Speed Prediction of IPSO-BP Neural Network Based on Lorenz Disturbance. IEEE Access 2018, 6, 53168–53179. [Google Scholar] [CrossRef]
  24. Liu, H.; Mi, X.W.; Li, Y.F.; Duan, Z.; Xu, Y.N. Smart wind speed deep learning based multi-step forecasting model using singular spectrum analysis, convolutional gated recurrent unit network and support vector regression. Renew. Energy 2019, 143, 842–854. [Google Scholar] [CrossRef]
  25. Ma, T.; Wang, C.; Wang, J.Z.; Cheng, J.J.; Chen, X.Y. Particle-swarm optimization of ensemble neural network with negative correlation learning for forecasting short-term wind speed of wind farms in western China. Inf. Sci. 2019, 505, 157–182. [Google Scholar] [CrossRef]
  26. Pei, S.Q.; Qin, H.; Zhang, Z.D.; Yao, L.Q.; Wang, Y.Q.; Wang, C.; Liu, Y.Q.; Jiang, Z.Q.; Zhou, J.Z.; Yi, T.L. Wind speed prediction method based on Empirical Wavelet Transform and New Cell Update Long Short-Term Memory network. Energy Convers. Manag. 2019, 196, 779–792. [Google Scholar] [CrossRef]
  27. Zhang, Z.D.; Ye, L.; Qin, H.; Liu, Y.Q.; Wang, C.; Yu, X.; Yin, X.L.; Li, J. Wind speed prediction method using Shared Weight Long Short-Term Memory Network and Gaussian Process Regression. Appl. Energy 2019, 247, 270–284. [Google Scholar] [CrossRef]
  28. Khosravi, A.; Machado, L.; Nunes, R.O. Time-series prediction of wind speed using machine learning algorithms: A case study Osorio wind farm, Brazil. Appl. Energy 2018, 224, 550–566. [Google Scholar] [CrossRef]
  29. Ronay, A.; Olga, F.; Enrico, Z. Two machine learning approaches for short-term wind speed time-series prediction. IEEE Trans. Neural Netw. Learn. Syst. 2015, 27, 1734–1747. [Google Scholar]
  30. Hong, Y.Y.; Paulo, P.; Rioflorido, C.L. A hybrid deep learning-based neural network for 24-h ahead wind power forecasting. Appl. Energy 2019, 250, 530–539. [Google Scholar] [CrossRef]
  31. Kumar, P.S.; Lopez, D. Feature Selection used for Wind Speed Forecasting with Data Driven Approaches. Res. J. Appl. Sci. Eng. Technol. 2015, 8, 921–926. [Google Scholar]
  32. Fan, S.; Liao, J.R.; Yokoyama, R.; Chen, L.N.; Lee, W.J. Forecasting the Wind Generation Using a Two-Stage Network Based on Meteorological Information. IEEE. Trans. Energy Conver. 2009, 24, 474–482. [Google Scholar] [CrossRef]
  33. Zhang, Z.D.; Qin, H.; Liu, Y.Q.; Wang, Y.Q.; Yao, L.Q.; Li, Q.Q.; Li, J.; Pei, S.Q. Long Short-Term Memory Network based on Neighborhood Gates for processing complex causality in wind speed prediction. Energy Convers. Manag. 2019, 192, 37–51. [Google Scholar] [CrossRef]
  34. Zhang, Y.G.; Zhang, C.H.; Zhao, Y.; Gao, S. Wind speed prediction with RBF neural network based on PCA and ICA. J. Electr. Eng. 2018, 69, 148–155. [Google Scholar] [CrossRef] [Green Version]
  35. Zhang, Y.G.; Chen, B.; Pan, G.F.; Zhao, Y. A novel hybrid model based on VMD-WT and PCA-BP-RBF neural network for short-term wind speed forecasting. Energy Convers. Manag. 2019, 195, 180–197. [Google Scholar] [CrossRef]
  36. Sarker, I.H.; Abushark, Y.B.; Khan, A.I. ContextPCA: Predicting Context-Aware Smartphone Apps Usage based on Machine Learning Techniques. Symmetry 2020, 12, 499. [Google Scholar] [CrossRef] [Green Version]
  37. Gupta, V.; Mittal, M. KNN and PCA classifier with Autoregressive modelling during different ECG signal interpretation. Procedia. Comput. 2017, 125, 18–24. [Google Scholar] [CrossRef]
  38. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural. Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  39. Liu, Y.; Guan, L.; Hou, C.; Han, H.; Liu, Z.J.; Sun, Y.; Zheng, M.H. Wind Power Short-Term Prediction Based on LSTM and Discrete Wavelet Transform. Appl. Sci. 2019, 9, 1108. [Google Scholar] [CrossRef] [Green Version]
  40. Wang, L.; Zeng, Y.; Chen, T. Back propagation neural network with adaptive differential evolution algorithm for time series forecasting. Expert Syst. Appl. 2015, 42, 855–863. [Google Scholar] [CrossRef]
  41. Cui, L.Z.; Huang, Q.L.; Li, G.H.; Yang, S.; Ming, Z.; Wen, Z.K.; Lu, N.; Lu, J. Differential Evolution Algorithm with Tracking Mechanism and Backtracking Mechanism. IEEE Access 2018, 6, 44252–44267. [Google Scholar] [CrossRef]
Figure 1. LSTM network.
Figure 1. LSTM network.
Applsci 10 04416 g001
Figure 2. LSTM unit.
Figure 2. LSTM unit.
Applsci 10 04416 g002
Figure 3. The proposed prediction framework.
Figure 3. The proposed prediction framework.
Applsci 10 04416 g003
Figure 4. Training set and test set.
Figure 4. Training set and test set.
Applsci 10 04416 g004
Figure 5. Forecast results in Case I.
Figure 5. Forecast results in Case I.
Applsci 10 04416 g005
Figure 6. Forecast results in Case II.
Figure 6. Forecast results in Case II.
Applsci 10 04416 g006
Figure 7. Forecast results in Case III.
Figure 7. Forecast results in Case III.
Applsci 10 04416 g007
Table 1. Basic information of meteorological factors.
Table 1. Basic information of meteorological factors.
Feature NumberFactorUnitMinMaxMeanStd
1Air pressurehpa902919.5909.963.42
2Sea level pressurehpa993.91017.41004.54.37
3Air temperature°C10.235.322.645.09
4Apparent temperature°C9.1632.3221.744.85
5Extreme wind speedm/s1.215.656.483.76
6Wind level\162.621.27
7Relative humidity%109140.8816.99
8Minimum relative humidity%89038.316.28
9Vapor pressurehpa4.318.410.362.41
10Precipitationmm00.70.000.04
11Horizontal visibilitym750035,00034,4622652.5
Table 2. Calculation results of each principal component.
Table 2. Calculation results of each principal component.
Principal ComponentEigenvaluesContribution RateCumulative Contribution Rate
First0.21890.52380.5238
Second0.08430.20170.7255
Third0.06610.15820.8837
Fourth0.02030.04850.9322
Fifth0.01380.03310.9653
Sixth0.00910.02170.9870
Seventh0.00270.00650.9935
Eighth0.00210.00511.0
Ninth0.00030.00071.0
Tenth0.00030.00061.0
Eleventh001.0
Table 3. Parameter settings.
Table 3. Parameter settings.
ModelParametersValueReason
LSTM
Feature-LSTM Proposed Model
Learning rate0.0047Obtained by DE
Hidden unit68Obtained by DE
Batch size29Obtained by DE
Epochs of training1000Converged
BPNN
Feature-BPNN PCA-BPNN
Hidden unit10Common value
Epochs of training1000Converged
SVR
Feature-SVR
PCA-SVR
Kernel functionRadial Basis FunctionA competitive kernel function
Parameter in Radial Basis Function1Common value
GPR
Feature-GPR
PCA-GPR
Kernel FunctionGaussian FunctionA competitive kernel function
Parameter in Gaussian Function2Common value
Parameter in Gaussian Function1Common value
RNN
Feature-RNN PCA-RNN
Hidden unit10Common value
Epochs of training1000Converged
Table 4. Evaluation prediction results in Case I.
Table 4. Evaluation prediction results in Case I.
ModelStatistical IndicatorsRMSEMAPEMAER2
BPNNMin0.80380.20220.61380.6967
Mean0.84350.21090.63440.879
Max1.0290.24470.7210.9518
GPR/0.80340.20540.6190.8688
LSTMMin0.31980.09300.23860.9471
Mean0.33270.10040.25980.9655
Max0.34830.10770.27530.9791
RNNMin0.80260.05070.61880.8601
Mean0.80440.05210.61930.8666
Max0.80750.05330.61990.8723
SVR/0.82920.18640.60990.8814
Table 5. Evaluation of prediction results using meteorological characteristics model.
Table 5. Evaluation of prediction results using meteorological characteristics model.
ModelStatistical IndicatorsRMSEMAPEMAER2
Feature-BPNNMin0.24250.07020.18320.9650
Mean0.33520.09750.24980.9806
Max0.45750.14180.32180.9902
Feature-GPR/0.18980.05350.13790.9546
Feature-LSTMMin0.15770.04290.10570.9592
Mean0.17450.04880.12120.9749
Max0.19260.05350.13340.9942
Feature-RNNMin0.16930.04710.12320.9066
Mean0.20770.06130.15490.9648
Max0.26130.08710.19370.9997
Feature-SVR/0.99760.17260.57260.704
Table 6. Evaluation index of prediction results.
Table 6. Evaluation index of prediction results.
ModelStatistical IndicatorsRMSEMAPEMAER2
PCA-BPNNMin0.20580.05690.1480.8455
Mean0.2680.07930.20410.9199
Max0.48540.11550.35580.9739
PCA-GPRMean0.17060.0460.12310.9686
Proposed ModelMin0.13800.03560.0958F0.9988
Mean0.14740.03820.10150.9989
Max0.15920.04150.10880.9991
PCA-RNNMin0.16930.04710.12320.9167
Mean0.20770.06130.15490.9768
Max0.26130.08710.19370.9897
PCA-SVRMean0.30710.07260.22180.8684

Share and Cite

MDPI and ACS Style

Geng, D.; Zhang, H.; Wu, H. Short-Term Wind Speed Prediction Based on Principal Component Analysis and LSTM. Appl. Sci. 2020, 10, 4416. https://doi.org/10.3390/app10134416

AMA Style

Geng D, Zhang H, Wu H. Short-Term Wind Speed Prediction Based on Principal Component Analysis and LSTM. Applied Sciences. 2020; 10(13):4416. https://doi.org/10.3390/app10134416

Chicago/Turabian Style

Geng, Dawei, Haifeng Zhang, and Hongyu Wu. 2020. "Short-Term Wind Speed Prediction Based on Principal Component Analysis and LSTM" Applied Sciences 10, no. 13: 4416. https://doi.org/10.3390/app10134416

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop