In order to handle the high-nonlinear regression problem in a day-ahead photovoltaic forecasting problem, an ensemble algorithm titled as the evidential extreme learning machine (EELM) is proposed in this paper, which inserts the ELM into the framework of evidential regression. The proposed algorithm not only regards the sample points as training data, but also takes the model information of nearest neighbors into consideration, which can improve the accuracy of prediction or regression. Moreover, by integrating multiple ELM models, the stability of the prediction model is promoted. The proposed EELM algorithm is employed in the proposed photovoltaic forecasting process based on similar days.
3.1. Evidential Extreme Learning Machine Algorithm
In this section, a novel regression algorithm named the evidential extreme learning machine is proposed, which can be considered as an ensemble model of ELM under the framework of evidential theory. The basic concept of the proposed ensemble method is to build multiple ELM models and weigh the optimal ELM models, while the weights are acquired by the normalized mass function. The EELM algorithm is on the basis of evidential regression with k-nearest neighbors, as shown in
Figure 2.
For a testing point to be predicted, the predicted output is determined by multiple ELM models of the k-nearest neighbors. For each neighbor, the distance of inputs, which can be regarded as the similarity, are converted to a mass through a discounting and combination process. Furthermore, instead of the original outputs of nearest neighbor points, the predicted outputs are accessed by ELM models of neighbors with the inputs of the point to be predicted. The general predicted value equals the expectation which is calculated by mass function and ELM prediction results. To build a suitable and accurate ELM model for each neighbor point, several data pre-processing methods can be employed, such as data clustering, which clusters the data with similar features for model training to ensure separate and accurate models for sample points in space. Meanwhile, another promising method is to reconstruct the database, through which a certain database is divided into multiple sub-databases.
As the general predicted output is accessed by the outputs of the ELM model as well as the mass functions which are regarded as the weights of outputs in the EELM model, the overall outputs can be expressed as a linear equation of outputs gained from multiple ELM models:
where
mi,
I = 1, 2, ⋯,
K, Ω is the belief assignment of each ELM model. To be more precise, the overall output can be calculated as:
where
K is the number of neighbors.
is the estimated output with the input
x through the ELM model of the
k-th neighbor:
where
ωE,j,k and
bj,k are the weights and bias of the
jth hidden layer node in the
k-th ELM model, respectively. The forecasting accuracy of the ensemble model is:
To improve the forecasting accuracy, parameter
γ can be optimized according to:
Thereby, the
BetPm and dynamic confidence interval
are calculated according to the confidence level
αpv, which are:
The output weights of ELM can be accessed through the desired output
yi. In the condition where the desired output of each ELM can be generated, the output weights of ELM are trained independently.
where
Nk is the number of training samples in TR corresponding to the
k-th ELM model. In case of a special situation where the desired outputs are unknown and the ELM model share the same coefficients
β, the output weight matrix can be:
The flowchart of the proposed EELM algorithm with certain ELM outputs is exhibited in
Figure 3, and the main steps can be summarized as follows:
The input variables, output variables, training database, testing database objective function and algorithm parameters are specified, including the number of hidden layer nodes L and the number of nearest neighbors K.
The database, including inputs and corresponding outputs, is reconstructed and separated into several sub-databases according to specific regulations, and for each sub-database an ELM model is constructed.
For ELM models, the weights and bias of input are generated randomly, and the penalty factor is set as a constant value. Nevertheless, an artificial bee colony (ABC) algorithm is employed in this model to optimize these parameters subsequently.
Concerning the corresponding output of the ELM model, the output weights are achieved. For regression problems with a regularized term, the output weights can be calculated as:
Differing from the randomly assigned weights and bias of input in a traditional ELM model, the weights, basis as well as the penalty factor are optimized by employing the ABC algorithm to enhance and ensure the accuracy of the forecasting algorithm [
43]. Furthermore, if the end condition of the optimization process is satisfied, proceed to the following step; otherwise, return to
Step 4 with updated parameters to continue execution.
Referring to the given metric ǁ·ǁ, k-nearest neighbors of the predicted point are searched. Afterwards, the mass function is obtained through discounting and combination by the conjunctive rule of combination.
As the most important parameter for regression, the parameter γ is optimized by using the leave-one-out method with an optimization objective of CVE. Unlike a traditional evidential regression algorithm, for a certain point to be predicted with an input of x, the corresponding predicted values of k-nearest neighbors are the outputs given by ELM models with the identical input x.
Regarding a testing point to be predicted, the output can be calculated by equations with the optimized weights and bias of ELMs as well as parameter γ of evidential regression. In accordance with the value of BetPm, the probabilistic forecasting results can be generated following Equations (30) and (31).
3.2. Photovoltaic Power Forecasting Based on Evidential Extreme Learning Machine
To guarantee economic and steady operation of power systems under growing PV penetration, precise algorithms for day-ahead PV power are essentially required. Accurate PV power forecasting is a complex issue on account of the fluctuated and volatile nature of weather. The prediction of PV power is a high nonlinear regression problem, and the power generation depends mostly on the solar radiance. The remaining potential influencing parameters such as the atmospheric temperature, humidity, precipitation and atmospheric pressure are also regarded as available inputs for PV power prediction. The day-ahead PV power forecasting is specified as:
where
yPV,t is the photovoltaic generation at the
t-th hour to be predicted;
xpv,i is the corresponding influencing factors.
The proposed forecasting approach comprises five stages: data preprocessing and reconstruction; ELM model training and parameter optimization; mass function calculation and evidential parameter optimization; deterministic forecasting; and probabilistic forecasting.
Figure 4 exhibits the detailed flowchart of the EELM algorithm, in which
xPV represents available meteorological elements, while
xPV,i and
yPV,i indicate available meteorological elements and PV power of the
ith similar days, respectively.
In the data preprocessing stage, the abnormal data are abandoned, and the missing parts are filled in. The optimal input vectors are arranged by Pearson correlation coefficient (
PCC) [
44], which is considered as a typical measurement for the interdependence of variables. The PCC between two vectors
S and
T can be mentioned as:
where
cov(
S,
T) is the covariance between
S and
T, while
σS and
σT are standard deviations of
S and
T, respectively. While the absolute value of the
PCC is generally above 0.8, the two factors are considered to have a strong correlation. In addition, it is considered to have a moderate correlation in the case of taking values between 0.3 and 0.8. Consequently, the parameters with an absolute Pearson correlation coefficient under 0.3 are abandoned.
Since the power of similar days have an obvious influence on PV power forecasting, the PV power and meteorological parameters of similar days are added to input variables at the database reconstruction stage. Among this phase, historical PV power and meteorological parameters series are decomposed into K sub-databases. For the k-th sub-database, PV power generations and corresponding meteorological parameters of the k-th similar days for samples are added to input vectors. In general terms, the k-th database contains the PV power and related meteorological parameters, as well as PV power and the corresponding meteorological parameters of the k-th similar days for samples.
During the training process, the EELM model is trained, and hidden layer parameters are optimized. Specifically, an ELM model is built for every sub-database, and the weights and basis of the hidden layer are optimized by an ABC algorithm. Afterwards, a series of predicted PV power is generated according to the optimized ELM models, which can be regarded as the nearest neighbor outputs for a proceeding evidence regression process. The similarity of sample and nearest neighbors is calculated by the Euclidean distance of meteorological parameters, and a normalized belief assignment is obtained through discounting and combination. At the following step, a general PV power prediction is acquired according to the ELM prediction results and mass functions. By minimizing the forecasting error, the parameter γ is optimized. To emphasize, the basic concept of EELM in PV forecasting is that it replaces the power generations of similar days by the predicted power through ELM, which are calculated by input parameters of the predicted samples and parameters of similar days.
With the optimized parameters in EELM, a PV power forecasting framework is constructed. For a specific point to be predicted, the meteorological parameters are used to search for similar days, and an expectation of power forecasting can be calculated based on ELM outputs and mass function. In many cases, the forecasting outcomes are supposed to be in the form of an estimation interval. Differing from the roughly modeled normal distribution in previous studies, the EELM generates the confidence interval via the accumulation of probabilities for potential values.