Next Article in Journal
Design of Light-Weight High-Entropy Alloys
Previous Article in Journal
Solution of Higher Order Nonlinear Time-Fractional Reaction Diffusion Equation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Short Term Electrical Load Forecasting Using Mutual Information Based Feature Selection with Generalized Minimum-Redundancy and Maximum-Relevance Criteria

School of Electrical Engineering, Northeast Dianli University, Jilin 132012, China
*
Author to whom correspondence should be addressed.
Entropy 2016, 18(9), 330; https://doi.org/10.3390/e18090330
Submission received: 21 July 2016 / Revised: 5 September 2016 / Accepted: 5 September 2016 / Published: 8 September 2016
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
A feature selection method based on the generalized minimum redundancy and maximum relevance (G-mRMR) is proposed to improve the accuracy of short-term load forecasting (STLF). First, mutual information is calculated to analyze the relations between the original features and the load sequence, as well as the redundancy among the original features. Second, a weighting factor selected by statistical experiments is used to balance the relevance and redundancy of features when using the G-mRMR. Third, each feature is ranked in a descending order according to its relevance and redundancy as computed by G-mRMR. A sequential forward selection method is utilized for choosing the optimal subset. Finally, a STLF predictor is constructed based on random forest with the obtained optimal subset. The effectiveness and improvement of the proposed method was tested with actual load data.

1. Introduction

A short-term load forecasting (STLF) predicts future electric loads with a particular prediction limit from one hour extending up to several days. The primary target of smart grids, such as reducing the difference between peak and valley electric loads, large-scale renewable energy absorption, demand side response, and optimal economic operation of the power grid, needs accurate STLF results [1]. In addition, with the development of competitive electricity markets, an accurate STLF is an important basis for drafting a reasonable electricity price and improving the stability of electricity market operation [2].
The existing STLF methods can be divided into traditional methods and artificial intelligence methods. In the traditional methods, such as autoregressive integrated moving average (ARIMA) [3] and regression analysis [4], Kalman filter [5] and exponential smoothing [6] are commonly used. The combination of autoregressive and moving average in ARIMA is a better time series model for STLF [7]. According to the historical time-varying load data, the ARIMA is established and applied for predicting the forthcoming electrical load. The regression analysis uses historical data to establish simple but highly efficient regression models [8]. The Kalman filter improves the accuracy of STLF by estimating each component of load which is apportioned into random and fixed components. The exponential smoothing eliminates the noise in the load time series, and the degree of future load influenced by recent load data can be reflected by adjusting the weight of both data, which is helpful for improving the accuracy of STLF [9]. Overall, the traditional STLF methods can analyze the linear relationships between input and output, but not the nonlinear relationships [10]. If the load presents large fluctuations caused by environmental factors, the traditional methods may provide inaccurate forecasts.
In recent years, predictors based on artificial intelligence algorithms were widely used in the STLF of power systems [10,11,12,13,14,15,16,17]. Such processes like fuzzy logic [14], expert systems [16,17], artificial neural networks (ANNs) [18,19], and support vector machines (SVMs) [20,21] are currently used in STLF. Fuzzy logic methods divide the input and the output into different kinds of membership functions, and then the relationship between input and output is established by a set of fuzzy rules for fuzzy systems for STLF [22]. However, the fuzzy systems with single if-then rules lack self-learning and adaptive ability to be able to learn the input information effectively. An ANN acquires the complicated non-linear relationship between input and output variables by learning the training samples. However, there is no scientific way of acquiring the optimal network architecture when establishing an ANN model. In addition, it also encounters the problems of falling into local optima and over-fitting [15,23]. SVMs overcome the deficiencies of ANNs by dealing with quadratic programming problems in acquiring the global optimal solution. As compared to an ANN, the SVM has many advantages. However, the SVM parameters, such as the type and variance of the kernel function, and penalty factor, are selected empirically. To achieve the optimal parameters, a SVM combined with genetic and particle swarm optimization algorithm is utilized [24,25]. The random forest (RF) is a combination of classification and regression trees (CARTs) and a bagging learning method. Randomly, by sampling from the training samples and selecting features for splitting node, the RF provides the ability to resist noise and is free from over-fitting problems [26]. Furthermore, in actual practice, there are only two parameters (the tree number and the number of the features for node splitting) that need to be set when RF is applied for STLF [15], making RF highly suitable for STLF.
Considering the effect of various factors, artificial intelligence methods analyze the complicated nonlinear relationships between power load and related factors to achieve higher precision of prediction. However, the features that the predictor employs will influence the accuracy and efficiency of STLF. Therefore, a feature selection schedule should be generated for choosing the optimal feature subset for a predictor. The common features, including historical load, time, and meteorology, are used for STLF modeling [11,27,28]. Historical load can reflect the variation of load accurately, which contains plenty of information. The features of time, such as hour point, day of week, and on/off work day, can also indirectly show the load pattern. In addition, a short-term power load is mainly affected by the changing weather conditions which have a strong correlation with load demand. The accurate meteorological information of the numerical weather prediction (NWP) can improve the accuracy of STLF effectively. Consequently, NWP errors will reduce the accuracy of STLF [29].
A feature selection is a process of choosing the most effective features from an original feature set. The optimal feature subset extracted from a given feature set can improve the efficiency and accuracy of predictor in STLF [30]. Nowadays, the manner of selecting the features has become a hot topic in short-term load forecasting research. Reference [31] adopted conditional mutual information for feature selection. The mutual information values between features and load was measured and subsequently ranked through their values. The first 50 features were used as a threshold parameter for filtering out the irrelevant and weakly relevant features. Reference [10] constructed an original feature set by using the phase space reconstruction theory. The correlation between features and load was analyzed, discovering the optimal feature subset. In reference [29], the mutual information was applied for extracting the effective features from the weather features, as well as, the historical load data features were also extracted for improving the accuracy of holiday load forecasting. Reference [32] used a memetic algorithm to extract a proper feature subset from an original feature set for medium-term load forecasting. Reference [33] analyzed the daily and weekly pattern by autocorrelation function, and chose 50 features as the best features for very short-term load forecasting. The mutual information based on feature selection was used in reference [23]. By calculating the mutual information values between feature vectors and target variable, we can temporarily define a lower boundary criterion to filter the features. The optimal feature subset with best features was achieved for STLF. All of the researches [10,23,29,31,32,33] made important contributions to the feature selection in STLF. However, these feature selection methods were just carried out by analyzing the correlation between features and load and the redundancy among these features was not considered.
To improve the accuracy of STLF, the mutual information based on generalized minimum redundancy and maximum relevance feature selection and RF for STLF is proposed. First, an original feature set is formed by extracting historical load features and time features from the original load data. Second, G-mRMR is used for generating the candidate feature, which is ranked in a descending order. Third, the sequential forward selection (SFS) method and a decision criteria based on mean absolute percentage error (MAPE) are utilized for obtaining optimal feature subset by adding one feature at a time to the input feature set of RF. Finally, the RF-based predictor is constructed with the optimal feature subset to achieve the optimal predictor. The proposed method is validated through STLF experiments using the actual load data from a city in Northeast China. The experimental results are compared with different feature selection methods and predictors.

2. Methodology

2.1. Mutual Information-Based Generalized Minimal-Redundancy and Maximal-Relevance

The minimum-redundancy and maximum-relevance (mRMR) is the method which uses mutual information (MI) to measure the dependence between two variables. The MI-based mRMR not only considers the effective information between feature and target variable, but also acquires the repetitive information among features [34]. It has the advantage of obtaining helpful features accurately when dealing with high dimensional data.
Given two random variables X and Y, the MI between them can be estimated as:
I ( X , Y ) = X , Y P ( x , y ) log P ( x , y ) P ( x ) P ( y )
where P(x) and P(y) are the marginal density functions, and P(x, y) is the joint probability density function.
The target of feature selection methods based on MI is finding a feature subset J with n features which reflect the largest dependency on the target variable l from a feature set Fm with m features ( n m ).
The maximum-relevance criterion uses the mean value of MI between feature xi and target l is described as follows:
max D ( J , l ) ,   D = 1 | J | x i J I ( x i , l )
The redundancy indicated by MI value describes the overlapping information among features, wherein a larger MI signifies more overlapping information and vice versa. In the process of feature selection, the features selected by maximum-relevance criterion can have more redundancy, and the redundant features have similar information as the prior selected feature cannot improve the accuracy of predictor. Therefore, the redundancy among features should also be evaluated in the process of feature selection.
The minimum-redundancy requires a minimum dependency among each feature:
min R ( J ) ,   D = 1 | J | 2 x i , x j J I ( x i , x j )
The mRMR criterion combined with Equations (2) and (3) is computed as follows:
max ψ ( D , R ) ,   ψ = D R
Generally, an incremental search method is used to search for the optimal features [34]. Supposing there is a feature set Jn−1 with n−1 features that has been selected. The aim is to select the nth feature from the rest of set {Fm-Jn−1} according to Equation (4). The incremental search method with respect to the condition is as follows:
m R M R : max x j F m - J n 1 [ I ( x j , l ) 1 | J n 1 | x i J n 1 I ( x j , x i ) ]
where |Jn−1| refers to the number of features in Jn−1.
Restructuring Equation (5) by using a weighting factor to balance the redundancy and relevance of feature subset develops into the generalized mRMR (G-mRMR) presented as follows [35]:
G - m R M R : max x j F m - J n 1 [ I ( x j , l ) α x i J n 1 I ( x j , x i ) ]

2.2. Random Forest

The random forest (RF) is a kind of machine learning algorithm presented by Leo Breiman, who integrates classification and regression tree (CART) and bagging algorithm [26]. A RF generates many different CARTs by sampling with replacement, wherein each CART achieves one result. The final forecasting result is achieved by computing the average value of all CARTs’ results.

2.2.1. CART

The CART employs binary recursive partitioning technology for solving classification and regression issues [36]. A CART, which consists of a root node, non-leaf nodes, branches, and leaf nodes, is shown in Figure 1. Each non-leaf node must be divided according to the Gini index when CART grows.
Supposing there is a dataset D with d samples which includes C classes, the Gini index of D can be defined as:
G ( D ) = 1 i = 1 C ( d i d ) 2
where di is the number of ith class.
Afterward, the feature f is used to divide D into D1 and D2 subset, wherein the Gini index after the split is:
G s p l i t ( D ) = d 1 d G ( D 1 ) + d 2 d G ( D 2 )

2.2.2. Bagging

The bagging is an integrated learning algorithm proposed by Leo Breiman [37]. Given dataset B with M features and learning rule H, a bootstrapping is carried out to generate training sets { B 1 , B 2 , , B q } . The samples in dataset B may be appraised many times or not at all. A forecasting system consists of a group of learning rule { H 1 , H 2 , , H q } which have learned the training set is achieved. Breiman pointed out that bagging can improve the accuracy of predicting the instability of learning algorithms such as CART and ANN [37].

2.2.3. RF

The RF is a group of predictors { p ( x , Θ k ) , k = 1 , 2 , } , which is composed of numbers of CARTs, where x is the input vector and { Θ k } represents the independent identically distributed random vectors. The modeling process of RF is:
(1)
k training sets are sampled with replacement from the dataset B by bootstrap.
(2)
Each training set grows up to a tree according to CART algorithm. Supposing dataset B has M features and mtry features are randomly selected from B for each non-leaf node. Afterward, the node is split by a feature selected from these mtry features.
(3)
Each tree grows completely without pruning.
(4)
The forecasting result is solved by calculating the mean value of the consequences of each tree predicted.
The flow chart of RF model is illustrated in Figure 2.
The bagging and the random selection of feature for splitting ensure the good performance of RF, wherein:
(1)
The same capacity of the training set sampled by bootstrap guarantees each sample in dataset B to be appraised equally. A situation that one sample may appear many times in the same training set and some may not causes low correlation among the trees.
(2)
The manner of selecting feature for node split applies randomness, and ensures the generalized performance of RF.
The number of feature mtry and the number of tree of RF nTree should be set when applying RF. Generally, mtry suggested setting is either mtry = [log2 (M) + 1] or mtry = M or mtry = M/3. The scale of RF generally selected empirically the largest size in order to improve the diversity of trees and guarantee the performance of RF.

3. Data Analysis

The historical load data used in this paper is archived data from a city in Northeast China from 2005 to 2012. As shown in Figure 3a,b, the load demand from 2005 to 2012 increased rapidly with the increase in population and development of the local society. It is difficult to generate a highly accurate STLF in this kind of load pattern. Figure 3c shows the correlation analysis results of the historical load by autocorrelation function [38]. Evidently, the autocorrelation coefficient is reduced gradually along with the increasing of hour lag. According to Figure 3c, the load far from current has low correlation. Only the correlation of the load data from 2011 to 2012 is above the confidence interval which is positive correlation (above of the blue line). With the increasing of the load, the historic load with large lag has very low correlation with the forecasting point. Therefore,we prefer the data from 2011 to 2012 to be used for further research.
Figure 4 shows the average daily load pattern occurring in different seasons. These loads have visibly different patterns which are caused by the varying climate.
By observing Figure 5, it is possible to know that the load demand presents a kind of cycling mode with a period of 7 days. The load demand from Monday to Friday is similar, whereas on Saturday and Sunday they are dissimilar from each other. This pattern is due to the concurrent changing of load level with the varying electricity consumption behavior of people within a week.
The load point predicts the highly correlated load points similar from the day before as well as relevant with previous week. As shown in Figure 6, the load points throughout the week at lag 1, lag 24, lag 48, lag 72, lag 96, lag 120, lag 144, and lag 168 have strong relevance assuming each lag is 1 h difference. Furthermore, other moment load values also have different dependence.
The original feature set for STLF can be achieved based on the above analysis. The 168 load variables {Lt-168, Lt-167, …, Lt-2, Lt-1} are extracted as part of original feature set. When doing a day ahead load forecasting, assuming the current moment is t, the load values from the moment t-1 to t-24 are unknown. Therefore, the variables {Lt-24, Lt-23, …, Lt-1} are eliminated from the original feature set. In addition, the features, such as hour of day, the day is within weekday or weekend, day of week and season, are considered for constructing the original feature set.
Though meteorological factor affects the load demand, it is not considered in this paper because the error of NWP influences the accuracy of STLF [29]. If needed, the meteorological can be added into the original feature set for feature selection in the same manner. There are 168 features in the original feature set F, as shown in Table 1.
The meaning of features in Table 1 is:
Exogenous features:
  • FHour means the moment of hour, which is tagged by the numbers from 1 to 24.
  • FWW is either weekday or weekend marked by binary numbers, wherein 0 means weekend and 1 means weekday.
  • FDW refers to the day of week, which is labeled by the numbers from 1 to 7.
  • FS uses the numbers from 1 to 4.
Endogenous features:
  • FL(t-25) is the load 25 h before, FL(t-26) means the load 26 h before, and so on.

4. The Proposed Feature Selection Method and STLF Model

A feature selection method combined with G-mRMR and RF is proposed. First, the redundancy of features in F and the relevance between features and load are measured by G-mRMR. Each feature with mRMR value is ranked in a descending order. Afterward, a SFS-based RF is used to search for the optimal feature subset. The MAPE used as a performance index in the feature subset selection process is defined as:
M A P E = 1 N i = 1 N | Z i Z ^ i Z i | × 100 %
where Zi is the actual value of load, Z ^ i is the forecasting value, N is the number of sample.

4.1. G-mRMR for Feature Selection

Supposed an original feature set Fm including m features and a selected feature set J. The detail of feature selection process is enumerated below:
(1)
Initialization ØJ.
(2)
Compute the relevance between each feature and target variable l. Pick out the feature from Fm which satisfies Equation (2) and add it into J.
(3)
Find the feature in the rest of m−1 features in Fm that satisfies Equation (4) and add it in to J.
(4)
Repeat step (3) until Fm becomes Ø.
(5)
Rank the features in feature set J in descending order in accordance with the measured mRMR value.

4.2. Wrapper for Feature Selection

The common wrapper is a sequential forward and backward selection, both of which do not consider the feature weighting [34,39]. Therefore, the effects of different dimensional features are must be measured, making wrapper a complex and computational feature selection method. According to the result of feature selection of G-mRMR, a wrapper for finding a feature subset can be applied in simpler manner. Considering the features selected by mRMR are ranked in a descending order, the features in the front of the ranking list contain more effective information, thus SFS is used for finding a small feature subset.
A SFS, in which features are sequentially added to an empty candidate set until the addition of another features, does not decrease the criterion. By defining an empty set S and an original feature set Fm, in the first step, the wrapper searched for the feature subset with only one feature, marked as S1, wherein the feature x1 selected in S1 leads to the largest prediction error reduction. In the second step, the wrapper selects the feature x2 from {Fm-S1} and combines with S1 lead to the largest prediction error reduction. The search schedule is repeated until the prediction stops decreasing.

4.3. The Proposed STLF Model

Based on the methods in Section 4.1 and Section 4.2, the method of feature selection with RF for STLF is proposed. The feature selection and short-term load forecasting process are shown in Figure 7, where p is the number of feature and α is the weighting factor from 0.1 to 0.9, with an increment of 0.1.

5. Case Study and Results Analysis

The data for the experiment consists of the actual data from 2011 to 2012 from a city in Northeast China. For the purpose of feature selection and STLF, the data is divided into three parts: (1) training set (extract eight months randomly from 2011); (2) validation set (the remaining four months of 2011); and (3) test set (extract one week from each season from the data of 2012). More information about the data set is shown in Table 2.
The number of variable mtry, which RF is not overly sensitive to, is recommended as mtry = p/3 [40]. The complexity of RF is affected by the number of tree. Under the premise of non-reduction of prediction accuracy, the initial number of trees nTree is set as 500 [15].
Let Equation (9) to be one of the criteria of RF. In addition, the root mean square error (RMSE) is also used. The RMSE is defined in the follow equation:
R M S E = 1 N i = 1 N ( Z i Z ^ i ) 2

5.1. Feature Selection Results Based on G-mRMR and RF

In this subsection, the optimal subset is achieved according to the minimum MAPE by setting different weighting factor values of G-mRMR. Figure 8 shows the MAPE curves of the results from RF predictions under different weighting factor α. As shown in Figure 8a, the MAPE is reduced and reaches a minimum value with the increase in the number of feature. Subsequently, it ceases to decrease and gradually increases, indicating that the later addition of features does not improve the performance of RF, but only brings adverse effect. As shown in Figure 8b, the error is reduced rapidly when adopting a small value of α, for instance α = 0.1, which indicates that features have useful information for improving the performance of RF. By excessively considering the redundancy among features when using a large value of α, the selected feature subset does not provide enough relevant information for the prediction of RF-based predictor.
Table 3 presents the results of feature selection. When α = 0.4, the feature subset has the least number of feature and the RF generates the minimum MAPE. The optimal feature subset is selected.
The RF will do poor forecasting with less trees, while excessive trees will make it a complicated predictor. In order to obtain a reasonable number of trees of RF, an experiment is designed as follows:
(1)
The training set and test set with optimal features are used for the experiment.
(2)
The initial number of tree nTree = 1.
(3)
Training RF and testing with different nTree value with increment of 1 until nTree = 500.
The experimental result is shown in Figure 9.
The prediction error decreases with the increasing number of tree. When nTree > 100, the error tends to be steady. By analyzing the result, nTree = 184 with minimum MAPE = 2.5389% is obtained, using this number of trees as the parameter of RF in the future experiment.

5.2. Comparison Experiments for STLF

The data shown in Table 2 are used in the comparison of experiments.

5.2.1. Comparison of Different Feature Selection Methods

By using RF as the predictor, the feature selection methods such as Pearson Correlation Coefficient (PCC), MI, and SFS, are compared with the proposed method for estimating the effect of feature selection of G-mRMR. The results of these feature selection methods are presented in Figure 10.
In Figure 10, with the same predictor, the SFS provides the best performance, followed by G-mRMR (α = 0.4) and MI, and finally the PCC. The SFS, which convolves with RF, selects 22 features and achieves the minimum MAPE = 2.4925%. Considering the relevance between feature and load and the redundancy among features, G-mRMR (α = 0.4) selects 15 features with the minimum MAPE = 2.5597%. The feature subset selected by MI, which does not consider the redundancy among features, is higher than G-mRMR (α = 0.4). Only the PCC analyzes the linear relation between features and load, however the feature subset selected through this method is not as good as G-mRMR (α = 0.4).
In order to verify the validity of the feature subset applied for STLF, there were four weeks distributed among four seasons in 2012 are used to test each feature subset with RF. For comparison, the full set of features with RF is also tested. The experimental results is shown in Figure 11. By examining the results in Figure 11a–d, generalized minimum redundancy and maximum relevance-random forest (G-mRMR-RF) (α = 0.4), mutual information-random forest (MI-RF), sequential forward selection-random forest (SFS-RF), and RF (full features) can fit with true load value accurately, whereas the accuracy of pearson correlation coefficient-random forest (PCC-RF) is low. The results of fifth day prediction in Figure 11a and the seventh day in Figure 11c show G-mRMR-RF has a better fit than MI-RF, indicating the necessity of considering the redundancy among features. The results of fifth day prediction in Figure 11a show that SFS-RF has better prediction performance than G-mRMR, while the seventh day prediction results in Figure 11c indicates G-mRMR-RF predicts better.
By analyzing Figure 10 and Figure 11 and Table 4, Table 5, Table 6 and Table 7 comprehensively, although SFS achieved the best forecasting results in the feature selection process, the proposed method achieved the better result in the testing schedule. When predicting the 28 days in the test set, the proposed method yields the best forecasting in 20 days and the MAPE in the remaining eight days is higher than other methods, ranging from 0.04% to 0.37%. The average MAPE and the average RMSE indicate G-mRMR-RF performs the best among the methods which demonstrates the validity and advancement of G-mRMR.
The new method also has the minimum value of the maximum error of STLF in the testing set. As shown in Table 6, the maximum MAPE and maximum RMSE of the proposed method are 6.12% and 208.00 MW. Although the maximum error of the new method is high, but compared with other methods, the proposed method still performed better. The high prediction error can be caused by two factors. On the one hand, the load of forecasting day is much larger than the historical load data in the training set. In this paper, most features in the original feature set are extracted from the historical load data. Without the consideration of other features, the prediction results cannot advance just by improving the feature selection and forecasting method. On the other hand, with the significant economic rise of China from 2005 to 2012, the growth rate of gross domestic product of the city is more than 10%. Under this premise, the electric load of the city increases rapidly which makes STLF a challenging work.

5.2.2. Comparison of Different Intelligent Methods

For comparing the influence of different predictors to STLF, support vector regression (SVR) and back propagation neural network (BPNN) are examined with G-mRMR for feature selecting in this subsection. The parameters of SVR are set as follows: the penalty factor is C = 100, the insensitive loss function is ɛ = 0.1, and the kernel width is δ2 = 2 [41].
The parameters of BPNN are set as follows: the number of neurons in hidden layer is Nneu = 2p+1 [42], and the iteration is T = 2000 [43].
Data consist of training set, validation set, and test set are similar with Section 4.2. The SVR and BPNN are used to generate the optimal feature subsets.
Table 8 presents feature subsets that different intelligent STLF methods had selected. With different predictors, the weighting factors are diverse, thus features are varying. Although the final number of feature selected by SVR and BPNN are less than RF, the RF-based predictor has higher precision of prediction which is the main target of STLF.
The test sets, with four weeks being distributed over the four seasons, are used for estimating each predictor with the features chosen above. Figure 12 shows the MAPE for comparison and Table 9 gives the predictive accuracy of each model through maximum, minimum, and average MAPE. In addition, a direct comparison between G-mRMR-RF, generalized minimum redundancy and maximum relevance-back propagation neural network (G-mRMR-BPNN), and generalized minimum redundancy and maximum relevance-support vector regression (G-mRMR-SVR), in terms of MAPE, are also presented in this figure. Except for the MAPE prediction in the seventh day, as shown in Figure 12c, the accuracy of G-mRMR-RF is between 1% and 2%; one point is above 2%. In the whole experiment, only four days show that G-mRMR-RF forecasted worse than other models. Clearly, the G-mRMR-RF is the best prediction model for its low MAPE and small fluctuation of error. The G-mRMR-BPNN shows a little better performance than G-mRMR-SVR. We can observe the maximum MAPE of these four weeks of G-mRMR-RF is 2.26%, 2.04%, 6.12%, 1.98%, respectively, which is smaller than other models. Same conclusion can be drawn by analyzing the minimum and average MAPE.
Based on the comprehensive analysis above, as compared to BPNN and SVR, the RF combines with G-mRMR is more suitable for STLF.

6. Conclusions

For the issues regarding the selection of reasonable features for STLF, a feature selection method based on G-mRMR and RF is proposed in this paper. The experimental results show that the proposed feature selection approach can select fewer features than other feature selection methods, and the features identified by the proposed approach are useful for STLF. In addition, the experimental results show that the forecasting consequences by RF are better than other predictors.
The advantages of the proposed method are as follows:
(1)
MI is adopted as the criterion to measure the relevance between features and time series of load and the dependency among features, which is the basis of quantitative analysis of feature selection by mRMR.
(2)
The correlation between features and load as well as the redundancy of these features are considered. As compared to the maximum relevance method, the G-mRMR method for feature selection reduces the number of optimal feature subset and avoids the association of STLF accuracy with the redundancy of features. For the time being, the relevance and redundancy are balanced by using a variable weighting factor. The features selected by G-mRMR make the accuracy of RF more precise than mRMR.
(3)
The optimal structure of RF is designed for reducing the complexity of the model and for improving the accuracy of STLF.

Acknowledgments

This work is supported by the National Nature Science Foundation of China (No. 51307020), the Science and Technology Development Project of Jilin Province (No. 20160411003XH) and the Science and Technology Foundation of Department of Education of Jilin Province (2016, No. 90).

Author Contributions

Nantian Huang put forward to the main idea and design the whole venation of this paper. Zhiqiang Hu did the experiments and prepared the manuscript. Guowei Cai guided the experiments and paper writing. Dongfeng Yang provided materials. All authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Moslehi, K.; Kumar, R. A reliability perspective of the smart grid. IEEE Trans. Smart Grid 2010, 1, 57–64. [Google Scholar] [CrossRef]
  2. Ren, Y.; Suganthan, P.N.; Srikanth, N.; Amaratunga, G. Random vector functional link network for short-term electricity load demand forecasting. Inf. Sci. 2016, 367–368, 1078–1093. [Google Scholar] [CrossRef]
  3. Lee, C.-M.; Ko, C.-N. Short-term load forecasting using lifting scheme and ARIMA models. Expert Syst. Appl. 2011, 38, 5902–5911. [Google Scholar] [CrossRef]
  4. Goia, A.; May, C.; Fusai, G. Functional clustering and linear regression for peak load forecasting. Int. J. Fofrecast. 2010, 26, 700–711. [Google Scholar] [CrossRef]
  5. Al-Hamadi, H.M.; Soliman, S.A. Fuzzy short-term electric load forecasting using Kalman filter. IEE Proc. Gener. Transm. Distrib. 2006, 153, 217–227. [Google Scholar] [CrossRef]
  6. Ramos, S.; Soares, J.; Vale, Z. Short-term load forecasting based on load profiling. In Proceedings of the 2013 IEEE Power and Energy Society General Meeting, Vancouver, BC, Canada, 21–25 July 2013; pp. 1–5.
  7. Li, W.; Zhang, Z.G. Based on Time Sequence of ARIMA Model in the Application of Short-Term Electricity Load Forecasting. In Proceedings of the 2009 International Conference on Research Challenges in Computer Science, Shanghai, China, 28–29 December 2009; pp. 11–14.
  8. Deshmukh, M.R.; Mahor, A. Comparisons of Short Term Load Forecasting using Artificial Neural Network and Regression Method. Int. J. Adv. Comput. Res. 2011, 1, 96–100. [Google Scholar]
  9. Taylor, J.W. Short-Term Load Forecasting With Exponentially Weighted Methods. IEEE Trans. Power Syst. 2012, 27, 458–464. [Google Scholar] [CrossRef]
  10. Kouhi, S.; Keynia, F.; Ravadanegh, S.N. A new short-term load forecast method based on neuro-evolutionary algorithm and chaotic feature selection. Int. J. Electr. Power Energy Syst. 2014, 62, 862–867. [Google Scholar] [CrossRef]
  11. Raza, M.Q.; Khosravi, A. A review on artificial intelligence based load demand forecasting techniques for smart grid and buildings. Renew. Sustain. Energy Rev. 2015, 50, 1352–1372. [Google Scholar] [CrossRef]
  12. Lin, C.T.; Chou, L.D.; Chen, Y.M.; Tseng, L.M. A hybrid economic indices based short-term load forecasting system. Int. J. Electr. Power Energy Syst. 2014, 54, 293–305. [Google Scholar] [CrossRef]
  13. Yu, F.; Xu, X. A short-term load forecasting model of natural gas based on optimized genetic algorithm and improved BP neural network. Appl. Energy 2014, 134, 102–113. [Google Scholar] [CrossRef]
  14. Çevik, H.H.; Çunkaş, M. Short-term load forecasting using fuzzy logic and ANFIS. Neural Comput. Appl. 2015, 26, 1355–1367. [Google Scholar] [CrossRef]
  15. Lahouar, A.; Slama, J.B.H. Day-ahead load forecast using random forest and expert input selection. Energy Convers. Manag. 2015, 103, 1040–1051. [Google Scholar] [CrossRef]
  16. Ho, K.L.; Hsu, Y.Y.; Chen, C.F.; Lee, T.E. Short term load forecasting of Taiwan power system using a knowledge-based expert system. IEEE Trans. Power Syst. 1990, 5, 1214–1221. [Google Scholar]
  17. Srinivasan, D.; Tan, S.S.; Cheng, C.S.; Chan, E.K. Parallel neural network-fuzzy expert system strategy for short-term load forecasting: System implementation and performance evaluation. IEEE Trans. Power Syst. 1999, 14, 1100–1106. [Google Scholar] [CrossRef]
  18. Quan, H.; Srinivasan, D.; Khosravi, A. Uncertainty handling using neural network-based prediction intervals for electrical load forecasting. Energy 2014, 73, 916–925. [Google Scholar] [CrossRef]
  19. Hernández, L.; Baladrón, C.; Aguiar, J.M.; Carro, B.; Sánchez-Esguevillas, A.; Lloret, J. Artificial neural networks for short-term load forecasting in microgrids environment. Energy 2014, 75, 252–264. [Google Scholar] [CrossRef]
  20. Ko, C.N.; Lee, C.M. Short-term load forecasting using SVR (support vector regression)-based radial basis function neural network with dual extended Kalman filter. Energy 2013, 49, 413–422. [Google Scholar] [CrossRef]
  21. Che, J.X.; Wang, J.Z. Short-term load forecasting using a kernel-based support vector regression combination model. Appl. Energy 2014, 132, 602–609. [Google Scholar] [CrossRef]
  22. Pandian, S.C.; Duraiswamy, K.; Rajan, C.C.A.; Kanagaraj, N. Fuzzy approach for short term load forecasting. Electr. Power Syst. Res. 2006, 76, 541–548. [Google Scholar] [CrossRef]
  23. Božić, M.; Stojanović, M.; Stajić, Z.; Stajić, N. Mutual Information-Based Inputs Selection for Electric Load Time Series Forecasting. Entropy 2013, 15, 926–942. [Google Scholar] [CrossRef]
  24. Ma, L.; Zhou, S.; Lin, M. Support Vector Machine Optimized with Genetic Algorithm for Short-Term Load Forecasting. In Proceedings of the 2008 International Symposium on Knowledge Acquisition and Modeling, Wuhan, China, 21–22 December 2008; pp. 654–657.
  25. Gao, R.; Liu, X. Support vector machine with PSO algorithm in short-term load forecasting. In Proceedings of the 2008 Chinese Control and Decision Conference, Yantai, China, 2–4 July 2008; pp. 1140–1142.
  26. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  27. Jurado, S.; Nebot, À.; Mugica, F.; Avellana, N. Hybrid methodologies for electricity load forecasting: Entropy-based feature selection with machine learning and soft computing techniques. Energy 2015, 86, 276–291. [Google Scholar] [CrossRef] [Green Version]
  28. Wilamowski, B.M.; Cecati, C.; Kolbusz, J.; Rozycki, P. A Novel RBF Training Algorithm for short-term Electric Load Forecasting and Comparative Studies. IEEE Trans. Ind. Electron. 2015, 62, 6519–6529. [Google Scholar]
  29. Wi, Y.M.; Joo, S.K.; Song, K.B. Holiday load forecasting using fuzzy polynomial regression with weather feature selection and adjustment. IEEE Trans. Power Syst. 2012, 27, 596–603. [Google Scholar] [CrossRef]
  30. Viegas, J.L.; Vieira, S.M.; Melício, M.; Mendes, V.M.F.; Sousa, J.M.C. GA-ANN Short-Term Electricity Load Forecasting. In Proceedings of the 7th IFIP WG 5.5/SOCOLNET Advanced Doctoral Conference on Computing, Electrical and Industrial Systems, Costa de Caparica, Portugal, 11–13 April 2016; pp. 485–493.
  31. Li, S.; Wang, P.; Goel, L. A novel wavelet-based ensemble method for short-term load forecasting with hybrid neural networks and feature selection. IEEE Trans. Power Syst. 2015, 1788–1798. [Google Scholar] [CrossRef]
  32. Hu, Z.; Bao, Y.; Chiong, R.; Xiong, T. Mid-term interval load forecasting using multi-output support vector regression with a memetic algorithm for feature selection. Energy 2015, 84, 419–431. [Google Scholar] [CrossRef]
  33. Koprinska, I.; Rana, M.; Agelidis, V.G. Yearly and seasonal models for electricity load forecasting. In Proceedings of the 2011 International Joint Conference on Neural Networks (IJCNN), San Jose, CA, USA, 31 July–5 August 2011; pp. 1474–1481.
  34. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information: Criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef] [PubMed]
  35. Nguyen, X.V.; Chan, J.; Romano, S.; Bailey, J. Effective global approaches for mutual information based feature selection. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 512–521.
  36. Speybroeck, N. Classification and regression trees. Int. J. Public Health 2012, 57, 243–246. [Google Scholar] [CrossRef] [PubMed]
  37. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
  38. Sood, R.; Koprinska, I.; Agelidis, V.G. Electricity load forecasting based on autocorrelation analysis. In Proceedings of the 2010 International Joint Conference on Neural Networks (IJCNN), Barcelona, Spain, 18–23 July 2010; pp. 1–8.
  39. Kohavi, R.; John, G.H. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef]
  40. Dudek, G. Short-Term Load Forecasting Using Random Forests. In Intelligent Systems’2014; Springer: Cham, Switzerland, 2015; pp. 821–828. [Google Scholar]
  41. Che, J.X.; Wang, J.Z.; Tang, Y.J. Optimal training subset in a support vector regression electric load forecasting model. Appl. Soft Comput. 2012, 12, 1523–1531. [Google Scholar] [CrossRef]
  42. Sheela, K.G.; Deepa, S.N. Review on Methods to Fix Number of Hidden Neurons in Neural Networks. Math. Probl. Eng. 2013, 2013, 425740. [Google Scholar] [CrossRef]
  43. Rana, M.; Koprinska, I. Forecasting electricity load with advanced wavelet neural networks. Neurocomputing 2016, 182, 118–132. [Google Scholar] [CrossRef]
Figure 1. A simple CART.
Figure 1. A simple CART.
Entropy 18 00330 g001
Figure 2. Random Forest modeling and predicting process.
Figure 2. Random Forest modeling and predicting process.
Entropy 18 00330 g002
Figure 3. Yearly load curve analysis: (a) Average daily load from 8 January 2005 to 31 December 2012; (b) The population and GDP from 2005 to 2012; (c) Hourly load autocorrelation of historical load data.
Figure 3. Yearly load curve analysis: (a) Average daily load from 8 January 2005 to 31 December 2012; (b) The population and GDP from 2005 to 2012; (c) Hourly load autocorrelation of historical load data.
Entropy 18 00330 g003
Figure 4. Four seasons average daily load profile from December 2010 to November 2011.
Figure 4. Four seasons average daily load profile from December 2010 to November 2011.
Entropy 18 00330 g004
Figure 5. Load curve from 15 to 28 August 2011.
Figure 5. Load curve from 15 to 28 August 2011.
Entropy 18 00330 g005
Figure 6. Autocorrelation coefficient of load with 168 lags.
Figure 6. Autocorrelation coefficient of load with 168 lags.
Entropy 18 00330 g006
Figure 7. The feature selection process based on G-mRMR and RF for STLF.
Figure 7. The feature selection process based on G-mRMR and RF for STLF.
Entropy 18 00330 g007
Figure 8. Prediction error curves: (a) Prediction error curves corresponding to different weighting factor α; (b) The enlarged figure of red box in (a).
Figure 8. Prediction error curves: (a) Prediction error curves corresponding to different weighting factor α; (b) The enlarged figure of red box in (a).
Entropy 18 00330 g008
Figure 9. Correlation between tree number and prediction of RF.
Figure 9. Correlation between tree number and prediction of RF.
Entropy 18 00330 g009
Figure 10. Prediction error curves: (a) Prediction error curves corresponding to different feature selection methods; (b) The enlarge figure of red box in (a).
Figure 10. Prediction error curves: (a) Prediction error curves corresponding to different feature selection methods; (b) The enlarge figure of red box in (a).
Entropy 18 00330 g010
Figure 11. Load curves of forecasting results of four weeks in four seasons and the true values: (a) Forecasting from 23 to 29 February 2012; (b) Forecasting from 13 to 19 May 2012; (c) Forecasting from 21 to 27 August 2012; (d) Forecasting from 24 to 30 November 2012.
Figure 11. Load curves of forecasting results of four weeks in four seasons and the true values: (a) Forecasting from 23 to 29 February 2012; (b) Forecasting from 13 to 19 May 2012; (c) Forecasting from 21 to 27 August 2012; (d) Forecasting from 24 to 30 November 2012.
Entropy 18 00330 g011
Figure 12. Forecasting error profiles of different predictors: (a) Forecasting from 23 to 29 February 2012; (b) Forecasting from 13 to 19 May 2012; (c) Forecasting from 21 to 27 August 2012; (d) Forecasting from 24 to 30 November 2012.
Figure 12. Forecasting error profiles of different predictors: (a) Forecasting from 23 to 29 February 2012; (b) Forecasting from 13 to 19 May 2012; (c) Forecasting from 21 to 27 August 2012; (d) Forecasting from 24 to 30 November 2012.
Entropy 18 00330 g012
Table 1. The original feature set.
Table 1. The original feature set.
Feature TypeOriginal Feature
Exogenous features1.FHour, 2.FWW, 3.FDW, 4.FS
Endogenous features5.FL(t-25), 6.FL(t-26), 7.FL(t-27), 8.FL(t-28), …, 146.FL(t-166), 147.FL(t-167), 148.FL(t-168)
Table 2. Detail information about the data set.
Table 2. Detail information about the data set.
Data SetInformationPurpose
Training SetJanuary, February, May, June, August, September, October, DecemberTrain RF
Validation SetMarch, April, July, NovemberUse for obtain the best weighting factor
Test Set23–29 February 2012 (Winter)
13–19 May 2012 (Spring)
21–27 August 2012 (Summer)
24–30 November 2012 (Fall)
Test performance of RF
Table 3. Feature subsets selected by minimum MAPE under different weighting factors.
Table 3. Feature subsets selected by minimum MAPE under different weighting factors.
αMin MAPE (%)Number of FeaturesFeature Subset
0.12.564026FL(t-168), FL(t-25), FL(t-48), FL(t-144), FL(t-72), FHour, FL(t-47), FL(t-26), FL(t-120), FL(t-167), FWW, FS, FDW, FL(t-34), FL(t-158), FL(t-103), FL(t-27), FL(t-96), FL(t-162), FL(t-132), FL(t-44), FL(t-88), FL(t-149), FL(t-153), FL(t-37), FL(t-107)
0.22.585725FL(t-168), FL(t-25), FL(t-48), FL(t-144), FHour, FS, FWW, FL(t-71), FDW, FL(t-27), FL(t-106), FL(t-162), FL(t-38), FL(t-127), FL(t-93), FL(t-156), FL(t-88), FL(t-32), FL(t-29), FL(t-96), FL(t-44), FL(t-134), FL(t-26), FL(t-166), FL(t-59)
0.32.585827FL(t-168), FL(t-25), FL(t-48), FHour, FWW, FS, FL(t-144), FDW, FL(t-103), FL(t-37), FL(t-162), FL(t-70), FL(t-131), FL(t-28), FL(t-88), FL(t-153), FL(t-106), FL(t-75), FL(t-159), FL(t-34), FL(t-125), FL(t-96), FL(t-43), FL(t-165), FL(t-109), FL(t-31), FL(t-26)
0.42.559715FL(t-168), FL(t-25), FL(t-48), FWW, FS, FL(t-127), FL(t-85), FL(t-139), FDW, FL(t-34), FL(t-160), FL(t-70), FL(t-28), FL(t-120), FL(t-141)
0.52.589780FL(t-168), FL(t-25), FL(t-47), FWW, FS, FL(t-127), FL(t-86), FDW, FL(t-139), FL(t-35), FL(t-99), FL(t-160), FL(t-69), FL(t-29), FL(t-154), FL(t-120), FL(t-41), FL(t-81), FL(t-133), FL(t-148), FL(t-166), FL(t-32), FL(t-63), FL(t-92), FL(t-26), FL(t-108), FL(t-162), FL(t-78), …
0.62.586846FL(t-168), FL(t-25), FHour, FL(t-47), FS, FL(t-127), FL(t-88), FDW, FL(t-156), FL(t-139), FL(t-76), FL(t-34), FL(t-110), FL(t-69), FL(t-149), FL(t-120), FL(t-41), FL(t-81), FL(t-27), FL(t-165), FL(t-37), FL(t-162), FL(t-98), FL(t-30), FL(t-131), FL(t-159), FL(t-104), FL(t-44), …
0.72.589188FL(t-168), FL(t-25), FWW, FS, FL(t-103), FL(t-61), FL(t-139), FDW, FL(t-47), FL(t-160), FL(t-82), FL(t-124), FL(t-30), FL(t-93), FL(t-156), FL(t-41), FL(t-146), FL(t-33), FL(t-110), FL(t-72), FL(t-152), FL(t-164), FL(t-27), FL(t-90), FL(t-131), FL(t-39), FL(t-118), FL(t-77), …
0.82.604693FL(t-168), FL(t-25), FWW, FS, FL(t-103), FL(t-61), FL(t-139), FDW, FL(t-47), FL(t-160), FL(t-82), FL(t-124), FL(t-30), FL(t-93), FL(t-156), FL(t-41), FL(t-146), FL(t-33), FL(t-110), FL(t-166), FL(t-75), FL(t-152), FL(t-90), FL(t-72), FL(t-44), FL(t-131), FL(t-28), FL(t-39), …
0.92.591835FL(t-168), FL(t-25), FWW, FS, FL(t-103), FL(t-67), FL(t-133), FDW, FL(t-34), FL(t-160), FL(t-46), FL(t-148), FL(t-96), FL(t-29), FL(t-84), FL(t-140), FL(t-39), FL(t-153), FL(t-75), FL(t-114), FL(t-165), FL(t-56), FL(t-122), FL(t-62), FL(t-155), FL(t-126), FL(t-41), FL(t-119), …
Table 4. Comparison of prediction error (MAPE (%) and RMSE (MW)) from 23 to 29 February 2012.
Table 4. Comparison of prediction error (MAPE (%) and RMSE (MW)) from 23 to 29 February 2012.
DayG-mRMR-RF (α = 0.4)MI-RFPCC-RFSFS-RFRF with Full Features
MAPERMSEMAPERMSEMAPERMSEMAPERMSEMAPERMSE
Day 11.9375.241.7969.4210.28401.012.0779.581.9170.74
Day 21.7766.631.7867.909.78388.262.2279.461.8069.13
Day 31.5853.241.6351.637.59285.511.4749.331.5050.49
Day 41.6979.281.5970.025.35189.652.52105.321.9876.33
Day 52.2690.722.66104.1611.14440.912.0483.682.91113.32
Day 61.5857.732.3783.879.78396.441.6157.412.5487.59
Day 71.2851.920.9736.359.26362.461.8773.031.2944.60
Average1.7267.821.8269.059.02352.031.9775.401.9973.17
Table 5. Comparison of prediction error (MAPE (%) and RMSE (MW)) from 13 to 19 May 2012.
Table 5. Comparison of prediction error (MAPE (%) and RMSE (MW)) from 13 to 19 May 2012.
DayG-mRMR-RF (α = 0.4)MI-RFPCC-RFSFS-RFRF with Full Features
MAPERMSEMAPERMSEMAPERMSEMAPERMSEMAPERMSE
Day 11.2042.361.2239.153.76110.391.4350.901.5747.92
Day 21.6460.321.3350.268.98273.781.2846.101.3753.34
Day 32.0466.882.0467.096.56246.642.0369.432.0066.78
Day 40.9434.380.9634.487.04263.290.8934.981.1141.54
Day 51.5553.261.4046.627.17261.541.4050.041.5052.38
Day 61.2841.451.3444.686.66237.551.2840.221.4540.03
Day 70.8426.820.9936.975.51178.830.9250.611.0149.05
Average1.3546.491.3348.036.53224.571.3248.901.4050.15
Table 6. Comparison of prediction error (MAPE (%) and RMSE (MW)) from 21 to 27 August 2012.
Table 6. Comparison of prediction error (MAPE (%) and RMSE (MW)) from 21 to 27 August 2012.
DayG-mRMR-RF (α = 0.4)MI-RFPCC-RFSFS-RFRF with Full Features
MAPERMSEMAPERMSEMAPERMSEMAPERMSEMAPERMSE
Day 12.8892.712.6183.056.69258.313.32104.412.8990.68
Day 21.4855.301.5556.818.22319.741.7762.531.5957.62
Day 30.9131.930.8229.027.04263.331.0038.681.0736.28
Day 41.8876.952.2790.868.97344.821.9984.882.1787.44
Day 51.7754.771.8756.566.42227.252.1670.951.9158.15
Day 62.0873.601.7871.445.91181.331.7165.131.8674.78
Day 76.12208.006.77237.0011.26458.196.98247.666.57227.17
Average2.4572.832.5289.257.79293.282.7096.322.5890.30
Table 7. Comparison of prediction error (MAPE (%) and RMSE (MW)) from 24 to 30 November 2012.
Table 7. Comparison of prediction error (MAPE (%) and RMSE (MW)) from 24 to 30 November 2012.
DayG-mRMR-RF (α = 0.4)MI-RFPCC-RFSFS-RFRF with Full Features
MAPERMSEMAPERMSEMAPERMSEMAPERMSEMAPERMSE
Day 11.6158.601.6458.266.80263.401.9668.901.7863.62
Day 21.0548.241.1243.266.97242.742.1178.431.0940.30
Day 31.9874.622.0674.6710.78427.391.9873.902.1276.64
Day 41.4757.141.3348.099.21387.301.8067.131.5057.63
Day 51.1242.840.9033.8310.29413.171.1546.871.2645.01
Day 61.3153.791.3352.339.03389.521.3254.741.2347.40
Day 71.1042.861.2245.319.53387.691.0644.211.1843.42
Average1.3854.011.3750.828.93358.741.6362.021.4553.43
Table 8. The optimal subset selected by using different intelligent STLF methods.
Table 8. The optimal subset selected by using different intelligent STLF methods.
PredictorMin MAPE (%)Number of FeaturesFeature Subset
G-mRMR-RF (α = 0.4)2.5389%15FL(t-168), FL(t-25), FL(t-48), FWW, FS, FL(t-127), FL(t-85), FL(t-139), FDW, FL(t-34), FL(t-160), FL(t-70), FL(t-28), FL(t-120), FL(t-141)
G-mRMR-SVR (α = 0.3)3.3293%5FL(t-168), FL(t-25), FL(t-48), FHour, FWW
G-mRMR-BPNN (α = 0.1)2.7186%11FL(t-168), FL(t-25), FL(t-48), FL(t-144), FL(t-72), FHour, FL(t-47), FL(t-26), FL(t-120), FL(t-167), FWW
Table 9. Max, Min and Average daily MAPEs of test set corresponding to different predictors.
Table 9. Max, Min and Average daily MAPEs of test set corresponding to different predictors.
Test SetMAPE (%)G-mRMR-RFG-mRMR-ANNG-mRMR-SVR
23–29 February 2012 (Winter)Max2.263.724.24
Min1.281.231.37
Ave1.722.182.61
13–19 May 2012 (Spring)Max2.043.144.42
Min0.841.241.87
Ave1.352.052.78
21–27 August 2012 (Summer)Max6.126.327.68
Min0.911.781.69
Ave2.452.873.50
24–30 November 2012 (Fall)Max1.982.844.01
Min1.051.381.80
Ave1.382.102.86

Share and Cite

MDPI and ACS Style

Huang, N.; Hu, Z.; Cai, G.; Yang, D. Short Term Electrical Load Forecasting Using Mutual Information Based Feature Selection with Generalized Minimum-Redundancy and Maximum-Relevance Criteria. Entropy 2016, 18, 330. https://doi.org/10.3390/e18090330

AMA Style

Huang N, Hu Z, Cai G, Yang D. Short Term Electrical Load Forecasting Using Mutual Information Based Feature Selection with Generalized Minimum-Redundancy and Maximum-Relevance Criteria. Entropy. 2016; 18(9):330. https://doi.org/10.3390/e18090330

Chicago/Turabian Style

Huang, Nantian, Zhiqiang Hu, Guowei Cai, and Dongfeng Yang. 2016. "Short Term Electrical Load Forecasting Using Mutual Information Based Feature Selection with Generalized Minimum-Redundancy and Maximum-Relevance Criteria" Entropy 18, no. 9: 330. https://doi.org/10.3390/e18090330

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop