ReNFuzz-LF: A Recurrent Neurofuzzy System for Short-Term Load Forecasting
Abstract
:1. Introduction
2. The Architecture of ReNFuzz-LF
- The fuzzy rules have static premise parts, which comprise m-dimensional hyper-cells, composed of single-dimension Gaussian membership functions along each input axis:
- The firing strength is calculated as the algebraic product of the Gaussian membership functions.
- The consequent parts of the fuzzy rules are three-layer recurrent neural networks in the form of m–H–1. Such a configuration is shown in Figure 1. The network has internal recurrence, since the outputs of the nodes in the hidden layer are fed back with unit delays (local output unit feedback). There exist no feedback connections of the fuzzy rules’ outputs or external feedback from the network’s output.The consequent part of the l-th fuzzy rule operates as follows:
- ➢
- The hyperbolic tangent is chosen to be the activation function of the neurons, , over other common activation functions, such as sigmoid or rectified linear unit (RELU). As mentioned in [40], the hyperbolic tangent performs better than the logistic sigmoid. In modern deep neural networks, the RELU function is preferred due to the fact that it overcomes the vanishing gradients problem ([40]) that occurs in multilayered networks. In the present case, the consequent parts of the fuzzy rules are too shallow for such a problem to occur. Moreover, the competing rivals of ReNFuzz-LF in Section 4 use the hyperbolic tangent as activation function.
- ➢
- is the output of the i-th hidden neuron of the l-th rule for the k-th sample.
- ➢
- is the output of the l-th fuzzy rule.
- ➢
- , and are the synaptic weights and bias terms, respectively, at the hidden layer of the consequent parts.
- ➢
- and are the synaptic weights and bias terms, respectively, at the output layer of the consequent parts.
- The defuzzification part is static. The output of ReNFuzz-LF is calculated using the weighted average method:
3. The Learning Process
Algorithm 1. SA-DRPROP | ||
1: | (a)Initialize the step sizes of the consequent weights : | |
2: | Repeat | |
3: | (b) For each weight , compute the SA-DRPROP error gradient: | |
4: | ||
5: | (c) For each weight , update its step size: (c.1) If | |
6: | Then, | |
7: | (c.2) Else, if | |
8: | Then, | |
9: | If | |
10: | Then | |
11: | Else, | |
12: | (c.3) Else, | |
13: | Update : | |
14: | Until convergence |
4. Experimental Results
4.1. Problem Statement—Data Preprocessing
4.2. ReNFuzz-LF’s Structural and Parameter Characteristics
4.3. Results and Discussion
- The working days of the four seasons bear similarity with respect to the appearances of morning and evening peaks, as well the first minimum load. As far as the load evolution during the day is concerned, autumn working day is smoother from 10 a.m. to 4 p.m.
- Sundays exhibit different patterns compared to their respective working days. Moreover, they differ considerably from season to season. It is evident that spring and autumn Sundays follow the respective seasonal pattern only during late evening and around midnight.
- Despite the differences in seasonal behavior and in the type of days, ReNFuzz-LF succeeds in tracking the actual time-series at all times. Maximum and minimum extremes are identified, and the transition from working days to weekend days is effectively modeled, as can be clearly seen in Figure 9, where a winter week is shown.
- The usual selection process with regard to past load values is skipped. No dimensionality reduction is necessary, which adds additional cost to the preprocessing phase. The internal dependencies of the time-series are identified through the recurrent processing that the consequent parts of the fuzzy rules perform.
- The model does not make use of climate data, such as temperature of humidity.
- No seasonal models or models dependent on the nature of the day (weekday, weekend, holiday) are necessary. With the exemption of August 15th, ReNFuzz-LF succeeded in predicting the electric load of irregular days quite accurately.
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Nomenclature
membership function of the i-th input axis for the l-th fuzzy rule | |
mean of a Gaussian membership function of the i-th input axis for the l-th fuzzy rule | |
standard deviation of a Gaussian membership function of the i-th input axis for the l-th fuzzy rule | |
H | number of hidden neurons |
RMSE | root mean squared error |
APE | average percentage error |
n | number of samples |
FIR | finite impulse response |
IIR | infinite impulse response |
maximum electric actual load of day d | |
actual electric load at hour h of day d-1 | |
the output of the i-th hidden neuron of the l-th fuzzy rule | |
the output of the l-th fuzzy rule | |
synaptic weights at the hidden layer of the fuzzy rules | |
bias terms at the hidden layer of the fuzzy rules | |
synaptic weights at the output layer of the fuzzy rules | |
bias terms at the output layer of the fuzzy rules | |
membership degree that a sample belongs to the l-th cluster (FCM) | |
c | scale parameter (FCM) |
ordered partial derivative of an error measure with respect to a consequent weight | |
Lagrange multiplier | |
actual electric load value | |
increase factor for step size | |
attenuation factor for step size | |
r | scale parameter for noise (SA-DRPROP) |
Temp | temperature (SA-DRPROP) |
Scale coefficients for SA term (SA-DRPROP) | |
minimum step size | |
maximum step size | |
initial step size |
References
- Park, D.C.; El-Sharkawi, M.; Marks, R.; Atlas, L.; Damborg, M. Electric load forecasting using an artificial neural network. IEEE Trans. Power Syst. 1991, 6, 442–449. [Google Scholar] [CrossRef] [Green Version]
- Papalexopoulos, A.; How, S.; Peng, T. An implementation of a neural network based load forecasting model for the EMS. IEEE Trans. Power Syst. 1994, 9, 1956–1962. [Google Scholar] [CrossRef]
- Bakirtzis, A.; Theocharis, J.; Kiartzis, S.; Satsios, K. Short term load forecasting using fuzzy neural networks. IEEE Trans. Power Syst. 1995, 10, 1518–1524. [Google Scholar] [CrossRef]
- Mastorocostas, P.; Theocharis, J.; Bakirtzis, A. Fuzzy modeling for short term load forecasting using the orthogonal least squares method. IEEE Trans. Power Syst. 1999, 14, 29–36. [Google Scholar] [CrossRef]
- Papadakis, S.; Theocharis, J.; Bakirtzis, A. A load curve based fuzzy modeling technique for short-term load forecasting. Fuzzy Sets Syst. 2003, 135, 279–303. [Google Scholar] [CrossRef]
- Bansal, R. Bibliography on the fuzzy set theory applications in power systems. IEEE Trans. Power Syst. 2003, 18, 1291–1299. [Google Scholar] [CrossRef]
- Dash, P.; Liew, A.; Rahman, S.; Dash, S. Fuzzy and neuro-fuzzy computing models for electric load forecasting. Eng. Appl. Artif. Intell. 1995, 8, 423–433. [Google Scholar] [CrossRef]
- Chen, B.; Chang, M. Load forecasting using support vector machines: A study on EUNITE competition 2001. IEEE Trans. Power Syst. 2004, 19, 1821–1830. [Google Scholar] [CrossRef] [Green Version]
- Ghelardoni, L.; Ghio, A.; Anguita, D. Energy load forecasting using empirical mode decomposition and support vector regression. IEEE Trans. Smart Grid 2013, 4, 549–556. [Google Scholar] [CrossRef]
- Yang, A.; Li, W.; Yang, X. Short-term electricity load forecasting based on feature selection and least squares support vector machines. Knowl. Based Syst. 2019, 163, 159–173. [Google Scholar] [CrossRef]
- Giasemidis, G.; Haben, S.; Lee, T.; Singleton, C.; Grindrod, P. A genetic algorithm approach for modelling low voltage network demands. Appl. Energy 2017, 203, 463–473. [Google Scholar] [CrossRef] [Green Version]
- Yang, X.; Yuan, J.; Yuan, J.; Mao, H. An improved WM method based on PSO for electric load forecasting. Expert Syst. Appl. 2010, 37, 8036–8041. [Google Scholar] [CrossRef]
- Dudek, G. Artificial immune system with local feature selection for short-term load forecasting. IEEE Trans. Evol. Comput. 2017, 21, 116–130. [Google Scholar] [CrossRef]
- Li, S.; Wang, P.; Goel, L. A novel wavelet-based ensemble method for short-term load forecasting with hybrid neural networks and feature selection. IEEE Trans. Power Syst. 2016, 31, 1788–1798. [Google Scholar] [CrossRef]
- Zhou, M.; Jin, M. Holographic ensemble forecasting method for short-term power load. IEEE Trans. Smart Grid 2019, 10, 425–434. [Google Scholar] [CrossRef]
- Jang, J.-S.R. ANFIS: Adaptive-network-based fuzzy inference system. IEEE Trans. Syst. Man, Cybern. 1993, 23, 665–685. [Google Scholar] [CrossRef]
- Khan, A.; Rizwan, M. ANN and ANFIS Based Approach for Very Short Term Load Forecasting: A Step Towards Energy Management System. In Proceedings of the 8th International Conference on Signal Processing and Integrated Networks, Noida, India, 26–27 August 2021. [Google Scholar]
- Elkazaz, M.; Sumner, M.; Thomas, D. Real-time energy management for a small scale PV-battery microgrid: Modeling, design, and experimental verification. Energies 2019, 12, 2712. [Google Scholar] [CrossRef] [Green Version]
- Shah, S.; Nagraja, H.; Chakravorty, J. ANN and ANFIS for short term load forecasting. Eng. Technol. Appl. Sci. Res. 2018, 8, 2818–2820. [Google Scholar] [CrossRef]
- Krzywanski, J.; Grabowska, K.; Sosnowski, M.; Zylka, A.; Szteklr, K.; Kalawa, W.; Wojcik, T.; Nowak, W. An adaptive neuro-fuzzy model of a re-heat two-stage adsorption chiller. Therm. Sci. 2019, 23, S1053–S1063. [Google Scholar] [CrossRef] [Green Version]
- Lasheen, M.; Abdel-Salam, M. Maximum power point tracking using hill climbing and ANFIS techniques for PV applications: A review and a novel hybrid approach. Energy Convers. Manag. 2018, 171, 1002–1019. [Google Scholar] [CrossRef]
- Karim, F.; Majumdar, S.; Darabi, H.; Harford, S. Multivariate LSTM-FCNs for time series classification. Neural Netw. 2019, 116, 237–245. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Demir, F. DeepCoroNet: A deep LSTM approach for automated detection of COVID-19 cases from chest X-ray images. Appl. Soft Comput. 2021, 103, 107160. [Google Scholar] [CrossRef] [PubMed]
- Wang, M.; Lin, T.; Jhan, K.; Wu, S. Abnormal event detection, identification and isolation in nuclear power plants using LSTM networks. Prog. Nucl. Energy 2021, 140, 103928. [Google Scholar] [CrossRef]
- Tsakiridis, N.; Keramaris, K.; Theocharis, J.; Zalidis, G. Simultaneous prediction of soil properties from VNIR-SWIR spectra using a localized multi-channel 1-D convolutional neural network. Geoderma 2020, 367, 114208. [Google Scholar] [CrossRef]
- Massaoudi, M.; Chihi, I.; Abu-Rub, H.; Refaat, S.; Oueslati, F. Convergence of photovoltaic power forecasting and deep learning: State-of-art review. IEEE Access. 2021, 9, 136593–136615. [Google Scholar] [CrossRef]
- Agga, A.; Abbou, A.; Labbadi, M.; El Houm, Y.; Ali, I. CNN-LSTM: An efficient hybrid deep leanring architecture for predicting short-term photovoltaic power production. Electr. Power Syst. Res. 2022, 208, 107908. [Google Scholar] [CrossRef]
- Skrobek, D.; Krzywanski, J.; Sosnowski, M.; Kulakowska, A.; Zylka, A.; Grabowska, K.; Ciesielska, K.; Nowak, W. Prediction of sorption processes using the deep learning methods. Energies 2020, 13, 6601. [Google Scholar] [CrossRef]
- Gao, Z.; Hu, S.; Sun, H.; Liu, J.; Zhi, Y.; Li, J. Dynamic state estimation of new energy power systems considering multi-level false data identification based on LSTM-CNN. IEEE Access. 2021, 9, 142411–142424. [Google Scholar] [CrossRef]
- Veeramsetty, V.; Chandra, D.R.; Salkuti, S.R. Short-term electric power load forecasting using factor analysis and long short-term memory for smart cities. Int. J. Circuit Theory Appl. 2021, 49, 1678–1703. [Google Scholar] [CrossRef]
- Kong, W.; Dong, Z.; Jia, Y.; Hill, D.; Xu, Y.; Zhang, Y. Short-term residential load forecasting based on LSTM recurrent neural network. IEEE Trans. Smart Grid 2019, 10, 844–851. [Google Scholar] [CrossRef]
- Han, L.; Peng, Y.; Li, Y.; Yong, B.; Zhou, Q.; Shu, L. Enhanced deep networks for short-term and medium-term load forecasting. IEEE Access 2018, 7, 4045–4055. [Google Scholar] [CrossRef]
- Bedi, J.; Toshniwal, D. Deep learning framework to forecast electricity demand. Appl. Energy 2019, 238, 1312–1326. [Google Scholar] [CrossRef]
- He, W. Load forecasting via deep neural networks. Procedia Comput. Sci. 2017, 122, 308–314. [Google Scholar] [CrossRef]
- Veeramsetty, V.; Chandra, D.R.; Grimaccia, F.; Mussetta, M. Short term electric load forecasting using principal component analysis and recurrent neural networks. Forecasting 2022, 4, 149–164. [Google Scholar] [CrossRef]
- Eskandari, H.; Imani, M.; Moghaddam, M. Convolutional and recurrent neural network based model for short-term load forecasting. Electr. Power Syst. Res. 2021, 195, 107173. [Google Scholar] [CrossRef]
- Sheng, Z.; Wang, H.; Chen, G.; Zhou, B.; Sun, J. Convolutional residual network to short-term load forecasting. Appl. Intell. 2021, 51, 2485–2499. [Google Scholar] [CrossRef]
- Takagi, T.; Sugeno, M. Fuzzy identification of systems and its applications. IEEE Trans. Syst. Man Cybern. 1985, 15, 116–132. [Google Scholar] [CrossRef]
- Mastorocostas, P.; Hilas, C. ReNFFor: A recurrent neurofuzzy forecaster for telecommunications data. Neural Comput. Appl. 2013, 22, 1727–1734. [Google Scholar] [CrossRef]
- Goodfellow, I.; Bengio, J.; Courville, A. Deep Learning; The MIT Press: Cambridge, MA, USA, 2017; pp. 187–189. [Google Scholar]
- Shihabudheen, K.; Pillai, G. Recent advances in neuro-fuzzy system: A Survey. Knowl. Based Syst. 2018, 152, 136–162. [Google Scholar] [CrossRef]
- Ojha, V.; Abraham, A.; Snasel, V. Heuristic design of fuzzy inference systems: A review of three decades of research. Eng. Appl. Artif. Intel. 2019, 85, 845–864. [Google Scholar] [CrossRef] [Green Version]
- Jassar, S.; Liao, Z.; Zhao, L. A recurrent neuro-fuzzy system and its application in inferential sensing. Appl. Soft Comput. 2011, 11, 2935–2945. [Google Scholar] [CrossRef]
- Juang, C.-F.; Lin, Y.-Y.; Tu, C.-C. A recurrent self-evolving fuzzy neural network with local feedbacks and its application to dynamic system processing. Fuzzy Sets Syst. 2010, 161, 2552–2568. [Google Scholar] [CrossRef]
- Stavrakoudis, D.; Theocharis, J. Pipelined recurrent fuzzy networks for nonlinear adaptive speech prediction. IEEE Trans. Syst. Man Cybern. B. Cybern. 2007, 37, 1305–1320. [Google Scholar] [CrossRef] [PubMed]
- Mandic, D.; Chambers, J. Recurrent Neural Networks for Prediction: Learning Algorithms, Architectures and Stability; John Wiley & Sons, Inc.: Hoboke, NJ, USA, 2001. [Google Scholar]
- Tsoi, A.; Back, A. Locally recurrent Ggobally feedforward networks: A critical review of architectures. IEEE Trans. Neural Netw. 1994, 5, 229–239. [Google Scholar] [CrossRef]
- Mastorocostas, P.; Theocharis, J. A Recurrent fuzzy neural model for dynamic system identification. IEEE Trans. Syst. Man Cybern. B. Cybern. 2002, 32, 176–190. [Google Scholar] [CrossRef]
- Dunn, J. A fuzzy relative of the ISODATA process and its use in detecting compact, well-separated clusters. J. Cybernet. 1974, 3, 32–57. [Google Scholar] [CrossRef]
- Bezdek, J. Fuzzy Mathematics in Pattern Recognition. Ph.D. Thesis, Cornell University, Ithaca, NY, USA, 1973. [Google Scholar]
- Bezdek, J. Cluster validity with fuzzy sets. J. Cybernet. 1973, 3, 58–73. [Google Scholar] [CrossRef]
- Wingham, M. Geometrical fuzzy clustering algorithms. Fuzzy Sets Syst. 1983, 10, 271–279. [Google Scholar] [CrossRef]
- Zhou, T.; Chung, F.-L.; Wang, S. Deep TSK fuzzy classifier with stacked generalization and triplely concise interpretability guarantee for large data. IEEE Trans. Fuzzy Syst. 2017, 25, 1207–1221. [Google Scholar] [CrossRef]
- Arbelaitz, O.; Gurrutxaga, I.; Muguerza, J.; Perez, J.; Perona, I. An extensive comparative study of cluster validity indices. Pattern Recognit. 2013, 46, 243–256. [Google Scholar] [CrossRef]
- Davies, D.; Bouldin, D. A clustering separation measure. IEEE Trans. Pattern Anal. Mach. Intell. 1979, 1, 224–227. [Google Scholar] [CrossRef] [PubMed]
- Mastorocostas, P.; Rekanos, I. Simulated Annealing Dynamic RPROP for Training Recurrent Fuzzy Systems. In Proceedings of the 14th IEEE International Conference on Fuzzy Systems, Reno, NV, USA, 22–25 May 2005. [Google Scholar]
- Treadgold, N.; Gedeon, T. Simulated annealing and weight decay in adaptive learning: The SARPROP algorithm. IEEE Trans. Neural Netw. 1998, 9, 662–668. [Google Scholar] [CrossRef] [PubMed]
- Werbos, P. Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. Ph.D. Thesis, Harvard University, Cambridge, MA, USA, 1974. [Google Scholar]
- Piche, S. Steepest descent algorithms for neural network controllers and filters. IEEE Trans. Neural Netw. 1994, 5, 198–212. [Google Scholar] [CrossRef] [PubMed]
- Greek Independent Power Transmission Operator. Available online: https://www.admie.gr/en/market/market-statistics/detail-data (accessed on 10 April 2022).
Parameter | Number |
---|---|
m | |
Premise | |
Consequent | |
Total |
Temp | |||||||
---|---|---|---|---|---|---|---|
1.2 | 1.05 | 0.5 | 0.0001 | 0.5 | 0.01 | 0.01 | 0.4 |
Grid Partition (6 Rules) | |||
APE Training | RMSE Training (Denormalized) | APE Testing | RMSE Testing (Denormalized) |
1.24% | 121 | 1.41% | 137 |
FCM Partition (3 rules) | |||
APE training | RMSE training (Denormalized) | APE testing | RMSE testing (Denormalized) |
1.19% | 113 | 1.35% | 122 |
Season | APE Training | RMSE Training (Denormalized) | APE Testing | RMSE Testing (Denormalized) |
---|---|---|---|---|
Winter | 1.05% | 113 | 1.10% | 112 |
Spring | 1.60% | 143 | 1.97% | 162 |
Summer | 0.93% | 88 | 1.08% | 100 |
Autumn | 1.18% | 103 | 1.21% | 104 |
Electric Load | >100 MW | >200 MW | >400 MW | >500 MW |
---|---|---|---|---|
Hours | 2787 | 794 | 91 | 40 |
Time | 31.80% | 9.04% | 1.04% | 0.45% |
DFNN | Structural Parameters | ||||||
---|---|---|---|---|---|---|---|
H | Membership function | Activation function | |||||
1 | 2 | 2 | 1 | 2 | Gaussian | tanh | |
Learning parameters | |||||||
1.05 | 0.5 | 0.0001 | 0.01 | 0.85 | |||
ANFIS | Learning parameters | ||||||
Initial Step size | Step size increase rate | Step size decrease rate | |||||
0.01 | 1.1 | 0.9 | |||||
Membership function: Gaussian | |||||||
LSTM | Hyperparameters | ||||||
Activation function | Bias | Dropout | Batch size | Optimizer | learning rate | ||
tanh | Yes | 0.2 | 24 | Adam | 0.001 |
Model | APE (Testing) | No. of Parameters |
---|---|---|
ReNFuzz-LF | 1.35% | 33 |
DFNN | 1.36% | 48 |
ANFIS | 1.48% | 279 |
LSTM-1 | 1.73% | 2726 |
LSTM-2 | 2.06% | 7826 |
LSTM-3 | 1.51% | 10451 |
LSTM-4 | 1.23% | 30651 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kandilogiannakis, G.; Mastorocostas, P.; Voulodimos, A. ReNFuzz-LF: A Recurrent Neurofuzzy System for Short-Term Load Forecasting. Energies 2022, 15, 3637. https://doi.org/10.3390/en15103637
Kandilogiannakis G, Mastorocostas P, Voulodimos A. ReNFuzz-LF: A Recurrent Neurofuzzy System for Short-Term Load Forecasting. Energies. 2022; 15(10):3637. https://doi.org/10.3390/en15103637
Chicago/Turabian StyleKandilogiannakis, George, Paris Mastorocostas, and Athanasios Voulodimos. 2022. "ReNFuzz-LF: A Recurrent Neurofuzzy System for Short-Term Load Forecasting" Energies 15, no. 10: 3637. https://doi.org/10.3390/en15103637
APA StyleKandilogiannakis, G., Mastorocostas, P., & Voulodimos, A. (2022). ReNFuzz-LF: A Recurrent Neurofuzzy System for Short-Term Load Forecasting. Energies, 15(10), 3637. https://doi.org/10.3390/en15103637