A Review of Smart Grid Evolution and Reinforcement Learning: Applications, Challenges and Future Directions
Abstract
:1. Introduction
- This work first reviews the development history of smart grids. By analyzing the evolution of smart grids and related advanced technologies in detail, it expounds the core issues and challenges faced in the process of smart grid optimal scheduling, providing background knowledge in the field of smart grid optimal scheduling and laying the foundation for subsequent discussions.
- This study reviews the state of the art in deep reinforcement learning for smart grids and analyzes the strengths and weaknesses of existing approaches. In particular, there are challenges in cooperative control of multi-agent systems, convergence of algorithms, and stability. Then, future research directions are discussed in-depth, and key open problems and potential research areas are proposed.
- A two-layer reinforcement learning optimization framework is introduced to achieve efficient optimization control for smart grids. The upper layer agents are responsible for the coordination and regulation of the global grid, and the lower layer agents perform specific device optimization to enable cooperative optimization of multiple devices. This scheme provides ideas for future research directions.
2. Evolution of Smart Grid Technologies
3. Application of Reinforcement Learning in Smart Grids
3.1. Key Research Directions in Smart Grids
3.2. Evolution of Reinforcement Learning in Smart Grid
4. Two-Layer Reinforcement Learning Architecture for Distributed Smart Grid Management
5. Emerging Trends and Challenges in Smart Grid Research
- Insufficient handling of safety constraints: Current reinforcement learning frameworks usually adopt the “a posteriori penalty” policy, which simply superimposes a constraint violation penalty term in the reward function. However, this approach leads to a conflict between safety constraints and the exploration space of the agent, which can lead to safety risks such as voltage overruns and line overloads while the agent pursues optimal economy.
- Dimensional catastrophes due to multi-timescale coupling: Smart grid operations involve a strong coupling between millisecond-level transient control and hourly level scheduling decisions, resulting in a dramatic growth of the action space dimension. In the face of the diverse operational requirements of smart grids, it is urgent to build a more refined hierarchical distributed cooperative control mechanism that ensures efficient cooperation of each grid node on the basis of independent decision making and guarantees global stability and optimization.
- Insufficient robustness and risk sensitivity: Under the influence of uncertain disturbances and complex constraints, smart grid control algorithms need to be more robust and risk-sensitive. Robustness ensures that the grid can remain stable in uncertain environments, while risk sensitivity requires that the control system can make adaptive decisions under uncertain conditions to minimize potential risks and losses.
- Deep integration of safety reinforcement learning and physical models: Lyapunov function constraints and safety layer embedding architecture are used to construct grid state maps to improve decision interpretability in grid control tasks. An organic combination of reinforcement learning and physical models of dynamical systems will be implemented based on safety-constrained algorithms.
- Collaborative optimization across time scales: Through a hierarchical reinforcement learning architecture, a coupled model across time scales is constructed to achieve efficient collaboration of control units in smart grids. Each agent makes independent decisions based on local information and collaborates with other agents to form a global optimal scheme to realize cross-regional collaborative scheduling, which not only protects data privacy but also reduces the redundancy of reserve capacity, so as to ensure the overall security and stability of the power grid system.
- Risk assessment and adaptive decision making: Risk assessment can be carried out for different scenarios, as well as the flexible adjustment of strategies. By quantifying the probability distribution of extreme events, stochastic optimization and robust optimization are integrated into the adversarial training framework to balance the worst-case and expected performance, and a decentralized robust reinforcement learning algorithm is developed.
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Gill, S.; Kockar, I.; Ault, G.W. Dynamic optimal power flow for active distribution networks. IEEE Trans. Power Syst. 2014, 29, 121–131. [Google Scholar] [CrossRef]
- Lin, C.; Wu, W.; Chen, X.; Zheng, W. Decentralized dynamic economic dispatch for integrated transmission and active distribution networks using multi-parametric programming. IEEE Trans. Smart Grid 2018, 9, 4983–4993. [Google Scholar] [CrossRef]
- Sun, H.; Guo, Q.; Qi, J.; Ajjarapu, V.; Bravo, R.; Chow, J.; Li, Z.; Moghe, R.; Nasr-Azadani, E.; Tamrakar, U.; et al. Review of challenges and research opportunities for voltage control in smart grids. IEEE Trans. Power Syst. 2019, 34, 2790–2801. [Google Scholar] [CrossRef]
- Chen, Y.; Pan, F.; Qiu, F.; Xavier, A.S.; Zheng, T.; Marwali, M.; Knueven, B.; Guan, Y.; Luh, P.B.; Wu, L.; et al. Security-constrained unit commitment for electricity market: Modeling, solution methods, and future challenges. IEEE Trans. Power Syst. 2023, 38, 4668–4681. [Google Scholar] [CrossRef]
- Molzahn, D.K.; Dörfler, F.; Sandberg, H.; Low, S.H.; Chakrabarti, S.; Baldick, R.; Lavaei, J. A survey of distributed optimization and control algorithms for electric power systems. IEEE Trans. Smart Grid 2017, 8, 2941–2962. [Google Scholar] [CrossRef]
- Dsouza, A.K.; Thammaiah, A.; Venkatesh, L.K.M. An intelligent management of power flow in the smart grid system using hybrid npo-atla approach. Artif. Intell. Rev. 2022, 55, 6461–6503. [Google Scholar] [CrossRef]
- Zhou, M.; Zhai, J.; Li, G.; Ren, J. Distributed dispatch approach for bulk AC/DC hybrid systems with high wind power penetration. IEEE Trans. Power Syst. 2018, 33, 3325–3336. [Google Scholar] [CrossRef]
- Antoniadou-Plytaria, K.E.; Kouveliotis-Lysikatos, I.N.; Georgilakis, P.S.; Hatziargyriou, N.D. Distributed and decentralized voltage control of smart distribution networks: Models, methods, and future research. IEEE Trans. Smart Grid 2017, 8, 2999–3008. [Google Scholar] [CrossRef]
- Kekatos, V.; Zhang, L.; Giannakis, G.B.; Baldick, R. Voltage regulation algorithms for multiphase power distribution grids. IEEE Trans. Power Syst. 2016, 31, 3913–3923. [Google Scholar] [CrossRef]
- Evangelopoulos, V.A.; Georgilakis, P.S.; Hatziargyriou, N.D. Optimal operation of smart distribution networks: A review of models, methods and future research. Electr. Power Syst. Res. 2016, 140, 95–106. [Google Scholar] [CrossRef]
- Safdarian, F.; Kargarian, A.; Hasan, F. Multiclass learning-aided temporal decomposition and distributed optimization for power systems. IEEE Trans. Power Syst. 2021, 36, 4941–4952. [Google Scholar] [CrossRef]
- Chen, Y.; Zhu, J.; Liu, Y.; Zhang, L.; Zhou, J. Distributed hierarchical deep reinforcement learning for large-scale grid emergency control. IEEE Trans. Power Syst. 2024, 39, 4446–4458. [Google Scholar]
- Mu, C.; Wang, K.; Ni, Z.; Sun, C. Cooperative differential game-based optimal control and its application to power systems. IEEE Trans. Ind. Inform. 2020, 16, 5169–5179. [Google Scholar]
- Gao, Y.; Wang, W.; Yu, N. Consensus multi-agent reinforcement learning for volt-var control in power distribution networks. IEEE Trans. Smart Grid 2021, 12, 3594–3604. [Google Scholar] [CrossRef]
- Naidu, B.R.; Bajpai, P.; Chakraborty, C.; Malakondaiah, M.; Kumar, B.K. Adaptive dynamic voltage support scheme for fault ride-through operation of a microgrid. IEEE Trans. Sustain. Energy 2023, 14, 974–986. [Google Scholar] [CrossRef]
- Hu, D.; Ye, Z.; Gao, Y.; Ye, Z.; Peng, Y.; Yu, N. Multi-agent deep reinforcement learning for voltage control with coordinated active and reactive power optimization. IEEE Trans. Smart Grid 2022, 13, 4873–4886. [Google Scholar] [CrossRef]
- Xu, S.; Xue, Y.; Chang, L. Review of power system support functions for inverter-based distributed energy resources-standards, control algorithms, and trends. IEEE Open J. Power Electron. 2021, 2, 88–105. [Google Scholar]
- Wang, S.; Duan, J.; Shi, D.; Xu, C.; Li, H.; Diao, R.; Wang, Z. A data-driven multi-agent autonomous voltage control framework using deep reinforcement learning. IEEE Trans. Power Syst. 2020, 35, 4644–4654. [Google Scholar]
- Yu, Y.; Ju, P.; Peng, Y.; Lou, B.; Huang, H. Analysis of dynamic voltage fluctuation mechanism in interconnected power grid with stochastic power disturbances. J. Mod. Power Syst. Clean Energy 2020, 8, 38–45. [Google Scholar]
- Liu, Y.; Qu, Z.; Xin, H.; Gan, D. Distributed real-time optimal power flow control in smart grid. IEEE Trans. Power Syst. 2017, 32, 3403–3414. [Google Scholar]
- Mohammadi, J.; Hug, G.; Kar, S. Agent-based distributed security constrained optimal power flow. IEEE Trans. Smart Grid 2018, 9, 1118–1130. [Google Scholar] [CrossRef]
- Jain, H.; Mather, B.; Jain, A.K.; Baldwin, S.F. Grid-supportive loads? a new approach to increasing renewable energy in power systems. IEEE Trans. Smart Grid 2022, 13, 2959–2972. [Google Scholar]
- Tazi, K.; Abbou, F.M.; Abdi, F. Multi-agent system for microgrids: Design, optimization and performance. Artif. Intell. Rev. 2020, 53, 1233–1292. [Google Scholar] [CrossRef]
- Mu, C.; Liu, W.; Xu, W. Hierarchically adaptive frequency control for an EV-integrated smart grid with renewable energy. IEEE Trans. Ind. Inform. 2018, 14, 4254–4263. [Google Scholar] [CrossRef]
- Gambuzza, L.V.; Frasca, M. Distributed control of multiconsensus. IEEE Trans. Autom. Control 2021, 66, 2032–2044. [Google Scholar]
- Gregoratti, D.; Matamoros, J. Distributed energy trading: The multiple-microgrid case. IEEE Trans. Ind. Electron. 2015, 62, 2551–2559. [Google Scholar]
- Huang, Q.; Huang, R.; Hao, W.; Tan, J.; Fan, R.; Huang, Z. Adaptive power system emergency control using deep reinforcement learning. IEEE Trans. Smart Grid 2020, 11, 1171–1182. [Google Scholar]
- Lowe, R.; Wu, Y.; Tamar, A.; Harb, J. Multi-agent Actor-Critic for mixed cooperative-competitive environments. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Rashid, T.; Samvelyan, M.; Witt, C.S.D.; Farquhar, G.; Foerster, J.; Whiteson, S. Qmix: Monotonic value function factorisation for deep multi-agent reinforcement learning. J. Mach. Learn. Res. 2021, 21, 1–51. [Google Scholar]
- Hao, J.; Yang, T.; Tang, H.; Bai, C.; Liu, J.; Meng, Z.; Liu, P.; Wang, Z. Exploration in deep reinforcement learning: From single-agent to multiagent domain. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 8762–8782. [Google Scholar] [CrossRef]
- Nguyen, V.; Wang, C.; Hsieh, Y. Electrification of highway transportation with solar and wind energy. Sustainability 2021, 13, 5456. [Google Scholar] [CrossRef]
- Li, X.; Fang, Z.; Li, F.; Xie, S.; Cheng, S. Game-based optimal dispatching strategy for distribution network with multiple microgrids leasing shared energy storage. Proc. CSEE 2022, 42, 6611–6625. [Google Scholar]
- Liu, H.; Wu, W. Two-stage deep reinforcement learning for inverter-based Volt-VAR control in active distribution networks. IEEE Trans. Smart Grid 2021, 12, 2037–2047. [Google Scholar]
- DAsl, K.; Seifi, A.R.; Rastegar, M.; Dabbaghjamanesh, M.; Hatziargyriou, N.D. Distributed two-level energy scheduling of networked regional integrated energy systems. IEEE Syst. J. 2022, 16, 5433–5444. [Google Scholar]
- Xu, G.; Lin, Z.; Wu, Q.; Tan, J.; Chan, W.K.V. Bi-level hierarchical model with deep reinforcement learning-based extended horizon scheduling for integrated electricity-heat systems. Electr. Power Syst. 2024, 229, 110195. [Google Scholar]
- Siu, J.Y.; Kumar, N.; Panda, S.K. Command authentication using multiagent system for attacks on the economic dispatch problem. IEEE Trans. Ind. Appl. 2022, 58, 4381–4393. [Google Scholar]
- Zheng, S.; Trott, A.; Srinivasa, S.; Parkes, D.C.; Socher, R. The AI economist: Optimal economic policy design via two-level deep reinforcement learning. Sci. Adv. 2022, 8, 13332. [Google Scholar]
- Chamandoust, H. Optimal hybrid participation of customers in a smart micro-grid based on day-ahead electrical market. Artif. Rev. 2022, 55, 5891–5915. [Google Scholar] [CrossRef]
- She, B.; Li, F.; Cui, H.; Zhang, J.; Bo, R. Fusion of microgrid control with model-free reinforcement learning: Review and vision. IEEE Trans. Smart Grid 2023, 14, 3232–3245. [Google Scholar]
- Zhou, Q.; Shahidehpour, M.; Li, Z.; Xu, X. Two-layer control scheme for maintaining the frequency and the optimal economic operation of hybrid ac/dc microgrids. IEEE Trans. Power Syst. 2019, 34, 64–75. [Google Scholar]
- Gong, Z.; Liu, C.; Shang, L.; Lai, Q.; Terriche, Y. Power decoupling strategy for voltage modulated direct power control of voltage source inverters connected to weak grids. IEEE Trans. Sustain. Energy 2023, 14, 152–167. [Google Scholar]
- Brandao, D.I.; Ferreira, W.M.; Alonso, A.M.S.; Tedeschi, E.; Marafão, F.P. Optimal multiobjective control of low-voltage AC microgrids: Power flow regulation and compensation of reactive power and unbalance. IEEE Trans. Smart Grid 2020, 11, 1239–1252. [Google Scholar]
- Das, A.; Wu, D.; Ni, Z. Approximate dynamic programming with policy-based exploration for microgrid dispatch under uncertainties. Int. Electr. Power Energy Syst. 2022, 142, 108359. [Google Scholar]
- Lee, J.T.; Anderson, S.; Vergara, C.; Callaway, D.S. Non-intrusive load management under forecast uncertainty in energy constrained microgrids. Electr. Power Syst. Res. 2021, 190, 106632. [Google Scholar]
- Liu, W.; Zhuang, P.; Liang, H.; Peng, J.; Huang, Z. Distributed economic dispatch in microgrids based on cooperative reinforcement learning. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2192–2203. [Google Scholar] [PubMed]
- Mu, C.; Zhang, Y.; Gao, Z.; Sun, C. ADP-based robust tracking control for a class of nonlinear systems with unmatched uncertainties. IEEE Trans. Syst. Man Cybern. Syst. 2020, 50, 4056–4067. [Google Scholar] [CrossRef]
- Mu, C.; Sun, C.; Wang, D.; Song, A. Adaptive tracking control for a class of continuous-time uncertain nonlinear systems using the approximate solution of HJB equation. Neurocomputing 2017, 260, 432–442. [Google Scholar]
- Shi, Z.; Wang, W.; Huang, Y.; Li, P.; Dong, L. Simultaneous optimization of renewable energy and energy storage capacity with the hierarchical control. CSEE J. Power Energy Syst. 2022, 8, 95–104. [Google Scholar]
- Esfahani, M.M.; Hariri, A.; Mohammed, O.A. A multiagent-based game-theoretic and optimization approach for market operation of multimicrogrid systems. IEEE Trans. Ind. Inform. 2019, 15, 280–292. [Google Scholar]
- Hong, S.-H.; Lee, H.-S. Robust energy management system with safe reinforcement learning using short-horizon forecasts. IEEE Trans. Smart Grid 2023, 14, 2485–2488. [Google Scholar]
- Wang, K.; Mu, C.; Ni, Z.; Liu, D. Safe reinforcement learning and adaptive optimal control with applications to obstacle avoidance problem. IEEE Trans. Autom. Sci. Eng. 2024, 21, 4599–4612. [Google Scholar]
- Wen, S.; Xiong, W.; Qiu, J. MPC-based frequency control strategy with a dynamic energy interaction scheme for the grid-connected microgrid system. J. Frankl. Inst. 2020, 357, 2736–2751. [Google Scholar]
- Bi, W.; Shu, Y.; Dong, W.; Yang, Q. Real-time energy management of microgrid using reinforcement learning. In Proceedings of the 2020 19th International Symposium on Distributed Computing and Applications for Business Engineering and Science (DCABES), Xuzhou, China, 16–19 October 2020; pp. 38–41. [Google Scholar]
- Xiong, L.; Tang, Y.; Mao, S.; Liu, H.; Meng, K.; Dong, Z.; Qian, F. A two-level energy management strategy for multi-microgrid systems with interval prediction and reinforcement learning. IEEE Trans. Circuits Syst. I Regul. Pap. 2022, 69, 1788–1799. [Google Scholar]
- Zhu, Y.; Nie, C.; Chen, B. Study on multi game cooperative scheduling of microgrid cluster system under hybrid time-scale. Power Syst. 2020, 47, 3249–3260. [Google Scholar]
- Dreglea, A.; Foley, A.; Häger, U.; Sidorov, D.; Tomin, N. Hybrid renewable energy systems, load and generation forecasting, new grids structure, and smart technologies. In Solving Urban Infrastructure Problems Using Smart City Technologies; Vacca, J.R., Ed.; Elsevier: Amsterdam, The Netherlands, 2021; pp. 475–484. [Google Scholar]
- Ding, L.; Lin, Z.; Shi, X.; Yan, G. Target-value-competition-based multi-agent deep reinforcement learning algorithm for distributed nonconvex economic dispatch. IEEE Trans. Power Syst. 2023, 38, 204–217. [Google Scholar]
- Chen, W.; Li, T. Distributed economic dispatch for energy internet based on multiagent consensus control. IEEE Trans. Autom. Control 2021, 66, 137–152. [Google Scholar]
- He, Q.; Ding, L.; Kong, Z.-M.; Hu, P.; Guan, Z.-H. Distributed scheme for line overload mitigation with linearized ac power flow. IEEE Trans. Circuits Syst. II Express Briefs 2022, 69, 2877–2881. [Google Scholar]
- Liu, C.; Xu, X.; Hu, D. Multiobjective reinforcement learning: A comprehensive overview. IEEE Trans. Syst. Man Cybern. Syst. 2015, 45, 385–398. [Google Scholar]
- Khamis, M.A.; Gomaa, W. Adaptive multi-objective reinforcement learning with hybrid exploration for traffic signal control based on cooperative multi-agent framework. Eng. Appl. Artif. 2014, 29, 134–151. [Google Scholar] [CrossRef]
- Dong, P.; Xu, L.; Lin, Y.; Liu, M. Multi-objective coordinated control of reactive compensation devices among multiple substations. IEEE Trans. Power Syst. 2018, 33, 2395–2403. [Google Scholar] [CrossRef]
- Nowak, S.; Chen, Y.C.; Wang, L. Distributed measurement-based optimal der dispatch with estimated sensitivity models. IEEE Trans. Smart Grid 2022, 13, 2197–2208. [Google Scholar]
- Zhang, Q.; Dehghanpour, K.; Wang, Z.; Qiu, F.; Zhao, D. Multi-agent safe policy learning for power management of networked microgrids. IEEE Trans. Smart Grid 2021, 12, 1048–1062. [Google Scholar] [CrossRef]
- Yan, Z.; Xu, Y. Real-time optimal power flow: A lagrangian based deep reinforcement learning approach. IEEE Trans. Power Syst. 2020, 35, 3270–3273. [Google Scholar] [CrossRef]
- Liu, Y.; Li, Y.; Xin, H.; Gooi, H.B.; Pan, J. Distributed optimal tie-line power flow control for multiple interconnected ac microgrids. IEEE Trans. Power Syst. 2019, 34, 1869–1880. [Google Scholar] [CrossRef]
- Mu, C.; Shi, Y.; Xu, N.; Wang, X.; Tang, Z.; Jia, H.; Geng, H. Multi-objective interval optimization dispatch of microgrid via deep reinforcement learning. IEEE Trans. Smart Grid 2024, 15, 2957–2970. [Google Scholar] [CrossRef]
- Lin, N.; Orfanoudakis, S.; Cardenas, N.O.; Giraldo, J.S.; Vergara, P.P. Powerflownet: Power flow approximation using message passing graph neural networks. Int. J. Electr. Power Energy Syst. 2024, 160, 110112. [Google Scholar] [CrossRef]
- Haarnoja, T.; Zhou, A.; Abbeel, P.; Levine, S. Soft Actor-Critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In Proceedings of the International Conference on Machine Learning (ICML), Stockholm, Sweden, 10–15 July 2018; pp. 1861–1870. [Google Scholar]
- Yang, Y.; Yang, Z.; Yu, J.; Xie, K.; Jin, L. Fast economic dispatch in smart grids using deep learning: An active constraint screening approach. IEEE Internet Things J. 2020, 7, 11030–11040. [Google Scholar] [CrossRef]
- Han, X.; Mu, C.; Yan, J.; Niu, Z. An autonomous control technology based on deep reinforcement learning for optimal active power dispatch. Int. J. Electr. Power Energy Syst. 2023, 145, 108686. [Google Scholar] [CrossRef]
- Li, H.; Wang, L.; Lin, D.; Zhang, X. A nash game model of multi-agent participation in renewable energy consumption and the solving method via transfer reinforcement learning. Proc. CSEE 2019, 39, 3249–3260. [Google Scholar]
- Li, M.; Wei, W.; Chen, Y.; Ge, M.-F.; Catalão, J.P.S. Learning the optimal strategy of power system operation with varying renewable generations. IEEE Trans. Sustain. Energy 2021, 12, 2293–2305. [Google Scholar] [CrossRef]
- Nguyen, T.T.; Nguyen, N.D.; Nahavandi, S. Deep reinforcement learning for multiagent systems: A review of challenges, solutions, and applications. IEEE Trans. Cybern. 2020, 50, 3826–3839. [Google Scholar] [CrossRef]
- Liu, H.; Wu, W. Online multi-agent reinforcement learning for decentralized inverter-based volt-var control. IEEE Trans. Smart Grid 2021, 12, 2980–2990. [Google Scholar]
- Sun, X.; Qiu, J. Two-Stage Volt/Var Control in Active Distribution Networks With Multi-Agent Deep Reinforcement Learning Method. IEEE Trans. Smart Grid 2021, 12, 2903–2912. [Google Scholar]
- Zhang, X.; Liu, Y.; Duan, J.; Qiu, G.; Liu, T.; Liu, J. DDPG-based multi-agent framework for SVC tuning in urban power grid with renewable energy resources. IEEE Trans. Power Syst. 2021, 36, 5465–5475. [Google Scholar] [CrossRef]
- Li, X.; Luo, F.; Li, C. Multi-agent deep reinforcement learning-based autonomous decision-making framework for community virtual power plants. Appl. Energy 2024, 360, 122813. [Google Scholar]
- Cao, D.; Zhao, J.; Hu, W.; Ding, F.; Huang, Q.; Chen, Z.; Blaabjerg, F. Data-driven multi-agent deep reinforcement learning for distribution system decentralized voltage control with high penetration of PVs. IEEE Trans. Smart Grid 2021, 12, 4137–4150. [Google Scholar]
- Hou, X.; Guo, Z.; Wang, X.; Qian, T.; Zhang, J.; Qi, S.; Xiao, J. Parallel learner: A practical deep reinforcement learning framework for multi-scenario games. Knowl.-Based Syst. 2022, 236, 107753. [Google Scholar] [CrossRef]
- Dong, L.; Lin, H.; Qiao, J.; Zhang, T.; Zhang, S.; Pu, T. A coordinated active and reactive power optimization approach for multi-microgrids connected to distribution networks with multi-actor-attention-critic deep reinforcement learning. Appl. Energy 2024, 373, 123870. [Google Scholar]
- Abdessameud, A.; Polushin, I.G.; Tayebi, A. Distributed coordination of dynamical multi-agent systems under directed graphs and constrained information exchange. IEEE Trans. Autom. Control 2017, 62, 1668–1683. [Google Scholar] [CrossRef]
- Li, J.; Zhang, R.; Wang, H.; Liu, Z.; Lai, H.; Zhang, Y. Deep reinforcement learning for voltage control and renewable accommodation using spatial-temporal graph information. IEEE Trans. Sustain. 2024, 15, 249–262. [Google Scholar] [CrossRef]
- Banerjee, S.; Balaban, E.; Shirley, M.; Bradner, K.; Pavone, M. Contingency planning using Bi-level Markov Decision Processes for space missions. In Proceedings of the 2024 IEEE Aerospace Conference, Big Sky, MT, USA, 2–9 March 2024; pp. 1–9. [Google Scholar]
- Wang, K.; Mu, C. Learning-based control with decentralized dynamic event-triggering for vehicle systems. IEEE Trans. Ind. 2023, 19, 2629–2639. [Google Scholar]
- Chi, P.; Wang, Z.; Liao, H.; Li, T.; Wu, X.; Zhang, Q. Application of artificial intelligence in the new generation of underwater humanoid welding robots: A review. Artif. Intell. Rev. 2024, 57, 306. [Google Scholar]
- de Mars, P.; O’Sullivan, A. Applying reinforcement learning and tree search to the unit commitment problem. Appl. Energy 2021, 302, 117519. [Google Scholar]
- Mu, C.; Peng, J.; Sun, C. Hierarchical multiagent formation control scheme via Actor-Critic learning. IEEE Trans. Neural Networks Learn. Syst. 2023, 34, 8764–8777. [Google Scholar]
- Zhou, G.; Tian, W.; Buyya, R.; Xue, R.; Song, L. Deep reinforcement learning-based methods for resource scheduling in cloud computing: A review and future directions. Artif. Intell. Rev. 2024, 57, 1–42. [Google Scholar]
- Mu, C.; Wang, K.; Sun, C. Learning control supported by dynamic event communication applying to industrial systems. IEEE Trans. Ind. Inform. 2021, 17, 2325–2335. [Google Scholar]
- Kozlov, A.N.; Tomin, N.V.; Sidorov, D.N.; Lora, E.E.S.; Kurbatsky, V.G. Optimal Operation Control of PV-Biomass Gasifier-Diesel-Hybrid Systems Using Reinforcement Learning Techniques. Energies 2020, 13, 2632. [Google Scholar] [CrossRef]
- Han, Y.; Zhang, K.; Li, H.; Coelho, E.A.A.; Guerrero, J.M. MAS-based distributed coordinated control and optimization in microgrid and microgrid clusters: A comprehensive overview. IEEE Trans. Power Electron. 2018, 33, 6488–6508. [Google Scholar]
- Xu, Y.; Dong, Z.; Li, Z.; Liu, Y.; Ding, Z. Distributed optimization for integrated frequency regulation and economic dispatch in microgrids. IEEE Trans. Smart Grid 2021, 12, 4595–4606. [Google Scholar]
- Xu, Y.; Dong, Z.; Li, Z.; Liu, Y.; Ding, Z. Day-ahead optimal dispatching of hybrid power system based ondeep reinforcement learning. Cogn. Comput. Syst. 2022, 4, 351–361. [Google Scholar]
- Mu, C.; Wang, K.; Qiu, T. Dynamic event-triggering neural learning control for partially unknown nonlinear systems. IEEE Trans. On Cybern. 2022, 52, 2200–2213. [Google Scholar]
- Guo, G.; Zhang, M.; Gong, Y.; Xu, Q. Safe multi-agent deep reinforcement learning for real-time decentralized control of inverter based renewable energy resources considering communication delay. Appl. Energy 2023, 349, 121648. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Xu, N.; Tang, Z.; Si, C.; Bian, J.; Mu, C. A Review of Smart Grid Evolution and Reinforcement Learning: Applications, Challenges and Future Directions. Energies 2025, 18, 1837. https://doi.org/10.3390/en18071837
Xu N, Tang Z, Si C, Bian J, Mu C. A Review of Smart Grid Evolution and Reinforcement Learning: Applications, Challenges and Future Directions. Energies. 2025; 18(7):1837. https://doi.org/10.3390/en18071837
Chicago/Turabian StyleXu, Na, Zhuo Tang, Chenyi Si, Jinshan Bian, and Chaoxu Mu. 2025. "A Review of Smart Grid Evolution and Reinforcement Learning: Applications, Challenges and Future Directions" Energies 18, no. 7: 1837. https://doi.org/10.3390/en18071837
APA StyleXu, N., Tang, Z., Si, C., Bian, J., & Mu, C. (2025). A Review of Smart Grid Evolution and Reinforcement Learning: Applications, Challenges and Future Directions. Energies, 18(7), 1837. https://doi.org/10.3390/en18071837