Discover the SciOpen Platform and Achieve Your Research Goals with Ease.
Search articles, authors, keywords, DOl and etc.
The complexity and uncertainty in power systems cause great challenges to controlling power grids. As a popular data-driven technique, deep reinforcement learning (DRL) attracts attention in the control of power grids. However, DRL has some inherent drawbacks in terms of data efficiency and explainability. This paper presents a novel hierarchical task planning (HTP) approach, bridging planning and DRL, to the task of power line flow regulation. First, we introduce a three-level task hierarchy to model the task and model the sequence of task units on each level as a task planning-Markov decision processes (TP-MDPs). Second, we model the task as a sequential decision-making problem and introduce a higher planner and a lower planner in HTP to handle different levels of task units. In addition, we introduce a two-layer knowledge graph that can update dynamically during the planning procedure to assist HTP. Experimental results conducted on the IEEE 118-bus and IEEE 300-bus systems demonstrate our HTP approach outperforms proximal policy optimization, a state-of-the-art deep reinforcement learning (DRL) approach, improving efficiency by 26.16% and 6.86% on both systems.
A. N. Venkat, I. A. Hiskens, J. B. Rawlings, and S. J. Wright, “Distributed MPC strategies with application to power system automatic generation control,” IEEE Transactions on Control Systems Technology, vol. 16, no. 6, pp. 1192–1206, Nov. 2008.
S. J. Russell, Artificial Intelligence: a Modern Approach. 3rd ed., Upper Saddle River, NJ: Prentice Hall, 2010.
J. Schrittwieser, I. Antonoglou, T. Hubert, K. Simonyan, L. Sifre, S. Schmitt, A. Guez, E. Lockhart, D. Hassabis, T. Graepel, T. Lillicrap, and D. Silver, “Mastering Atari, Go, chess and Shogi by planning with a learned model,” Nature, vol. 588, no. 7839, pp. 604–609, Dec, 2020.
H. J. Zhang, N. Yang, W. Huangfu, K. P. Long, and V. C. M. Leung, “Power control based on deep reinforcement learning for spectrum sharing,” IEEE Transactions on Wireless Communications, vol. 19, no. 6, pp. 4209–4219, Jun. 2020.
H. Tang, K. Lv, B. Bak-Jensen, J. R. Pillai, and Z. F. Wang, “Deep neural network-based hierarchical learning method for dispatch control of multi-regional power grid,” Neural Computing and Applications, vol. 34, no. 7, pp. 5063–5079, Apr. 2022.
Y. Ji, J. H. Wang, J. C. Xu, X. K. Fang, and H. G. Zhang, “Real-time energy management of a microgrid using deep reinforcement learning,” Energies, vol. 12, no. 12, pp. 2291, Jun. 2019.
Y. F. Zhang, Q. Ai and Z. Y. Li, “Intelligent demand response resource trading using deep reinforcement learning,” CSEE Journal of Power and Energy Systems, doi: 10.17775/CSEEJPES.2020.05540.
M. Kamel, R. C. Dai, Y. W. Wang, F. X. Li and G. Y. Liu, “Data-driven and model-based hybrid reinforcement learning to reduce stress on power systems branches,” CSEE Journal of Power and Energy Systems, vol. 7, no. 3, pp. 433–442, May 2021.
J. Li, S. Chen, X. Y. Wang and T. J. Pu, “Load Shedding Control Strategy in Power Grid Emergency State Based on Deep Reinforcement Learning,” CSEE Journal of Power and Energy Systems, vol. 8, no. 4, pp. 1175–1182, Jul. 2022.
R. S. Sutton and A. G. Barto, Reinforcement Learning: an Introduction, 2nd ed., Cambridge: MIT Press, 2018.
T. M. Moerland, J. Broekens, A. Plaat, and C. M. Jonker, “A unifying framework for reinforcement learning and planning,” Frontiers in Artificial Intelligence, vol. 5, p. 908353, 2022.
X. M. Dong, H. Sun, C. F. Wang, Z. H. Yun, Y. M. Wang, P. H. Zhao, Y. Y. Ding, and Y. Wang, “Power flow analysis considering automatic generation control for multi-area interconnection power networks,” IEEE Transactions on Industry Applications, vol. 53, no. 6, pp. 5200–5208, Nov./Dec. 2017.
R. S. Sutton and A. G. Barto, “Introduction to Reinforcement Learning,” Cambridge: MIT Press, 1998.
M. Hanheide, M. Göbelbecker, G. S. Horn, A. Pronobis, K. Sjöö, A. Aydemir, P. Jensfelt, C. Gretton, R. Dearden, M. Janicek, H. Zender, G. J. Kruijff, N. Hawes, and J. L. Wyatt, “Robot task planning and explanation in open and uncertain worlds,” Artificial Intelligence, vol. 247, pp. 119–150, Jun. 2017.
L. X. Yang, Q. Y. Sun, N. Zhang, and Z. W. Liu, “Optimal energy operation strategy for we-energy of energy internet based on hybrid reinforcement learning with human-in-the-loop,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 52, no. 1, pp. 32–42, Jan. 2022.
J. J. Duan, D. Shi, R. S. Diao, H. F. Li, Z. W. Wang, B. Zhang, D. S. Bian, and Z. H. Yi, “Deep-reinforcement-learning-based autonomous voltage control for power grid operations,” IEEE Transactions on Power Systems, vol. 35, no. 1, pp. 814–817, Jan. 2020.
R. Rocchetta, L. Bellani, M. Compare, E. Zio, and E. Patelli, “A reinforcement learning framework for optimal operation and maintenance of power grids,” Applied Energy, vol. 241, pp. 291–301, May 2019.
Q. H. Huang, R. K. Huang, W. T. Hao, J. Tan, R. Fan, and Z. Y. Huang, “Adaptive power system emergency control using deep reinforcement learning,” IEEE Transactions on Smart Grid, vol. 11, no. 2, pp. 1171–1182, Mar. 2020.
C. B. Browne, E. Powley, D. Whitehouse, S. M. Lucas, P. I. Cowling, P. Rohlfshagen, S. Tavener, D. Perez, S. Samothrakis, and S. Colton, “A survey of Monte Carlo tree search methods,” IEEE Transactions on Computational Intelligence and AI in Games, vol. 4, no. 1, pp. 1–43, Mar. 2012.
This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).