TY - GEN
T1 - Reactive Power Optimization Strategy for Power Grid with High Proportion of Renewable Energy Based on Reinforcement Learning Algorithm
AU - Chen, Hao
AU - Yang, Nan
AU - Ma, Zeliang
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - As the high proportion of renewable energy grid connection leads to intensified grid voltage fluctuations, traditional reactive power optimization methods have obvious deficiencies in dynamic regulation capability and real-time performance, making it difficult to cope with the voltage over-limit and network loss problems caused by the randomness of wind and solar output. To this end, this paper proposes a reactive power optimization strategy based on the deep deterministic policy gradient (DDPG) algorithm, constructs a hierarchical reinforcement learning framework that includes 12-dimensional state spaces such as node voltage, renewable energy output, and load rate, and continuous/discrete action spaces such as SVG and capacitor banks, designs a composite reward function that integrates voltage deviation, network loss cost, and equipment action penalty, and implements minute-level online training through the interactive interface between the power grid simulation platform PSS/E and Python. Tests in the improved IEEE 33-node system show that this method reduces the voltage deviation rate to 0.41 % and the average network loss to 123.2 kW, verifying that the strategy can effectively coordinate the collaborative control of new energy stations and traditional reactive equipment, providing a new way to solve the problem of dynamic reactive power optimization in high-penetration power grids.
AB - As the high proportion of renewable energy grid connection leads to intensified grid voltage fluctuations, traditional reactive power optimization methods have obvious deficiencies in dynamic regulation capability and real-time performance, making it difficult to cope with the voltage over-limit and network loss problems caused by the randomness of wind and solar output. To this end, this paper proposes a reactive power optimization strategy based on the deep deterministic policy gradient (DDPG) algorithm, constructs a hierarchical reinforcement learning framework that includes 12-dimensional state spaces such as node voltage, renewable energy output, and load rate, and continuous/discrete action spaces such as SVG and capacitor banks, designs a composite reward function that integrates voltage deviation, network loss cost, and equipment action penalty, and implements minute-level online training through the interactive interface between the power grid simulation platform PSS/E and Python. Tests in the improved IEEE 33-node system show that this method reduces the voltage deviation rate to 0.41 % and the average network loss to 123.2 kW, verifying that the strategy can effectively coordinate the collaborative control of new energy stations and traditional reactive equipment, providing a new way to solve the problem of dynamic reactive power optimization in high-penetration power grids.
KW - Action Space Decomposition
KW - Actor-Critic Network
KW - DDPG
KW - Power Grid Reactive Power Optimization
KW - Reinforcement Learning
UR - https://www.scopus.com/pages/publications/105023637484
U2 - 10.1109/AIC66080.2025.11212096
DO - 10.1109/AIC66080.2025.11212096
M3 - Conference contribution
AN - SCOPUS:105023637484
T3 - 2025 IEEE 4th World Conference on Applied Intelligence and Computing, AIC 2025
SP - 897
EP - 902
BT - 2025 IEEE 4th World Conference on Applied Intelligence and Computing, AIC 2025
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 4th IEEE World Conference on Applied Intelligence and Computing, AIC 2025
Y2 - 26 July 2025 through 27 July 2025
ER -