4.7 Article

Accelerated Derivative-Free Deep Reinforcement Learning for Large-Scale Grid Emergency Voltage Control

Journal

IEEE TRANSACTIONS ON POWER SYSTEMS
Volume 37, Issue 1, Pages 14-25

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TPWRS.2021.3095179

Keywords

Load shedding; Voltage control; Power system stability; Training; Power system dynamics; Power grids; Adaptation models; Deep reinforcement learning; voltage stability; load shedding; augmented random search

Funding

  1. U.S. Department of Energy (DOE) ARPA-E OPEN 2018 Program
  2. U.S. DOE [DE-AC0576RL01830]

Ask authors/readers for more resources

This study proposes a novel derivative-free DRL algorithm called PARS, tailored for load shedding in power system voltage control. It shows better computational efficiency, robustness in learning, scalability, and generalization capability compared to other methods.
Load shedding has been one of the most widely used and effective emergency control approaches against voltage instability. With increased uncertainties and rapidly changing operational conditions in power systems, existing methods have outstanding issues in terms of either speed, adaptiveness, or scalability. Deep reinforcement learning (DRL) was regarded and adopted as a promising approach for fast and adaptive grid stability control in recent years. However, existing DRL algorithms show two outstanding issues when being applied to power system control problems: 1) computational inefficiency that requires extensive training and tuning time; and 2) poor scalability making it difficult to scale to high dimensional control problems. To overcome these issues, an novel derivative-free DRL algorithm named PARS was developed and tailored for power system voltage stability control via load shedding. The method was tested on both the IEEE 39-bus and IEEE 300-bus systems, and the latter is by far the largest scale for such a study. Test results show that, compared to other methods including model-predictive control (MPC) and proximal policy optimization(PPO) methods, PARS shows better computational efficiency (faster convergence), more robustness in learning, excellent scalability and generalization capability.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available