• Corpus ID: 244488356

Reinforcement Learning for Volt-Var Control: A Novel Two-stage Progressive Training Strategy

  title={Reinforcement Learning for Volt-Var Control: A Novel Two-stage Progressive Training Strategy},
  author={Si Zhang and Mingzhi Zhang and Rongxing Hu and David Lubkeman and Yunan Liu and Ning Lu},
This paper develops a reinforcement learning (RL) approach to solve a cooperative, multi-agent Volt-Var Control (VVC) problem for high solar penetration distribution systems. The ingenuity of our RL method lies in a novel two-stage progressive training strategy that can effectively improve training speed and convergence of the machine learning algorithm. In Stage 1 (individual training), while holding all the other agents inactive, we separately train each agent to obtain its own optimal VVC… 

Figures and Tables from this paper


A Data-Driven Multi-Agent Autonomous Voltage Control Framework Using Deep Reinforcement Learning
A multi-agent AVC (MA-AVC) algorithm based on a multi- agent deep deterministic policy gradient (MADDPG) method that features centralized training and decentralized execution is developed to solve the AVC problem.
Deep Reinforcement Learning Based Volt-VAR Optimization in Smart Distribution Systems
This paper develops a model-free volt-VAR optimization (VVO) algorithm via multi-agent deep reinforcement learning (DRL) in unbalanced distribution systems and realizes the dual goals for VVO.
Two-Stage Volt/Var Control in Active Distribution Networks With Multi-Agent Deep Reinforcement Learning Method
A two-stage deep reinforcement learning (DRL)-based real-time VVC method to mitigate fast voltage violation while minimizing the network power loss is proposed.
Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments
An adaptation of actor-critic methods that considers action policies of other agents and is able to successfully learn policies that require complex multi-agent coordination is presented.
Continuous control with deep reinforcement learning
This work presents an actor-critic, model-free algorithm based on the deterministic policy gradient that can operate over continuous action spaces, and demonstrates that for many of the tasks the algorithm can learn policies end-to-end: directly from raw pixel inputs.
Asynchronous Methods for Deep Reinforcement Learning
A conceptually simple and lightweight framework for deep reinforcement learning that uses asynchronous gradient descent for optimization of deep neural network controllers and shows that asynchronous actor-critic succeeds on a wide variety of continuous motor control problems as well as on a new task of navigating random 3D mazes using a visual input.
IEEE Standard for Interconnection and Interoperability of Distributed Energy Resources with Associated Electric Power Systems Interfaces
  • IEEE, Tech. Rep