论文标题

稳定性限制了对分散实时电压控制的强化学习

Stability Constrained Reinforcement Learning for Decentralized Real-Time Voltage Control

论文作者

Feng, Jie, Shi, Yuanyuan, Qu, Guannan, Low, Steven H., Anandkumar, Anima, Wierman, Adam

论文摘要

深度强化学习被认为是解决电力系统实时控制挑战的有前途的工具。但是,由于缺乏明确的稳定性和安全保证,它在现实世界电力系统中的部署受到了阻碍。在本文中,我们提出了一种实时电压控制的稳定性约束强化学习(RL)方法,可以保证在策略学习和学习政策的部署过程中系统稳定性。我们方法基础的关键思想是明确构建的Lyapunov功能,它导致稳定策略的足够结构条件,即单调降低政策保证稳定性。我们通过使用单调神经网络对每个局部电压控制器进行参数化将每个局部电压控制器进行参数化,从而确保通过设计满足稳定性约束。我们证明了方法在单相和三相IEEE测试馈线中的有效性,与广泛使用的线性策略相比,该方法可以将瞬态控制成本降低超过25%,并将电压恢复时间平均缩短21.5%,而始终达到电压稳定性。相反,标准RL方法通常无法实现电压稳定性。

Deep reinforcement learning has been recognized as a promising tool to address the challenges in real-time control of power systems. However, its deployment in real-world power systems has been hindered by a lack of explicit stability and safety guarantees. In this paper, we propose a stability-constrained reinforcement learning (RL) method for real-time voltage control, that guarantees system stability both during policy learning and deployment of the learned policy. The key idea underlying our approach is an explicitly constructed Lyapunov function that leads to a sufficient structural condition for stabilizing policies, i.e., monotonically decreasing policies guarantee stability. We incorporate this structural constraint with RL, by parameterizing each local voltage controller using a monotone neural network, thus ensuring the stability constraint is satisfied by design. We demonstrate the effectiveness of our approach in both single-phase and three-phase IEEE test feeders, where the proposed method can reduce the transient control cost by more than 25% and shorten the voltage recovery time by 21.5% on average compared to the widely used linear policy, while always achieving voltage stability. In contrast, standard RL methods often fail to achieve voltage stability.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源