Robust preventive and corrective security-constrained OPF for worst contingencies with the adoption of VPP: A safe reinforcement learning approach

Xiang Wei, Ka Wing Chan, Guibin Wang, Ze Hu, Ziqing Zhu, Xian Zhang

Research output: Journal article publicationJournal articleAcademic researchpeer-review

Abstract

The rising frequency of extreme weather events calls for urgent measures to improve the resilience and reliability of power systems. This paper, therefore, presents a robust preventive-corrective security-constrained optimal power flow (PCSCOPF) model designed to strengthen power system reliability during N-k outages. The model integrates fast-response virtual power plants (VPPs), dynamically adjusting their injections to mitigate post-contingency overloads and maintain branch flows within emergency limits. Additionally, a novel approach combining deep reinforcement learning (DRL) with Lagrangian relaxation is introduced to efficiently solve the PCSCOPF decision-making problem. By framing the problem as a constrained Markov decision process (CMDP), the proposed Lagrangian-based soft actor-critic (L-SAC) algorithm optimizes control actions while ensuring constraint satisfaction during the exploration process. Extensive investigations have been conducted on the IEEE 30-bus and 118-bus systems to evaluate their computational efficiency and reliability.

Original languageEnglish
Article number124970
Pages (from-to)1-16
Number of pages16
JournalApplied Energy
Volume380
DOIs
Publication statusPublished - 15 Feb 2025

Keywords

  • Deep reinforcement learning
  • Lagrangian relaxation
  • Security-constrained optimal power flow
  • Virtual power plant

ASJC Scopus subject areas

  • Building and Construction
  • Renewable Energy, Sustainability and the Environment
  • Mechanical Engineering
  • General Energy
  • Management, Monitoring, Policy and Law

Fingerprint

Dive into the research topics of 'Robust preventive and corrective security-constrained OPF for worst contingencies with the adoption of VPP: A safe reinforcement learning approach'. Together they form a unique fingerprint.

Cite this