Assessment of Low-Carbon Flexibility in Self-Organized Virtual Power Plants Using Multi-Agent Reinforcement Learning

Gengsheng He*, Yu Huang, Guori Huang, Xi Liu, Pei Li, Yan Zhang

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Virtual power plants (VPPs) aggregate a large number of distributed energy resources (DERs) through IoT technology to provide flexibility to the grid. It is an effective means to promote the utilization of renewable energy, and enable carbon neutrality for future power systems. This paper addresses the evaluation issue of DERs‘ low-carbon benefits, proposes a flexibility assessment model for self-organized VPP to quantify the low-carbon value of DERs’ response behavior in different time periods. Firstly, we introduce the definition of zero-carbon index based on the curve simultaneous rate of renewable energy and load demand. Then, we establish a multi-level self-organized aggregation method for virtual power plants, define the basic rules of DER, and characterize its self-organized aggregation as a Markov game process. Moreover, we use QMIX to achieve a bottom-up, hierarchical construction of VPP from simple to complex. Experimental results show that when users track the zero-carbon curve, they can achieve zero carbon emissions without reducing the overall load, significantly enhancing the grid’s regulation capabilities and the consumption of renewable energy. Additionally, self-organized algorithms can optimize the combinations of DERs to improve the coordination efficiency of VPPs in complex environments.

Original languageEnglish
Article number3688
JournalEnergies
Volume17
Issue number15
DOIs
Publication statusPublished - Aug 2024
Externally publishedYes

Keywords

  • distributed energy resources
  • flexibility
  • self-organized aggregation
  • virtual power plant
  • zero-carbon index

Fingerprint

Dive into the research topics of 'Assessment of Low-Carbon Flexibility in Self-Organized Virtual Power Plants Using Multi-Agent Reinforcement Learning'. Together they form a unique fingerprint.

Cite this