The integration of massive distributed energy resources (DERs) brings new challenges for the distribution networks (DNs) operation control. To control DERs effectively, virtual power plants (VPPs) are introduced into DNs. Conventionally, the control strategy of VPPs is formulated as an optimization model without considering voltage control and transactive energy (TE) in DNs. In this article, multiple VPPs cooperative optimization with TE is studied and cast as a nonlinear programming model. It aims to optimize the voltage control of DNs and the operation profits of VPPs. To dynamically capture the optimal operation state of VPPs in DNs, the distribution locational marginal pricing and DN partition are incorporated into the model. In general, this type of model is computationally challenging and mainly solved by approximate algorithms with low efficiency. Recently, multi-agent deep reinforcement learning (MADRL) is emerging as a scalable and promising data-driven approach. We propose an augmented MADRL solution to this problem. In the training stage, one soft actor-critic learning agent augmented by parameter sharing strategy is employed to simultaneously capture the state-action patterns of multiple VPPs. The prioritized experience replay technique is applied to improve learning efficiency and speed. Numerical test results demonstrate the superiority of the proposed method over benchmark methods.
Read full abstract