Abstract

In this paper, a computationally efficient real-time control of a battery with lookahead state-of-energy constraints in active distribution grids with distributed energy sources is presented. The goal is to follow a previously computed dispatch plan or to optimize a monetary cost from buying and selling power at the point of common coupling. However, the lookahead constraints render the battery decisions non-trivial. The current practice in literature to solve this problem is Model Predictive Control (MPC), which does not scale for large grids. Instead, here, we propose a reinforcement learning approach based on the Deep Deterministic Policy Gradient (DDPG) algorithm. To satisfy the lookahead battery constraints we adapt the experience replay technique used in DDPG. To guarantee the satisfaction of the hard grid constraints, we introduce a safety layer that performs constrained optimization. Our approach does not need forecasts contrary to MPC. We perform evaluations on a realistic grid and comparisons with Lyapunov optimization and MPC. We show that we can achieve costs close to MPC and Lyapunov, while reducing the computational time by multiple orders of magnitude.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call