Do you want to publish a course? Click here

Optimal Management of the Peak Power Penalty for Smart Grids Using MPC-based Reinforcement Learning

56   0   0.0 ( 0 )
 Added by Wenqi Cai
 Publication date 2021
and research's language is English




Ask ChatGPT about the research

The cost of the power distribution infrastructures is driven by the peak power encountered in the system. Therefore, the distribution network operators consider billing consumers behind a common transformer in the function of their peak demand and leave it to the consumers to manage their collective costs. This management problem is, however, not trivial. In this paper, we consider a multi-agent residential smart grid system, where each agent has local renewable energy production and energy storage, and all agents are connected to a local transformer. The objective is to develop an optimal policy that minimizes the economic cost consisting of both the spot-market cost for each consumer and their collective peak-power cost. We propose to use a parametric Model Predictive Control (MPC)-scheme to approximate the optimal policy. The optimality of this policy is limited by its finite horizon and inaccurate forecasts of the local power production-consumption. A Deterministic Policy Gradient (DPG) method is deployed to adjust the MPC parameters and improve the policy. Our simulations show that the proposed MPC-based Reinforcement Learning (RL) method can effectively decrease the long-term economic cost for this smart grid problem.



rate research

Read More

This paper develops a model-free volt-VAR optimization (VVO) algorithm via multi-agent deep reinforcement learning (MADRL) in unbalanced distribution systems. This method is novel since we cast the VVO problem in unbalanced distribution networks to an intelligent deep Q-network (DQN) framework, which avoids solving a specific optimization model directly when facing time-varying operating conditions of the systems. We consider statuses/ratios of switchable capacitors, voltage regulators, and smart inverters installed at distributed generators as the action variables of the DQN agents. A delicately designed reward function guides these agents to interact with the distribution system, in the direction of reinforcing voltage regulation and power loss reduction simultaneously. The forward-backward sweep method for radial three-phase distribution systems provides accurate power flow results within a few iterations to the DQN environment. Finally, the proposed multi-objective MADRL method realizes the dual goals for VVO. We test this algorithm on the unbalanced IEEE 13-bus and 123-bus systems. Numerical simulations validate the excellent performance of this method in voltage regulation and power loss reduction.
Under voltage load shedding has been considered as a standard and effective measure to recover the voltage stability of the electric power grid under emergency and severe conditions. However, this scheme usually trips a massive amount of load which can be unnecessary and harmful to customers. Recently, deep reinforcement learning (RL) has been regarded and adopted as a promising approach that can significantly reduce the amount of load shedding. However, like most existing machine learning (ML)-based control techniques, RL control usually cannot guarantee the safety of the systems under control. In this paper, we introduce a novel safe RL method for emergency load shedding of power systems, that can enhance the safe voltage recovery of the electric power grid after experiencing faults. Unlike the standard RL method, the safe RL method has a reward function consisting of a Barrier function that goes to minus infinity when the system state goes to the safety bounds. Consequently, the optimal control policy can render the power system to avoid the safety bounds. This method is general and can be applied to other safety-critical control problems. Numerical simulations on the 39-bus IEEE benchmark is performed to demonstrate the effectiveness of the proposed safe RL emergency control, as well as its adaptive capability to faults not seen in the training.
Hybrid-electric propulsion systems powered by clean energy derived from renewable sources offer a promising approach to decarbonise the worlds transportation systems. Effective energy management systems are critical for such systems to achieve optimised operational performance. However, developing an intelligent energy management system for applications such as ships operating in a highly stochastic environment and requiring concurrent control over multiple power sources presents challenges. This article proposes an intelligent energy management framework for hybrid-electric propulsion systems using deep reinforcement learning. In the proposed framework, a Twin-Delayed Deep Deterministic Policy Gradient agent is trained using an extensive volume of historical load profiles to generate a generic energy management strategy. The strategy, i.e. the core of the energy management system, can concurrently control multiple power sources in continuous state and action spaces. The proposed framework is applied to a coastal ferry model with multiple fuel cell clusters and a battery, achieving near-optimal cost performance when applied to novel future voyages.
Building energy management is one of the core problems in modern power grids to reduce energy consumption while ensuring occupants comfort. However, the building energy management system (BEMS) is now facing more challenges and uncertainties with the increasing penetration of renewables and complicated interactions between humans and buildings. Classical model predictive control (MPC) has shown its capacity to reduce building energy consumption, but it suffers from labor-intensive modelling and complex on-line control optimization. Recently, with the growing accessibility to the building control and automation data, data-driven solutions have attracted more research interest. This paper presents a compact review of the recent advances in data-driven MPC and reinforcement learning based control methods for BEMS. The main challenges in these approaches and insights on the selection of a control method are discussed.
83 - Cunlai Pu , Pang Wu 2019
In modern power grids, a local failure or attack can trigger catastrophic cascading failures, which make it challenging to assess the attack vulnerability of power grids. In this Brief, we define the $K$-link attack problem and study the attack vulnerability of power grids under cascading failures. Particularly, we propose a link centrality measure based on both topological and electrical properties of power grids. According to this centrality, we propose a greedy attack algorithm and an optimal attack algorithm. Simulation results on standard IEEE bus test data show that the optimal attack is better than the greedy attack and the traditional PSO-based attack in fracturing power grids. Moreover, the greedy attack has smaller computational complexity than the optimal attack and the PSO-based attack with an adequate attack efficiency. Our work helps to understand the vulnerability of power grids and provides some clues for securing power grids.
comments
Fetching comments Fetching comments
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا