Do you want to publish a course? Click here

Deep Learning-Aided Model Predictive Control of Wind Farms for AGC Considering the Dynamic Wake Effect

75   0   0.0 ( 0 )
 Added by Yiwei Qiu PhD
 Publication date 2020
and research's language is English
 Authors Kaixuan Chen




Ask ChatGPT about the research

To provide automatic generation control (AGC) service, wind farms (WFs) are required to control their operation dynamically to track the time-varying power reference. Wake effects impose significant aerodynamic interactions among turbines, which remarkably influence the WF dynamic power production. The nonlinear and high-dimensional nature of dynamic wake model, however, brings extremely high computation complexity and obscure the design of WF controllers. This paper overcomes the control difficulty brought by the dynamic wake model by proposing a novel control-oriented reduced order WF model and a deep-learning-aided model predictive control (MPC) method. Leveraging recent advances in computational fluid dynamics (CFD) to provide high-fidelity data that simulates WF dynamic wake flows, two novel deep neural network (DNN) architectures are specially designed to learn a dynamic WF reduced-order model (ROM) that can capture the dominant flow dynamics. Then, a novel MPC framework is constructed that explicitly incorporates the obtained WF ROM to coordinate different turbines while considering dynamic wake interactions. The proposed WF ROM and the control method are evaluated in a widely-accepted high-dimensional dynamic WF simulator whose accuracy has been validated by realistic measurement data. A 9-turbine WF case and a larger 25-turbine WF case are studied. By reducing WF model states by many orders of magnitude, the computational burden of the control method is reduced greatly. Besides, through the proposed method, the range of AGC signals that can be tracked by the WF in the dynamic operation is extended compared with the existing greedy controller.

rate research

Read More

This paper introduces HPIPM, a high-performance framework for quadratic programming (QP), designed to provide building blocks to efficiently and reliably solve model predictive control problems. HPIPM currently supports three QP types, and provides interior point method (IPM) solvers as well (partial) condensing routines. In particular, the IPM for optimal control QPs is intended to supersede the HPMPC solver, and it largely improves robustness while keeping the focus on speed. Numerical experiments show that HPIPM reliably solves challenging QPs, and that it outperforms other state-of-the-art solvers in speed.
114 - Christoph Mark , Steven Liu 2021
In this paper, we propose a chance constrained stochastic model predictive control scheme for reference tracking of distributed linear time-invariant systems with additive stochastic uncertainty. The chance constraints are reformulated analytically based on mean-variance information, where we design suitable Probabilistic Reachable Sets for constraint tightening. Furthermore, the chance constraints are proven to be satisfied in closed-loop operation. The design of an invariant set for tracking complements the controller and ensures convergence to arbitrary admissible reference points, while a conditional initialization scheme provides the fundamental property of recursive feasibility. The paper closes with a numerical example, highlighting the convergence to changing output references and empirical constraint satisfaction.
We present a data-driven model predictive control scheme for chance-constrained Markovian switching systems with unknown switching probabilities. Using samples of the underlying Markov chain, ambiguity sets of transition probabilities are estimated which include the true conditional probability distributions with high probability. These sets are updated online and used to formulate a time-varying, risk-averse optimal control problem. We prove recursive feasibility of the resulting MPC scheme and show that the original chance constraints remain satisfied at every time step. Furthermore, we show that under sufficient decrease of the confidence levels, the resulting MPC scheme renders the closed-loop system mean-square stable with respect to the true-but-unknown distributions, while remaining less conservative than a fully robust approach.
Move blocking (MB) is a widely used strategy to reduce the degrees of freedom of the Optimal Control Problem (OCP) arising in receding horizon control. The size of the OCP is reduced by forcing the input variables to be constant over multiple discretization steps. In this paper, we focus on developing computationally efficient MB schemes for multiple shooting based nonlinear model predictive control (NMPC). The degrees of freedom of the OCP is reduced by introducing MB in the shooting step, resulting in a smaller but sparse OCP. Therefore, the discretization accuracy and level of sparsity is maintained. A condensing algorithm that exploits the sparsity structure of the OCP is proposed, that allows to reduce the computation complexity of condensing from quadratic to linear in the number of discretization nodes. As a result, active-set methods with warm-start strategy can be efficiently employed, thus allowing the use of a longer prediction horizon. A detailed comparison between the proposed scheme and the nonuniform grid NMPC is given. Effectiveness of the algorithm in reducing computational burden while maintaining optimization accuracy and constraints fulfillment is shown by means of simulations with two different problems.
In this technical note we analyse the performance improvement and optimality properties of the Learning Model Predictive Control (LMPC) strategy for linear deterministic systems. The LMPC framework is a policy iteration scheme where closed-loop trajectories are used to update the control policy for the next execution of the control task. We show that, when a Linear Independence Constraint Qualification (LICQ) condition holds, the LMPC scheme guarantees strict iterative performance improvement and optimality, meaning that the closed-loop cost evaluated over the entire task converges asymptotically to the optimal cost of the infinite-horizon control problem. Compared to previous works this sufficient LICQ condition can be easily checked, it holds for a larger class of systems and it can be used to adaptively select the prediction horizon of the controller, as demonstrated by a numerical example.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا