Do you want to publish a course? Click here

Unified Riccati theory for optimal permanent and sampled-data control problems in finite and infinite time horizons

77   0   0.0 ( 0 )
 Added by Emmanuel Trelat
 Publication date 2020
  fields
and research's language is English
 Authors Loic Bourdin




Ask ChatGPT about the research

We revisit and extend the Riccati theory, unifying continuous-time linear-quadratic optimal permanent and sampled-data control problems, in finite and infinite time horizons. In a nutshell, we prove that:-- when the time horizon T tends to $+infty$, one passes from the Sampled-Data Difference Riccati Equation (SD-DRE) to the Sampled-Data Algebraic Riccati Equation (SD-ARE), and from the Permanent Differential Riccati Equation (P-DRE) to the Permanent Algebraic Riccati Equation (P-ARE);-- when the maximal step of the time partition $Delta$ tends to $0$, one passes from (SD-DRE) to (P-DRE), and from (SD-ARE) to (P-ARE).Our notations and analysis provide a unified framework in order to settle all corresponding results.



rate research

Read More

We use the continuation and bifurcation package pde2path to numerically analyze infinite time horizon optimal control problems for parabolic systems of PDEs. The basic idea is a two step approach to the canonical systems, derived from Pontryagins maximum principle. First we find branches of steady or time-periodic states of the canonical systems, i.e., canonical steady states (CSS) respectively canonical periodic states (CPS), and then use these results to compute time-dependent canonical paths connecting to a CSS or a CPS with the so called saddle point property. This is a (high dimensional) boundary value problem in time, which we solve by a continuation algorithm in the initial states. We first explain the algorithms and then the implementation via some example problems and associated pde2path demo directories. The first two examples deal with the optimal management of a distributed shallow lake, and of a vegetation system, both with (spatially, and temporally) distributed controls. These examples show interesting bifurcations of so called patterned CSS, including patterned optimal steady states. As a third example we discuss optimal boundary control of a fishing problem with boundary catch. For the case of CPS-targets we first focus on an ODE toy model to explain and validate the method, and then discuss an optimal pollution mitigation PDE model.
104 - Chen Wang , Shuai Li , Weiguo Xia 2019
We study the formation control problem for a group of mobile agents in a plane, in which each agent is modeled as a kinematic point and can only use the local measurements in its local frame. The agents are required to maintain a geometric pattern while keeping a desired distance to a static/moving target. The prescribed formation is a general one which can be any geometric pattern, and the neighboring relationship of the N-agent system only has the requirement of containing a directed spanning tree. To solve the formation control problem, a distributed controller is proposed based on the idea of decoupled design. One merit of the controller is that it only uses each agents local measurements in its local frame, so that a practical issue that the lack of a global coordinate frame or a common reference direction for real multi-robot systems is successfully solved. Considering another practical issue of real robotic applications that sampled data is desirable instead of continuous-time signals, the sampled-data based controller is developed. Theoretical analysis of the convergence to the desired formation is provided for the multi-agent system under both the continuous-time controller with a static/moving target and the sampled-data based one with a static target. Numerical simulations are given to show the effectiveness and performance of the controllers.
We establish existence and uniqueness for infinite dimensional Riccati equations taking values in the Banach space L 1 ($mu$ $otimes$ $mu$) for certain signed matrix measures $mu$ which are not necessarily finite. Such equations can be seen as the infinite dimensional analogue of matrix Riccati equations and they appear in the Linear-Quadratic control theory of stochastic Volterra equations.
We reconsider the variational integration of optimal control problems for mechanical systems based on a direct discretization of the Lagrange-dAlembert principle. This approach yields discrete dynamical constraints which by construction preserve important structural properties of the system, like the evolution of the momentum maps or the energy behavior. Here, we employ higher order quadrature rules based on polynomial collocation. The resulting variational time discretization decreases the overall computational effort.
We present a Reinforcement Learning (RL) algorithm to solve infinite horizon asymptotic Mean Field Game (MFG) and Mean Field Control (MFC) problems. Our approach can be described as a unified two-timescale Mean Field Q-learning: The emph{same} algorithm can learn either the MFG or the MFC solution by simply tuning the ratio of two learning parameters. The algorithm is in discrete time and space where the agent not only provides an action to the environment but also a distribution of the state in order to take into account the mean field feature of the problem. Importantly, we assume that the agent can not observe the populations distribution and needs to estimate it in a model-free manner. The asymptotic MFG and MFC problems are also presented in continuous time and space, and compared with classical (non-asymptotic or stationary) MFG and MFC problems. They lead to explicit solutions in the linear-quadratic (LQ) case that are used as benchmarks for the results of our algorithm.
comments
Fetching comments Fetching comments
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا