ترغب بنشر مسار تعليمي؟ اضغط هنا

Linear-Quadratic Optimal Control in Maximal Coordinates

176   0   0.0 ( 0 )
 نشر من قبل Jan Br\\\"udigam
 تاريخ النشر 2020
  مجال البحث
والبحث باللغة English




اسأل ChatGPT حول البحث

The linear-quadratic regulator (LQR) is an efficient control method for linear and linearized systems. Typically, LQR is implemented in minimal coordinates (also called generalized or joint coordinates). However, other coordinates are possible and recent research suggests that there may be numerical and control-theoretic advantages when using higher-dimensional non-minimal state parameterizations for dynamical systems. One such parameterization is maximal coordinates, in which each link in a multi-body system is parameterized by its full six degrees of freedom and joints between links are modeled with algebraic constraints. Such constraints can also represent closed kinematic loops or contact with the environment. This paper investigates the difference between minimal- and maximal-coordinate LQR control laws. A case study of applying LQR to a simple pendulum and simulations comparing the basins of attraction and tracking performance of minimal- and maximal-coordinate LQR controllers suggest that maximal-coordinate LQR achieves greater robustness and improved tracking performance compared to minimal-coordinate LQR when applied to nonlinear systems.



قيم البحث

اقرأ أيضاً

95 - Jingrui Sun , Zhen Wu , Jie Xiong 2021
This paper is concerned with a backward stochastic linear-quadratic (LQ, for short) optimal control problem with deterministic coefficients. The weighting matrices are allowed to be indefinite, and cross-product terms in the control and state process es are present in the cost functional. Based on a Hilbert space method, necessary and sufficient conditions are derived for the solvability of the problem, and a general approach for constructing optimal controls is developed. The crucial step in this construction is to establish the solvability of a Riccati-type equation, which is accomplished under a fairly weak condition by investigating the connection with forward stochastic LQ optimal control problems.
This paper is concerned with the distributed linear quadratic optimal control problem. In particular, we consider a suboptimal version of the distributed optimal control problem for undirected multi-agent networks. Given a multi-agent system with ide ntical agent dynamics and an associated global quadratic cost functional, our objective is to design suboptimal distributed control laws that guarantee the controlled network to reach consensus and the associated cost to be smaller than an a priori given upper bound. We first analyze the suboptimality for a given linear system and then apply the results to linear multiagent systems. Two design methods are then provided to compute such suboptimal distributed controllers, involving the solution of a single Riccati inequality of dimension equal to the dimension of the agent dynamics, and the smallest nonzero and the largest eigenvalue of the graph Laplacian. Furthermore, we relax the requirement of exact knowledge of the smallest nonzero and largest eigenvalue of the graph Laplacian by using only lower and upper bounds on these eigenvalues. Finally, a simulation example is provided to illustrate our design method.
In this paper we consider the distributed linear quadratic control problem for networks of agents with single integrator dynamics. We first establish a general formulation of the distributed LQ problem and show that the optimal control gain depends o n global information on the network. Thus, the optimal protocol can only be computed in a centralized fashion. In order to overcome this drawback, we propose the design of protocols that are computed in a decentralized way. We will write the global cost functional as a sum of local cost functionals, each associated with one of the agents. In order to achieve good performance of the controlled network, each agent then computes its own local gain, using sampled information of its neighboring agents. This decentralized computation will only lead to suboptimal global network behavior. However, we will show that the resulting network will reach consensus. A simulation example is provided to illustrate the performance of the proposed protocol.
100 - Na Li , Xun Li , Jing Peng 2020
This paper applies a reinforcement learning (RL) method to solve infinite horizon continuous-time stochastic linear quadratic problems, where drift and diffusion terms in the dynamics may depend on both the state and control. Based on Bellmans dynami c programming principle, an online RL algorithm is presented to attain the optimal control with just partial system information. This algorithm directly computes the optimal control rather than estimating the system coefficients and solving the related Riccati equation. It just requires local trajectory information, greatly simplifying the calculation processing. Two numerical examples are carried out to shed light on our theoretical findings.
120 - Shanjian Tang 2014
We are concerned with the linear-quadratic optimal stochastic control problem with random coefficients. Under suitable conditions, we prove that the value field $V(t,x,omega), (t,x,omega)in [0,T]times R^ntimes Omega$, is quadratic in $x$, and has the following form: $V(t,x)=langle K_tx, xrangle$ where $K$ is an essentially bounded nonnegative symmetric matrix-valued adapted processes. Using the dynamic programming principle (DPP), we prove that $K$ is a continuous semi-martingale of the form $$K_t=K_0+int_0^t , dk_s+sum_{i=1}^dint_0^tL_s^i, dW_s^i, quad tin [0,T]$$ with $k$ being a continuous process of bounded variation and $$Eleft[left(int_0^T|L_s|^2, dsright)^pright] <infty, quad forall pge 2; $$ and that $(K, L)$ with $L:=(L^1, cdots, L^d)$ is a solution to the associated backward stochastic Riccati equation (BSRE), whose generator is highly nonlinear in the unknown pair of processes. The uniqueness is also proved via a localized completion of squares in a self-contained manner for a general BSRE. The existence and uniqueness of adapted solution to a general BSRE was initially proposed by the French mathematician J. M. Bismut (1976, 1978). It had been solved by the author (2003) via the stochastic maximum principle with a viewpoint of stochastic flow for the associated stochastic Hamiltonian system. The present paper is its companion, and gives the {it second but more comprehensive} adapted solution to a general BSRE via the DDP. Further extensions to the jump-diffusion control system and to the general nonlinear control system are possible.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا