No Arabic abstract
In real-world problems, uncertainties (e.g., errors in the measurement, precision errors) often lead to poor performance of numerical algorithms when not explicitly taken into account. This is also the case for control problems, where optimal solutions can degrade in quality or even become infeasible. Thus, there is the need to design methods that can handle uncertainty. In this work, we consider nonlinear multi-objective optimal control problems with uncertainty on the initial conditions, and in particular their incorporation into a feedback loop via model predictive control (MPC). In multi-objective optimal control, an optimal compromise between multiple conflicting criteria has to be found. For such problems, not much has been reported in terms of uncertainties. To address this problem class, we design an offline/online framework to compute an approximation of efficient control strategies. This approach is closely related to explicit MPC for nonlinear systems, where the potentially expensive optimization problem is solved in an offline phase in order to enable fast solutions in the online phase. In order to reduce the numerical cost of the offline phase, we exploit symmetries in the control problems. Furthermore, in order to ensure optimality of the solutions, we include an additional online optimization step, which is considerably cheaper than the original multi-objective optimization problem. We test our framework on a car maneuvering problem where safety and speed are the objectives. The multi-objective framework allows for online adaptations of the desired objective. Alternatively, an automatic scalarizing procedure yields very efficient feedback controls. Our results show that the method is capable of designing driving strategies that deal better with uncertainties in the initial conditions, which translates into potentially safer and faster driving strategies.
We present an algorithm for robust model predictive control with consideration of uncertainty and safety constraints. Our framework considers a nonlinear dynamical system subject to disturbances from an unknown but bounded uncertainty set. By viewing the system as a fixed point of an operator acting over trajectories, we propose a convex condition on control actions that guarantee safety against the uncertainty set. The proposed condition guarantees that all realizations of the state trajectories satisfy safety constraints. Our algorithm solves a sequence of convex quadratic constrained optimization problems of size n*N, where n is the number of states, and N is the prediction horizon in the model predictive control problem. Compared to existing methods, our approach solves convex problems while guaranteeing that all realizations of uncertainty set do not violate safety constraints. Moreover, we consider the implicit time-discretization of system dynamics to increase the prediction horizon and enhance computational accuracy. Numerical simulations for vehicle navigation demonstrate the effectiveness of our approach.
In this paper, we propose a chance constrained stochastic model predictive control scheme for reference tracking of distributed linear time-invariant systems with additive stochastic uncertainty. The chance constraints are reformulated analytically based on mean-variance information, where we design suitable Probabilistic Reachable Sets for constraint tightening. Furthermore, the chance constraints are proven to be satisfied in closed-loop operation. The design of an invariant set for tracking complements the controller and ensures convergence to arbitrary admissible reference points, while a conditional initialization scheme provides the fundamental property of recursive feasibility. The paper closes with a numerical example, highlighting the convergence to changing output references and empirical constraint satisfaction.
In this paper the optimal control of alignment models composed by a large number of agents is investigated in presence of a selective action of a controller, acting in order to enhance consensus. Two types of selective controls have been presented: an homogeneous control filtered by a selective function and a distributed control active only on a selective set. As a first step toward a reduction of computational cost, we introduce a model predictive control (MPC) approximation by deriving a numerical scheme with a feedback selective constrained dynamics. Next, in order to cope with the numerical solution of a large number of interacting agents, we derive the mean-field limit of the feedback selective constrained dynamics, which eventually will be solved numerically by means of a stochastic algorithm, able to simulate efficiently the selective constrained dynamics. Finally, several numerical simulations are reported to show the efficiency of the proposed techniques.
We present an approach for accelerating nonlinear model predictive control. If the current optimal input signal is saturated, also the optimal signals in subsequent time steps often are. We propose to use the open-loop optimal input signals whenever the first and some subsequent input signals are saturated. We only solve the next optimal control problem, when a non-saturated signal is encountered, or the end of the horizon is reached. In this way, we can save a significant number of NLPs to be solved while on the other hand keep the performance loss small. Furthermore, the NMPC is reactivated in time when it comes to controlling the system safely to its reference.
Robust control is a core approach for controlling systems with performance guarantees that are robust to modeling error, and is widely used in real-world systems. However, current robust control approaches can only handle small system uncertainty, and thus require significant effort in system identification prior to controller design. We present an online approach that robustly controls a nonlinear system under large model uncertainty. Our approach is based on decomposing the problem into two sub-problems, robust control design (which assumes small model uncertainty) and chasing consistent models, which can be solved using existing tools from control theory and online learning, respectively. We provide a learning convergence analysis that yields a finite mistake bound on the number of times performance requirements are not met and can provide strong safety guarantees, by bounding the worst-case state deviation. To the best of our knowledge, this is the first approach for online robust control of nonlinear systems with such learning theoretic and safety guarantees. We also show how to instantiate this framework for general robotic systems, demonstrating the practicality of our approach.