ترغب بنشر مسار تعليمي؟ اضغط هنا

We apply the generalized conditional gradient algorithm to potential mean field games and we show its well-posedeness. It turns out that this method can be interpreted as a learning method called fictitious play. More precisely, each step of the gene ralized conditional gradient method amounts to compute the best-response of the representative agent, for a predicted value of the coupling terms of the game. We show that for the learning sequence $delta_k = 2/(k+2)$, the potential cost converges in $O(1/k)$, the exploitability and the variables of the problem (distribution, congestion, price, value function and control terms) converge in $O(1/sqrt{k})$, for specific norms.
We propose and investigate a general class of discrete time and finite state space mean field game (MFG) problems with potential structure. Our model incorporates interactions through a congestion term and a price variable. It also allows hard constr aints on the distribution of the agents. We analyze the connection between the MFG problem and two optimal control problems in duality. We present two families of numerical methods and detail their implementation: (i) primal-dual proximal methods (and their extension with nonlinear proximity operators), (ii) the alternating direction method of multipliers (ADMM) and a variant called ADM-G. We give some convergence results. Numerical results are provided for two examples with hard constraints.
The objective of this paper is to analyze the existence of equilibria for a class of deterministic mean field games of controls. The interaction between players is due to both a congestion term and a price function which depends on the distributions of the optimal strategies. Moreover, final state and mixed state-control constraints are considered, the dynamics being nonlinear and affine with respect to the control. The existence of equilibria is obtained by Kakutanis theorem, applied to a fixed point formulation of the problem. Finally, uniqueness results are shown under monotonicity assumptions.
The value function associated with an optimal control problem subject to the Navier-Stokes equations in dimension two is analyzed. Its smoothness is established around a steady state, moreover, its derivatives are shown to satisfy a Riccati equation at the order two and generalized Lyapunov equations at the higher orders. An approximation of the optimal feedback law is then derived from the Taylor expansion of the value function. A convergence rate for the resulting controls and closed-loop systems is demonstrated.
Optimal control problems with a very large time horizon can be tackled with the Receding Horizon Control (RHC) method, which consists in solving a sequence of optimal control problems with small prediction horizon. The main result of this article is the proof of the exponential convergence (with respect to the prediction horizon) of the control generated by the RHC method towards the exact solution of the problem. The result is established for a class of infinite-dimensional linear-quadratic optimal control problems with time-independent dynamics and integral cost. Such problems satisfy the turnpike property: the optimal trajectory remains most of the time very close to the solution to the associated static optimization problem. Specific terminal cost functions, derived from the Lagrange multiplier associated with the static optimization problem, are employed in the implementation of the RHC method.
A class of optimal control problems of hybrid nature governed by semilinear parabolic equations is considered. These problems involve the optimization of switching times at which the dynamics, the integral cost, and the bounds on the control may chan ge. First- and second-order optimality conditions are derived. The analysis is based on a reformulation involving a judiciously chosen transformation of the time domains. For autonomous systems and time-independent integral cost, we prove that the Hamiltonian is constant in time when evaluated along the optimal controls and trajectories. A numerical example is provided.
An optimal control problem with a time-parameter is considered. The functional to be optimized includes the maximum over time-horizon reached by a function of the state variable, and so an $L^infty$-term. In addition to the classical control function , the time at which this maximum is reached is considered as a free parameter. The problem couples the behavior of the state and the control, with this time-parameter. A change of variable is introduced to derive first and second-order optimality conditions. This allows the implementation of a Newton method. Numerical simulations are developed, for selected ordinary differential equations and a partial differential equation, which illustrate the influence of the additional parameter and the original motivation.
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا