Do you want to publish a course? Click here

An entropy minimization approach to second-order variational mean-field games

62   0   0.0 ( 0 )
 Added by Luca Nenna
 Publication date 2018
  fields
and research's language is English




Ask ChatGPT about the research

We propose a new viewpoint on variational mean-field games with diffusion and quadratic Hamiltonian. We show the equivalence of such mean-field games with a relative entropy minimization at the level of probabilities on curves. We also address the time-discretization of such problems, establish $Gamma$-convergence results as the time step vanishes and propose an efficient algorithm relying on this entropic interpretation as well as on the Sinkhorn scaling algorithm.



rate research

Read More

We analyze a (possibly degenerate) second order mean field games system of partial differential equations. The distinguishing features of the model considered are (1) that it is not uniformly parabolic, including the first order case as a possibility, and (2) the coupling is a local operator on the density. As a result we look for weak, not smooth, solutions. Our main result is the existence and uniqueness of suitably defined weak solutions, which are characterized as minimizers of two optimal control problems. We also show that such solutions are stable with respect to the data, so that in particular the degenerate case can be approximated by a uniformly parabolic (viscous) perturbation.
We study a general class of entropy-regularized multi-variate LQG mean field games (MFGs) in continuous time with $K$ distinct sub-population of agents. We extend the notion of actions to action distributions (exploratory actions), and explicitly derive the optimal action distributions for individual agents in the limiting MFG. We demonstrate that the optimal set of action distributions yields an $epsilon$-Nash equilibrium for the finite-population entropy-regularized MFG. Furthermore, we compare the resulting solutions with those of classical LQG MFGs and establish the equivalence of their existence.
Entropy regularization has been extensively adopted to improve the efficiency, the stability, and the convergence of algorithms in reinforcement learning. This paper analyzes both quantitatively and qualitatively the impact of entropy regularization for Mean Field Game (MFG) with learning in a finite time horizon. Our study provides a theoretical justification that entropy regularization yields time-dependent policies and, furthermore, helps stabilizing and accelerating convergence to the game equilibrium. In addition, this study leads to a policy-gradient algorithm for exploration in MFG. Under this algorithm, agents are able to learn the optimal exploration scheduling, with stable and fast convergence to the game equilibrium.
Mean Field Games with state constraints are differential games with infinitely many agents, each agent facing a constraint on his state. The aim of this paper is to provide a meaning of the PDE system associated with these games, the so-called Mean Field Game system with state constraints. For this, we show a global semiconvavity property of the value function associated with optimal control problems with state constraints.
Mean field games (MFGs) and the best reply strategy (BRS) are two methods of describing competitive optimisation of systems of interacting agents. The latter can be interpreted as an approximation of the respective MFG system. In this paper we present a systematic analysis and comparison of the two approaches in the stationary case. We provide novel existence and uniqueness results for the stationary boundary value problems related to the MFG and BRS formulations, and we present an analytical and numerical comparison of the two paradigms in a variety of modelling situations.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا