Retour aux activités
Séminaire informel de théorie des systèmes (ISS)

Approximately solving mean field games via entropy-regularized deep reinforcement learning

iCalendar

26 mars 2021   10h00 — 11h00

Kai Cui Technical University of Darmstadt, Allemagne

Kai Cui

Présentation sur YouTube

The recent mean field game (MFG) formalism facilitates otherwise intractable computation of approximate Nash equilibria in many-agent settings. In this paper, we consider discrete-time finite MFGs subject to finite-horizon objectives. We show that all discrete-time finite MFGs with non-constant fixed point operators fail to be contractive as typically assumed in existing MFG literature, barring convergence via fixed point iteration. Instead, we incorporate entropy-regularization and Boltzmann policies into the fixed point iteration. As a result, we obtain provable convergence to approximate fixed points where existing methods fail, and reach the original goal of approximate Nash equilibria. All proposed methods are evaluated with respect to their exploitability, on both instructive examples with tractable exact solutions and high-dimensional problems where exact methods become intractable. In high-dimensional scenarios, we apply established deep reinforcement learning methods and empirically combine fictitious play with our approximations.

Bio: Kai Cui is a PhD candidate at the Bioinspired Communication Systems Lab under supervision of Professor Heinz Koeppl at Technical University of Darmstadt (Germany). Prior to joining BCS, he also received BSc and MSc degrees in Computer Science as well as Electrical Engineering and Information Technology at Technical University of Darmstadt. His current research interests are multi-agent systems, reinforcement learning, mean field games and UAV swarms.

Peter E. Caines responsable
Aditya Mahajan responsable
Shuang Gao responsable
Rinel Foguen Tchuendom responsable
Yaroslav Salii responsable

Lieu

Webinaire
Zoom
Montréal Québec
Canada

Axes de recherche

Applications de recherche