Heterogeneity, reinforcement learning, and chaos in population games.

Journal: Proceedings of the National Academy of Sciences of the United States of America
Published Date:

Abstract

Inspired by the challenges at the intersection of Evolutionary Game Theory and Machine Learning, we investigate a class of discrete-time multiagent reinforcement learning (MARL) dynamics in population/nonatomic congestion games, where agents have diverse beliefs and learn at different rates. These congestion games, a well-studied class of potential games, are characterized by individual agents having negligible effects on system performance, strongly aligned incentives, and well-understood advantageous properties of Nash equilibria. Despite the presence of static Nash equilibria, we demonstrate that MARL dynamics with heterogeneous learning rates can deviate from these equilibria, exhibiting instability and even chaotic behavior and resulting in increased social costs. Remarkably, even within these chaotic regimes, we show that the time-averaged macroscopic behavior converges to exact Nash equilibria, thus linking the microscopic dynamic complexity with traditional equilibrium concepts. By employing dynamical systems techniques, we analyze the interaction between individual-level adaptation and population-level outcomes, paving the way for studying heterogeneous learning dynamics in discrete time across more complex game scenarios.

Authors

  • Jakub Bielawski
    Department of Mathematics, Krakow University of Economics, Kraków 31-510, Poland.
  • Thiparat Chotibut
    Department of Physics, Chula Intelligent and Complex Systems, Chulalongkorn University, Bangkok 10330, Thailand.
  • Fryderyk Falniowski
    Department of Mathematics, Krakow University of Economics, Kraków 31-510, Poland.
  • Michał Misiurewicz
    Department of Mathematical Sciences, Indiana University-Purdue University Indianapolis, Indianapolis, IN 46202.
  • Georgios Piliouras
    Google DeepMind, London EC4A 3TW, United Kingdom.