Playing repeated games with large language models.

Journal: Nature human behaviour
Published Date:

Abstract

Large language models (LLMs) are increasingly used in applications where they interact with humans and other agents. We propose to use behavioural game theory to study LLMs' cooperation and coordination behaviour. Here we let different LLMs play finitely repeated 2 × 2 games with each other, with human-like strategies, and actual human players. Our results show that LLMs perform particularly well at self-interested games such as the iterated Prisoner's Dilemma family. However, they behave suboptimally in games that require coordination, such as the Battle of the Sexes. We verify that these behavioural signatures are stable across robustness checks. We also show how GPT-4's behaviour can be modulated by providing additional information about its opponent and by using a 'social chain-of-thought' strategy. This also leads to better scores and more successful coordination when interacting with human players. These results enrich our understanding of LLMs' social behaviour and pave the way for a behavioural game theory for machines.

Authors

  • Elif Akata
    Institute for Human-Centered AI, Helmholtz Munich, Oberschleißheim, Germany. elif.akata@helmholtz-munich.de.
  • Lion Schulz
    Max Planck Institute for Biological Cybernetics, Tübingen, Germany.
  • Julian Coda-Forno
    Institute for Human-Centered AI, Helmholtz Munich, Oberschleißheim, Germany.
  • Seong Joon Oh
    University of Tübingen, Tübingen, Germany.
  • Matthias Bethge
    University of Tübingen, Tübingen, Germany.
  • Eric Schulz
    Institute for Human-Centered AI, Helmholtz Munich, Oberschleißheim, Germany.

Keywords

No keywords available for this article.