TY - UNPB
T1 - Algorithmic Learning in Local and Global Public Goods Games
AU - Herings, P.J.J.
AU - Michaelides, Philippos
AU - Seel, Christian
N1 - CentER Discussion Paper Nr. 2026-002
PY - 2026/1/26
Y1 - 2026/1/26
N2 - In this paper, we consider two variations of an infinitely repeated public goods game. In the global version, the agents’ stage-game rewards depend on the contributions of all other agents, while in the local variant, introduced by Eshel, Samuelson, and Shaked (1998), these rewards depend only on the contributions of their neighbors. We define three nested solution concepts: one-shot deviation optimality (OSDO), Nash equilibrium in private stationary strategies (SNE), and subgame-perfect Nash equilibrium in private stationary strategies (SSPE), and derive all symmetric SSPEs for the local and global public goods games, depending on the altruism cost, discount factor, and tremble probability. For each solution concept, we develop a measure to determine how close a given strategy profile is to satisfying the solution concept. We then simulate the outcomes of interactions between Q-learning agents under different parameter settings. Almost no simulations converge to an SSPE. In the global variant, we obtain moderate convergence rates to an SNE, whereas, especially in the local variant, many simulation outcomes fail even to satisfy OSDO. Typically, Q-learners end up with conditional cooperation strategies supported by moderate punishment. Coordination on these strategies is easier in the global public goods game. Consequently, unlike learning-by-imitation, Q-learning results in higher cooperation levels in the global variant
AB - In this paper, we consider two variations of an infinitely repeated public goods game. In the global version, the agents’ stage-game rewards depend on the contributions of all other agents, while in the local variant, introduced by Eshel, Samuelson, and Shaked (1998), these rewards depend only on the contributions of their neighbors. We define three nested solution concepts: one-shot deviation optimality (OSDO), Nash equilibrium in private stationary strategies (SNE), and subgame-perfect Nash equilibrium in private stationary strategies (SSPE), and derive all symmetric SSPEs for the local and global public goods games, depending on the altruism cost, discount factor, and tremble probability. For each solution concept, we develop a measure to determine how close a given strategy profile is to satisfying the solution concept. We then simulate the outcomes of interactions between Q-learning agents under different parameter settings. Almost no simulations converge to an SSPE. In the global variant, we obtain moderate convergence rates to an SNE, whereas, especially in the local variant, many simulation outcomes fail even to satisfy OSDO. Typically, Q-learners end up with conditional cooperation strategies supported by moderate punishment. Coordination on these strategies is easier in the global public goods game. Consequently, unlike learning-by-imitation, Q-learning results in higher cooperation levels in the global variant
KW - Public goods
KW - network
KW - equilibrium
KW - Q-learning
M3 - Discussion paper
VL - 2026-002
T3 - CentER Discussion Paper
SP - 1
EP - 46
BT - Algorithmic Learning in Local and Global Public Goods Games
PB - CentER, Center for Economic Research
CY - Tilburg
ER -