Evolving general cooperation with a Bayesian theory of mind | PNAS

5 views
Skip to first unread message

David Rand

unread,
Jul 22, 2025, 7:16:30 AM7/22/25
to Human Cooperation Lab
https://www.pnas.org/doi/10.1073/pnas.2400993122 

Evolving general cooperation with a Bayesian theory of mind

Edited by Udari M. Sehwag, Stanford University, Stanford, CA; received January 16, 2024; accepted December 26, 2024, by Editorial Board Member Elke U. Weber
June 16, 2025
122 (25) e2400993122

Theories of the evolution of cooperation through reciprocity explain how unrelated self-interested individuals can accomplish more together than they can on their own. The most prominent theories of reciprocity, such as tit-for-tat or win-stay-lose-shift, are inflexible automata that lack a theory of mind—the human ability to infer the hidden mental states in others’ minds. Here, we develop a model of reciprocity with a theory of mind, the Bayesian Reciprocator. When making decisions, this model does not simply seek to maximize its own payoff. Instead, it also values the payoffs of others—but only to the extent it believes that those others are also cooperating in the same way. To compute its beliefs about others, the Bayesian Reciprocator uses a probabilistic and generative approach to infer the latent preferences, beliefs, and strategies of others through interaction and observation. We evaluate the Bayesian Reciprocator using a generator over games where every interaction is unique, as well as in classic environments such as the iterated prisoner’s dilemma. The Bayesian Reciprocator enables the emergence of both direct-reciprocity when games are repeated and indirect-reciprocity when interactions are one-shot but observable to others. In an evolutionary competition, the Bayesian Reciprocator outcompetes existing automata strategies and sustains cooperation across a larger range of environments and noise settings than prior approaches. This work quantifies the advantage of a theory of mind for cooperation in an evolutionary game theoretic framework and suggests avenues for building artificially intelligent agents with more human-like learning mechanisms that can cooperate across many environments.

--
David G. Rand (he/him)
Information Science & Marketing and Management Communication
Cornell University

Recent publication: Political motives help rather than hinder crowdsourced fact-checking Working paper

All misinformation/fake news related papers: [Dynamic list with PDF and Twitter links
Reply all
Reply to author
Forward
0 new messages