ArXiv TLDR

Trust, Lies, and Long Memories: Emergent Social Dynamics and Reputation in Multi-Round Avalon with LLM Agents

🐦 Tweet
2604.20582

Suveen Ellawela

cs.MAcs.AIcs.CL

TLDR

LLM agents playing multi-round Avalon with memory develop emergent reputation dynamics and strategic deception, showing complex social behavior.

Key contributions

  • LLM agents develop cross-game reputation dynamics in repeated Avalon games.
  • Reputations are role-conditional; agents are perceived differently based on their role.
  • High-reputation agents receive 46% more team inclusions, demonstrating trust.
  • Higher reasoning effort enables more strategic deception by evil players (passing early missions).

Why it matters

This paper reveals that LLM agents can develop complex social behaviors like reputation and strategic deception when given memory across repeated interactions. It highlights the potential for advanced social dynamics in AI systems, opening new avenues for studying AI ethics and social intelligence.

Original Abstract

We study emergent social dynamics in LLM agents playing The Resistance: Avalon, a hidden-role deception game. Unlike prior work on single-game performance, our agents play repeated games while retaining memory of previous interactions, including who played which roles and how they behaved, enabling us to study how social dynamics evolve. Across 188 games, two key phenomena emerge. First, reputation dynamics emerge organically when agents retain cross-game memory: agents reference past behavior in statements like "I am wary of repeating last game's mistake of over-trusting early success." These reputations are role-conditional: the same agent is described as "straightforward" when playing good but "subtle" when playing evil, and high-reputation players receive 46% more team inclusions. Second, higher reasoning effort supports more strategic deception: evil players more often pass early missions to build trust before sabotaging later ones, 75% in high-effort games vs 36% in low-effort games. Together, these findings show that repeated interaction with memory gives rise to measurable reputation and deception dynamics among LLM agents.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.