Skip to main content
eScholarship
Open Access Publications from the University of California

The Wisdom of Partisan Crowds: Comparing Collective Intelligence in Humans and LLM-based Agents

Creative Commons 'BY' version 4.0 license
Abstract

Human groups are able to converge to more accurate beliefs through deliberation, even in the presence of polarization and partisan bias --- a phenomenon known as the ``wisdom of partisan crowds.'' Large Language Models (LLMs) are increasingly being used to simulate human collective behavior, yet few benchmarks exist for evaluating their dynamics against the behavior of human groups. In this paper, we examine the extent to which the wisdom of partisan crowds emerges in groups of LLM-based agents that are prompted to role-play as partisan personas (e.g., Democrat or Republican). We find that they not only display human-like partisan biases, but also converge to more accurate beliefs through deliberation, as humans do. We then identify several factors that interfere with convergence, including the use of chain-of-thought prompting and lack of details in personas. Conversely, fine-tuning on human data appears to enhance convergence. These findings show the potential and limitations of LLM-based agents as a model of human collective intelligence.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View