Towards Strategic Persuasion with Language Models
By: Zirui Cheng, Jiaxuan You
Potential Business Impact:
Teaches computers to convince people better.
Large language models (LLMs) have demonstrated strong persuasive capabilities comparable to those of humans, offering promising benefits while raising societal concerns about their deployment. However, systematically evaluating the persuasive capabilities of LLMs is inherently challenging, as the effectiveness of persuasion among humans varies significantly across different domains. In this paper, we take a theory-driven approach to provide a scalable and principled framework for measuring the persuasive capabilities of LLMs. Grounded in the Bayesian Persuasion (BP) framework, we repurpose existing human-human persuasion datasets to construct environments for evaluating and training LLMs in strategic persuasion. Our results reveal that frontier models can consistently achieve high persuasion gains and exhibit sophisticated persuasion strategies that align with theoretical predictions. Building on this, we use reinforcement learning to train LLMs for strategic persuasion in our environments. Our results also demonstrate that even small LLMs can obtain significantly higher persuasion gains through reinforcement learning.
Similar Papers
A Meta-Analysis of the Persuasive Power of Large Language Models
Human-Computer Interaction
Computers persuade people as well as humans.
Large Language Models Are More Persuasive Than Incentivized Human Persuaders
Computation and Language
AI can trick people better than humans.
A Framework to Assess the Persuasion Risks Large Language Model Chatbots Pose to Democratic Societies
Computation and Language
Computers can now convince voters cheaper than ads.