Score: 0

Valid Survey Simulations with Limited Human Data: The Roles of Prompting, Fine-Tuning, and Rectification

Published: October 13, 2025 | arXiv ID: 2510.11408v1

By: Stefan Krsteski , Giuseppe Russo , Serina Chang and more

Potential Business Impact:

Makes surveys cheaper and more accurate.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Surveys provide valuable insights into public opinion and behavior, but their execution is costly and slow. Large language models (LLMs) have been proposed as a scalable, low-cost substitute for human respondents, but their outputs are often biased and yield invalid estimates. We study the interplay between synthesis methods that use LLMs to generate survey responses and rectification methods that debias population estimates, and explore how human responses are best allocated between them. Using two panel surveys with questions on nutrition, politics, and economics, we find that synthesis alone introduces substantial bias (24-86%), whereas combining it with rectification reduces bias below 5% and increases effective sample size by up to 14%. Overall, we challenge the common practice of using all human responses for fine-tuning, showing that under a fixed budget, allocating most to rectification results in far more effective estimation.

Country of Origin
🇨🇭 Switzerland

Page Count
19 pages

Category
Computer Science:
Computation and Language