Depth and Autonomy: A Framework for Evaluating LLM Applications in Social Science Research
By: Ali Sanaei, Ali Rajabzadeh
Potential Business Impact:
Helps scientists use AI without making mistakes.
Large language models (LLMs) are increasingly utilized by researchers across a wide range of domains, and qualitative social science is no exception; however, this adoption faces persistent challenges, including interpretive bias, low reliability, and weak auditability. We introduce a framework that situates LLM usage along two dimensions, interpretive depth and autonomy, thereby offering a straightforward way to classify LLM applications in qualitative research and to derive practical design recommendations. We present the state of the literature with respect to these two dimensions, based on all published social science papers available on Web of Science that use LLMs as a tool and not strictly as the subject of study. Rather than granting models expansive freedom, our approach encourages researchers to decompose tasks into manageable segments, much as they would when delegating work to capable undergraduate research assistants. By maintaining low levels of autonomy and selectively increasing interpretive depth only where warranted and under supervision, one can plausibly reap the benefits of LLMs while preserving transparency and reliability.
Similar Papers
Systematic Framework of Application Methods for Large Language Models in Language Sciences
Computation and Language
Guides scientists to use AI for language study.
Evaluating Medical LLMs by Levels of Autonomy: A Survey Moving from Benchmarks to Applications
Computation and Language
Helps doctors trust AI for patient care.
Knowing Your Uncertainty -- On the application of LLM in social sciences
Computers and Society
Helps scientists trust computer answers about people.