Scalable and consistent few-shot classification of survey responses using text embeddings
By: Jonas Timmann Mjaaland , Markus Fleten Kreutzer , Halvor Tyseng and more
Potential Business Impact:
Helps sort and understand many answers faster.
Qualitative analysis of open-ended survey responses is a commonly-used research method in the social sciences, but traditional coding approaches are often time-consuming and prone to inconsistency. Existing solutions from Natural Language Processing such as supervised classifiers, topic modeling techniques, and generative large language models have limited applicability in qualitative analysis, since they demand extensive labeled data, disrupt established qualitative workflows, and/or yield variable results. In this paper, we introduce a text embedding-based classification framework that requires only a handful of examples per category and fits well with standard qualitative workflows. When benchmarked against human analysis of a conceptual physics survey consisting of 2899 open-ended responses, our framework achieves a Cohen's Kappa ranging from 0.74 to 0.83 as compared to expert human coders in an exhaustive coding scheme. We further show how performance of this framework improves with fine-tuning of the text embedding model, and how the method can be used to audit previously-analyzed datasets. These findings demonstrate that text embedding-assisted coding can flexibly scale to thousands of responses without sacrificing interpretability, opening avenues for deductive qualitative analysis at scale.
Similar Papers
Labeling Free-text Data using Language Model Ensembles
Computation and Language
Helps computers understand people's thoughts faster.
Beyond the Hype: Embeddings vs. Prompting for Multiclass Classification Tasks
Machine Learning (CS)
Computers can sort jobs better than AI.
Small sample-based adaptive text classification through iterative and contrastive description refinement
Machine Learning (CS)
Teaches computers to sort text without new training.