Score: 5

LiteraryTaste: A Preference Dataset for Creative Writing Personalization

Published: November 12, 2025 | arXiv ID: 2511.09310v1

By: John Joon Young Chung , Vishakh Padmakumar , Melissa Roemmele and more

BigTech Affiliations: University of Washington University of California, Berkeley Stanford University

Potential Business Impact:

Teaches computers to write stories people like.

Business Areas:
Reading Apps Apps, Software

People have different creative writing preferences, and large language models (LLMs) for these tasks can benefit from adapting to each user's preferences. However, these models are often trained over a dataset that considers varying personal tastes as a monolith. To facilitate developing personalized creative writing LLMs, we introduce LiteraryTaste, a dataset of reading preferences from 60 people, where each person: 1) self-reported their reading habits and tastes (stated preference), and 2) annotated their preferences over 100 pairs of short creative writing texts (revealed preference). With our dataset, we found that: 1) people diverge on creative writing preferences, 2) finetuning a transformer encoder could achieve 75.8% and 67.7% accuracy when modeling personal and collective revealed preferences, and 3) stated preferences had limited utility in modeling revealed preferences. With an LLM-driven interpretability pipeline, we analyzed how people's preferences vary. We hope our work serves as a cornerstone for personalizing creative writing technologies.

Country of Origin
🇺🇸 United States


Page Count
19 pages

Category
Computer Science:
Computation and Language