ConspirED: A Dataset for Cognitive Traits of Conspiracy Theories and Large Language Model Safety
By: Luke Bates , Max Glockner , Preslav Nakov and more
Potential Business Impact:
Helps AI spot fake science stories and their tricks.
Conspiracy theories erode public trust in science and institutions while resisting debunking by evolving and absorbing counter-evidence. As AI-generated misinformation becomes increasingly sophisticated, understanding rhetorical patterns in conspiratorial content is important for developing interventions such as targeted prebunking and assessing AI vulnerabilities. We introduce ConspirED (CONSPIR Evaluation Dataset), which captures the cognitive traits of conspiratorial ideation in multi-sentence excerpts (80--120 words) from online conspiracy articles, annotated using the CONSPIR cognitive framework (Lewandowsky and Cook, 2020). ConspirED is the first dataset of conspiratorial content annotated for general cognitive traits. Using ConspirED, we (i) develop computational models that identify conspiratorial traits and determine dominant traits in text excerpts, and (ii) evaluate large language/reasoning model (LLM/LRM) robustness to conspiratorial inputs. We find that both are misaligned by conspiratorial content, producing output that mirrors input reasoning patterns, even when successfully deflecting comparable fact-checked misinformation.
Similar Papers
Do Androids Dream of Unseen Puppeteers? Probing for a Conspiracy Mindset in Large Language Models
Computation and Language
Computers can be tricked into believing conspiracies.
Modeling Narrative Archetypes in Conspiratorial Narratives: Insights from Singapore-Based Telegram Groups
Artificial Intelligence
Finds secret beliefs hidden in everyday chats.
Just Asking Questions: Doing Our Own Research on Conspiratorial Ideation by Generative AI Chatbots
Computers and Society
AI chatbots sometimes spread fake conspiracy stories.