Score: 0

Does Less Hallucination Mean Less Creativity? An Empirical Investigation in LLMs

Published: December 12, 2025 | arXiv ID: 2512.11509v1

By: Mohor Banerjee , Nadya Yuki Wangsajaya , Syed Ali Redha Alsagoff and more

Potential Business Impact:

Helps AI invent new science ideas, not just facts.

Business Areas:
Virtual Reality Hardware, Software

Large Language Models (LLMs) exhibit remarkable capabilities in natural language understanding and reasoning, but suffer from hallucination: the generation of factually incorrect content. While numerous methods have been developed to reduce hallucinations, their impact on creative generations remains unexplored. This gap is particularly critical for AI-assisted scientific discovery, which requires both factual accuracy and creative hypothesis generation. We investigate how three hallucination-reduction techniques: Chain of Verification (CoVe), Decoding by Contrasting Layers (DoLa), and Retrieval-Augmented Generation (RAG), affect creativity in LLMs. Evaluating multiple model families (LLaMA, Qwen, Mistral) at varying scales (1B - 70B parameters) on two creativity benchmarks (NeoCoder and CS4), we find that these methods have opposing effects on divergent creativity. CoVe enhances divergent thinking, DoLa suppresses it, and RAG shows minimal impact. Our findings provide guidance for selecting appropriate hallucination-reduction methods in scientific applications, where the balance between factual accuracy and creative exploration is crucial.

Country of Origin
πŸ‡ΈπŸ‡¬ Singapore

Page Count
12 pages

Category
Computer Science:
Computation and Language