Does Less Hallucination Mean Less Creativity? An Empirical Investigation in LLMs
By: Mohor Banerjee , Nadya Yuki Wangsajaya , Syed Ali Redha Alsagoff and more
Potential Business Impact:
Helps AI invent new science ideas, not just facts.
Large Language Models (LLMs) exhibit remarkable capabilities in natural language understanding and reasoning, but suffer from hallucination: the generation of factually incorrect content. While numerous methods have been developed to reduce hallucinations, their impact on creative generations remains unexplored. This gap is particularly critical for AI-assisted scientific discovery, which requires both factual accuracy and creative hypothesis generation. We investigate how three hallucination-reduction techniques: Chain of Verification (CoVe), Decoding by Contrasting Layers (DoLa), and Retrieval-Augmented Generation (RAG), affect creativity in LLMs. Evaluating multiple model families (LLaMA, Qwen, Mistral) at varying scales (1B - 70B parameters) on two creativity benchmarks (NeoCoder and CS4), we find that these methods have opposing effects on divergent creativity. CoVe enhances divergent thinking, DoLa suppresses it, and RAG shows minimal impact. Our findings provide guidance for selecting appropriate hallucination-reduction methods in scientific applications, where the balance between factual accuracy and creative exploration is crucial.
Similar Papers
Shakespearean Sparks: The Dance of Hallucination and Creativity in LLMs' Decoding Layers
Computation and Language
Makes AI more creative without making up fake stuff.
A Concise Review of Hallucinations in LLMs and their Mitigation
Computation and Language
Stops computers from making up fake information.
HalluClean: A Unified Framework to Combat Hallucinations in LLMs
Computation and Language
Fixes computer writing to be truthful and correct.