Investigating the Effects of Cognitive Biases in Prompts on Large Language Model Outputs
By: Yan Sun, Stanley Kok
Potential Business Impact:
Fixes AI to stop believing fake information.
This paper investigates the influence of cognitive biases on Large Language Models (LLMs) outputs. Cognitive biases, such as confirmation and availability biases, can distort user inputs through prompts, potentially leading to unfaithful and misleading outputs from LLMs. Using a systematic framework, our study introduces various cognitive biases into prompts and assesses their impact on LLM accuracy across multiple benchmark datasets, including general and financial Q&A scenarios. The results demonstrate that even subtle biases can significantly alter LLM answer choices, highlighting a critical need for bias-aware prompt design and mitigation strategy. Additionally, our attention weight analysis highlights how these biases can alter the internal decision-making processes of LLMs, affecting the attention distribution in ways that are associated with output inaccuracies. This research has implications for Al developers and users in enhancing the robustness and reliability of Al applications in diverse domains.
Similar Papers
Cognitive Bias Detection Using Advanced Prompt Engineering
Computers and Society
Finds and fixes unfair thinking in writing.
Rethinking Prompt-based Debiasing in Large Language Models
Computation and Language
AI might not really understand fairness.
Cognitive Debiasing Large Language Models for Decision-Making
Computation and Language
Makes AI smarter by fixing its thinking errors.