Do Code Models Suffer from the Dunning-Kruger Effect?
By: Mukul Singh , Somya Chatterjee , Arjun Radhakrishna and more
Potential Business Impact:
AI thinks it's better at coding than it is.
As artificial intelligence systems increasingly collaborate with humans in creative and technical domains, questions arise about the cognitive boundaries and biases that shape our shared agency. This paper investigates the Dunning-Kruger Effect (DKE), the tendency for those with limited competence to overestimate their abilities in state-of-the-art LLMs in coding tasks. By analyzing model confidence and performance across a diverse set of programming languages, we reveal that AI models mirror human patterns of overconfidence, especially in unfamiliar or low-resource domains. Our experiments demonstrate that less competent models and those operating in rare programming languages exhibit stronger DKE-like bias, suggesting that the strength of the bias is proportionate to the competence of the models.
Similar Papers
Is General-Purpose AI Reasoning Sensitive to Data-Induced Cognitive Biases? Dynamic Benchmarking on Typical Software Engineering Dilemmas
Human-Computer Interaction
AI can make mistakes like people.
Humans incorrectly reject confident accusatory AI judgments
Human-Computer Interaction
AI judges lies better than people, but we don't trust it.
Transitive Expert Error and Routing Problems in Complex AI Systems
Artificial Intelligence
Experts make mistakes at edges of knowledge.