Hallucination, reliability, and the role of generative AI in science
By: Charles Rathkopf
Potential Business Impact:
Fixes AI mistakes that trick scientists.
Generative AI is increasingly used in scientific domains, from protein folding to climate modeling. But these models produce distinctive errors known as hallucinations - outputs that are incorrect yet superficially plausible. Worse, some arguments suggest that hallucinations are an inevitable consequence of the mechanisms underlying generative inference. Fortunately, such arguments rely on a conception of hallucination defined solely with respect to internal properties of the model, rather than in reference to the empirical target system. This conception fails to distinguish epistemically benign errors from those that threaten scientific inference. I introduce the concept of corrosive hallucination to capture the epistemically troubling subclass: misrepresentations that are substantively misleading and resistant to systematic anticipation. I argue that although corrosive hallucinations do pose a threat to scientific reliability, they are not inevitable. Scientific workflows such as those surrounding AlphaFold and GenCast, both of which serve as case studies, can neutralize their effects by imposing theoretical constraints during training, and by strategically screening for errors at inference time. When embedded in such workflows, generative AI can reliably contribute to scientific knowledge.
Similar Papers
Engineering of Hallucination in Generative AI: It's not a Bug, it's a Feature
Computation and Language
Makes AI creative by letting it imagine things.
Hallucinating with AI: AI Psychosis as Distributed Delusions
Computers and Society
Helps us stop believing AI's fake stories.
HALoGEN: Fantastic LLM Hallucinations and Where to Find Them
Computation and Language
Tests computers to find when they make up facts.