From Symptoms to Systems: An Expert-Guided Approach to Understanding Risks of Generative AI for Eating Disorders
By: Amy Winecoff, Kevin Klyman
Potential Business Impact:
AI can accidentally help people with eating disorders.
Generative AI systems may pose serious risks to individuals vulnerable to eating disorders. Existing safeguards tend to overlook subtle but clinically significant cues, leaving many risks unaddressed. To better understand the nature of these risks, we conducted semi-structured interviews with 15 clinicians, researchers, and advocates with expertise in eating disorders. Using abductive qualitative analysis, we developed an expert-guided taxonomy of generative AI risks across seven categories: (1) providing generalized health advice; (2) encouraging disordered behaviors; (3) supporting symptom concealment; (4) creating thinspiration; (5) reinforcing negative self-beliefs; (6) promoting excessive focus on the body; and (7) perpetuating narrow views about eating disorders. Our results demonstrate how certain user interactions with generative AI systems intersect with clinical features of eating disorders in ways that may intensify risk. We discuss implications of our work, including approaches for risk assessment, safeguard design, and participatory evaluation practices with domain experts.
Similar Papers
Simulating Psychological Risks in Human-AI Interactions: Real-Case Informed Modeling of AI-Induced Addiction, Anorexia, Depression, Homicide, Psychosis, and Suicide
Human-Computer Interaction
Finds ways AI can hurt people's minds.
Multimodal Generative AI and Foundation Models for Behavioural Health in Online Gambling
Computers and Society
AI helps stop online gambling addiction.
Beyond the Benefits: A Systematic Review of the Harms and Consequences of Generative AI in Computing Education
Human-Computer Interaction
Finds dangers of AI in school learning.