SproutBench: A Benchmark for Safe and Ethical Large Language Models for Youth
By: Wenpeng Xing , Lanyi Wei , Haixiao Hu and more
Potential Business Impact:
Tests AI to keep kids safe online.
The rapid proliferation of large language models (LLMs) in applications targeting children and adolescents necessitates a fundamental reassessment of prevailing AI safety frameworks, which are largely tailored to adult users and neglect the distinct developmental vulnerabilities of minors. This paper highlights key deficiencies in existing LLM safety benchmarks, including their inadequate coverage of age-specific cognitive, emotional, and social risks spanning early childhood (ages 0--6), middle childhood (7--12), and adolescence (13--18). To bridge these gaps, we introduce SproutBench, an innovative evaluation suite comprising 1,283 developmentally grounded adversarial prompts designed to probe risks such as emotional dependency, privacy violations, and imitation of hazardous behaviors. Through rigorous empirical evaluation of 47 diverse LLMs, we uncover substantial safety vulnerabilities, corroborated by robust inter-dimensional correlations (e.g., between Safety and Risk Prevention) and a notable inverse relationship between Interactivity and Age Appropriateness. These insights yield practical guidelines for advancing child-centric AI design and deployment.
Similar Papers
Safe-Child-LLM: A Developmental Benchmark for Evaluating LLM Safety in Child-LLM Interactions
Computers and Society
Tests if AI is safe for kids and teens.
MinorBench: A hand-built benchmark for content-based risks for children
Computation and Language
Tests AI to keep kids safe from bad online stuff.
YouthSafe: A Youth-Centric Safety Benchmark and Safeguard Model for Large Language Models
Human-Computer Interaction
Protects kids from harmful AI chats.