STAGE: A Benchmark for Knowledge Graph Construction, Question Answering, and In-Script Role-Playing over Movie Screenplays
By: Qiuyu Tian , Yiding Li , Fengyi Chen and more
Movie screenplays are rich long-form narratives that interleave complex character relationships, temporally ordered events, and dialogue-driven interactions. While prior benchmarks target individual subtasks such as question answering or dialogue generation, they rarely evaluate whether models can construct a coherent story world and use it consistently across multiple forms of reasoning and generation. We introduce STAGE (Screenplay Text, Agents, Graphs and Evaluation), a unified benchmark for narrative understanding over full-length movie screenplays. STAGE defines four tasks: knowledge graph construction, scene-level event summarization, long-context screenplay question answering, and in-script character role-playing, all grounded in a shared narrative world representation. The benchmark provides cleaned scripts, curated knowledge graphs, and event- and character-centric annotations for 150 films across English and Chinese, enabling holistic evaluation of models' abilities to build world representations, abstract and verify narrative events, reason over long narratives, and generate character-consistent responses.
Similar Papers
STAGE: Storyboard-Anchored Generation for Cinematic Multi-shot Narrative
CV and Pattern Recognition
Makes videos tell stories with consistent characters.
StageDesigner: Artistic Stage Generation for Scenography via Theater Scripts
CV and Pattern Recognition
Creates amazing 3D stage designs from scripts.
Long Story Generation via Knowledge Graph and Literary Theory
Computation and Language
Writes longer, more interesting stories that don't get boring.