TaleDiffusion: Multi-Character Story Generation with Dialogue Rendering
By: Ayan Banerjee , Josep Lladós , Umapada Pal and more
Potential Business Impact:
Makes stories show characters acting right.
Text-to-story visualization is challenging due to the need for consistent interaction among multiple characters across frames. Existing methods struggle with character consistency, leading to artifact generation and inaccurate dialogue rendering, which results in disjointed storytelling. In response, we introduce TaleDiffusion, a novel framework for generating multi-character stories with an iterative process, maintaining character consistency, and accurate dialogue assignment via postprocessing. Given a story, we use a pre-trained LLM to generate per-frame descriptions, character details, and dialogues via in-context learning, followed by a bounded attention-based per-box mask technique to control character interactions and minimize artifacts. We then apply an identity-consistent self-attention mechanism to ensure character consistency across frames and region-aware cross-attention for precise object placement. Dialogues are also rendered as bubbles and assigned to characters via CLIPSeg. Experimental results demonstrate that TaleDiffusion outperforms existing methods in consistency, noise reduction, and dialogue rendering.
Similar Papers
Playmate2: Training-Free Multi-Character Audio-Driven Animation via Diffusion Transformer with Reward Feedback
CV and Pattern Recognition
Makes videos of people talking from sound.
Lay2Story: Extending Diffusion Transformers for Layout-Togglable Story Generation
CV and Pattern Recognition
Makes cartoon characters stay the same in stories.
TalkingPose: Efficient Face and Gesture Animation with Feedback-guided Diffusion Model
CV and Pattern Recognition
Creates long, smooth talking animations from pictures.