IdentityStory: Taming Your Identity-Preserving Generator for Human-Centric Story Generation
By: Donghao Zhou , Jingyu Lin , Guibao Shen and more
Potential Business Impact:
Creates stories with the same people in every picture.
Recent visual generative models enable story generation with consistent characters from text, but human-centric story generation faces additional challenges, such as maintaining detailed and diverse human face consistency and coordinating multiple characters across different images. This paper presents IdentityStory, a framework for human-centric story generation that ensures consistent character identity across multiple sequential images. By taming identity-preserving generators, the framework features two key components: Iterative Identity Discovery, which extracts cohesive character identities, and Re-denoising Identity Injection, which re-denoises images to inject identities while preserving desired context. Experiments on the ConsiStory-Human benchmark demonstrate that IdentityStory outperforms existing methods, particularly in face consistency, and supports multi-character combinations. The framework also shows strong potential for applications such as infinite-length story generation and dynamic character composition.
Similar Papers
Infinite-Story: A Training-Free Consistent Text-to-Image Generation
CV and Pattern Recognition
Creates matching pictures for stories, super fast.
Reverse Personalization
CV and Pattern Recognition
Changes faces in pictures without text.
2K-Characters-10K-Stories: A Quality-Gated Stylized Narrative Dataset with Disentangled Control and Sequence Consistency
CV and Pattern Recognition
Creates stories with the same characters every time.