Score: 1

MemFlow: Flowing Adaptive Memory for Consistent and Efficient Long Video Narratives

Published: December 16, 2025 | arXiv ID: 2512.14699v1

By: Sihui Ji , Xi Chen , Shuai Yang and more

Potential Business Impact:

Makes videos stay the same story over time.

Business Areas:
Motion Capture Media and Entertainment, Video

The core challenge for streaming video generation is maintaining the content consistency in long context, which poses high requirement for the memory design. Most existing solutions maintain the memory by compressing historical frames with predefined strategies. However, different to-generate video chunks should refer to different historical cues, which is hard to satisfy with fixed strategies. In this work, we propose MemFlow to address this problem. Specifically, before generating the coming chunk, we dynamically update the memory bank by retrieving the most relevant historical frames with the text prompt of this chunk. This design enables narrative coherence even if new event happens or scenario switches in future frames. In addition, during generation, we only activate the most relevant tokens in the memory bank for each query in the attention layers, which effectively guarantees the generation efficiency. In this way, MemFlow achieves outstanding long-context consistency with negligible computation burden (7.9% speed reduction compared with the memory-free baseline) and keeps the compatibility with any streaming video generation model with KV cache.

Repos / Data Links

Page Count
10 pages

Category
Computer Science:
CV and Pattern Recognition