Ming-UniVision: Joint Image Understanding and Generation with a Unified Continuous Tokenizer
By: Ziyuan Huang , DanDan Zheng , Cheng Zou and more
Potential Business Impact:
Makes computers understand and create pictures better.
Visual tokenization remains a core challenge in unifying visual understanding and generation within the autoregressive paradigm. Existing methods typically employ tokenizers in discrete latent spaces to align with the tokens from large language models, where the quantization errors can limit semantic expressiveness and degrade the capability of vision-language understanding. To address this, we introduce MingTok, a new family of visual tokenizers with a continuous latent space, for unified autoregressive generation and understanding. While understanding tasks favor discriminative high-dimensional features, generation tasks prefer compact low-level codes. Thus, to reconcile these competing demands, MingTok adopts a three-stage sequential architecture involving low-level encoding, semantic expansion, and visual reconstruction. Built on top of it, Ming-UniVision eliminates the need for task-specific visual representations, and unifies diverse vision-language tasks under a single autoregrsssive prediction paradigm. By formulating both understanding and generation as next-token prediction in a shared continuous space, it seamlessly supports multi-round, in-context tasks such as iterative understanding, generation and editing. Empirically, we find that using a unified continuous visual representation reconciles the competing requirements on the tokenizers by the understanding and generation tasks, thereby leading to state-of-the-art level performance across both domains. We hope our findings will facilitate unified visual tokenization in the continuous domain. Inference code and model weights are released to benefit community.
Similar Papers
UniToken: Harmonizing Multimodal Understanding and Generation through Unified Visual Encoding
CV and Pattern Recognition
Lets computers understand and create pictures.
Ming-UniAudio: Speech LLM for Joint Understanding, Generation and Editing with Unified Representation
Computation and Language
Lets you edit spoken words with just your voice.
Ming-Lite-Uni: Advancements in Unified Architecture for Natural Multimodal Interaction
CV and Pattern Recognition
Makes computers create and change pictures from words.