A Training-Free Style-Personalization via Scale-wise Autoregressive Model
By: Kyoungmin Lee , Jihun Park , Jongmin Gim and more
Potential Business Impact:
Makes computer art match your exact style.
We present a training-free framework for style-personalized image generation that controls content and style information during inference using a scale-wise autoregressive model. Our method employs a three-path design--content, style, and generation--each guided by a corresponding text prompt, enabling flexible and efficient control over image semantics without any additional training. A central contribution of this work is a step-wise and attention-wise intervention analysis. Through systematic prompt and feature injection, we find that early-to-middle generation steps play a pivotal role in shaping both content and style, and that query features predominantly encode content-specific information. Guided by these insights, we introduce two targeted mechanisms: Key Stage Attention Sharing, which aligns content and style during the semantically critical steps, and Adaptive Query Sharing, which reinforces content semantics in later steps through similarity-aware query blending. Extensive experiments demonstrate that our method achieves competitive style fidelity and prompt fidelity compared to fine-tuned baselines, while offering faster inference and greater deployment flexibility.
Similar Papers
A Training-Free Style-aligned Image Generation with Scale-wise Autoregressive Model
CV and Pattern Recognition
Makes AI pictures match the style you want.
StyleAR: Customizing Multimodal Autoregressive Model for Style-Aligned Text-to-Image Generation
CV and Pattern Recognition
Makes pictures match a chosen style
Personalized Text-to-Image Generation with Auto-Regressive Models
CV and Pattern Recognition
Makes AI draw pictures of *your* stuff.