Towards Long-window Anchoring in Vision-Language Model Distillation
By: Haoyi Zhou , Shuo Li , Tianyu Chen and more
Potential Business Impact:
Makes AI understand much longer text and pictures.
While large vision-language models (VLMs) demonstrate strong long-context understanding, their prevalent small branches fail on linguistics-photography alignment for a limited window size. We discover that knowledge distillation improves students' capability as a complement to Rotary Position Embeddings (RoPE) on window sizes (anchored from large models). Building on this insight, we propose LAid, which directly aims at the transfer of long-range attention mechanisms through two complementary components: (1) a progressive distance-weighted attention matching that dynamically emphasizes longer position differences during training, and (2) a learnable RoPE response gain modulation that selectively amplifies position sensitivity where needed. Extensive experiments across multiple model families demonstrate that LAid-distilled models achieve up to 3.2 times longer effective context windows compared to baseline small models, while maintaining or improving performance on standard VL benchmarks. Spectral analysis also suggests that LAid successfully preserves crucial low-frequency attention components that conventional methods fail to transfer. Our work not only provides practical techniques for building more efficient long-context VLMs but also offers theoretical insights into how positional understanding emerges and transfers during distillation.
Similar Papers
Enhancing Medical Large Vision-Language Models via Alignment Distillation
CV and Pattern Recognition
Makes AI doctors understand medical images better.
Online In-Context Distillation for Low-Resource Vision Language Models
CV and Pattern Recognition
Makes small AI understand pictures better with less cost.
Vi-LAD: Vision-Language Attention Distillation for Socially-Aware Robot Navigation in Dynamic Environments
Robotics
Teaches robots to navigate crowded places safely.