Score: 1

Distillation Dynamics: Towards Understanding Feature-Based Distillation in Vision Transformers

Published: November 10, 2025 | arXiv ID: 2511.06848v1

By: Huiyuan Tian, Bonan Xu Shijian Li

Potential Business Impact:

Makes AI models smaller and faster to train.

Business Areas:
Image Recognition Data and Analytics, Software

While feature-based knowledge distillation has proven highly effective for compressing CNNs, these techniques unexpectedly fail when applied to Vision Transformers (ViTs), often performing worse than simple logit-based distillation. We provide the first comprehensive analysis of this phenomenon through a novel analytical framework termed as ``distillation dynamics", combining frequency spectrum analysis, information entropy metrics, and activation magnitude tracking. Our investigation reveals that ViTs exhibit a distinctive U-shaped information processing pattern: initial compression followed by expansion. We identify the root cause of negative transfer in feature distillation: a fundamental representational paradigm mismatch between teacher and student models. Through frequency-domain analysis, we show that teacher models employ distributed, high-dimensional encoding strategies in later layers that smaller student models cannot replicate due to limited channel capacity. This mismatch causes late-layer feature alignment to actively harm student performance. Our findings reveal that successful knowledge transfer in ViTs requires moving beyond naive feature mimicry to methods that respect these fundamental representational constraints, providing essential theoretical guidance for designing effective ViTs compression strategies. All source code and experimental logs are provided in the supplementary material.

Country of Origin
🇨🇳 🇭🇰 China, Hong Kong

Page Count
13 pages

Category
Computer Science:
CV and Pattern Recognition