The Inductive Bottleneck: Data-Driven Emergence of Representational Sparsity in Vision Transformers
By: Kanishk Awadhiya
Potential Business Impact:
Makes computers understand pictures better by focusing on important parts.
Vision Transformers (ViTs) lack the hierarchical inductive biases inherent to Convolutional Neural Networks (CNNs), theoretically allowing them to maintain high-dimensional representations throughout all layers. However, recent observations suggest ViTs often spontaneously manifest a "U-shaped" entropy profile-compressing information in middle layers before expanding it for the final classification. In this work, we demonstrate that this "Inductive Bottleneck" is not an architectural artifact, but a data-dependent adaptation. By analyzing the layer-wise Effective Encoding Dimension (EED) of DINO-trained ViTs across datasets of varying compositional complexity (UC Merced, Tiny ImageNet, and CIFAR-100), we show that the depth of the bottleneck correlates strongly with the semantic abstraction required by the task. We find that while texture-heavy datasets preserve high-rank representations throughout, object-centric datasets drive the network to dampen high-frequency information in middle layers, effectively "learning" a bottleneck to isolate semantic features.
Similar Papers
Distillation Dynamics: Towards Understanding Feature-Based Distillation in Vision Transformers
CV and Pattern Recognition
Makes AI models smaller and faster to train.
From Low-Rank Features to Encoding Mismatch: Rethinking Feature Distillation in Vision Transformers
CV and Pattern Recognition
Helps AI learn better from images.
Mechanisms of Non-Monotonic Scaling in Vision Transformers
Machine Learning (CS)
Makes computer "eyes" learn better by changing how they see.