Sparsity Outperforms Low-Rank Projections in Few-Shot Adaptation
By: Nairouz Mrabah , Nicolas Richet , Ismail Ben Ayed and more
Potential Business Impact:
Teaches computers new things with very little data.
Adapting Vision-Language Models (VLMs) to new domains with few labeled samples remains a significant challenge due to severe overfitting and computational constraints. State-of-the-art solutions, such as low-rank reparameterization, mitigate these issues but often struggle with generalization and require extensive hyperparameter tuning. In this paper, a novel Sparse Optimization (SO) framework is proposed. Unlike low-rank approaches that typically constrain updates to a fixed subspace, our SO method leverages high sparsity to dynamically adjust very few parameters. We introduce two key paradigms. First, we advocate for \textit{local sparsity and global density}, which updates a minimal subset of parameters per iteration while maintaining overall model expressiveness. As a second paradigm, we advocate for \textit{local randomness and global importance}, which sparsifies the gradient using random selection while pruning the first moment based on importance. This combination significantly mitigates overfitting and ensures stable adaptation in low-data regimes. Extensive experiments on 11 diverse datasets show that SO achieves state-of-the-art few-shot adaptation performance while reducing memory overhead.
Similar Papers
Robust Few-Shot Vision-Language Model Adaptation
CV and Pattern Recognition
Helps AI understand images it hasn't seen.
Few-Shot Adaptation Benchmark for Remote Sensing Vision-Language Models
CV and Pattern Recognition
Helps AI learn from few pictures of Earth.
Sparse Optimization for Transfer Learning: A L0-Regularized Framework for Multi-Source Domain Adaptation
Machine Learning (Stat)
Makes computer learning faster and more accurate.