Task Matrices: Linear Maps for Cross-Model Finetuning Transfer
By: Darrin O' Brien, Dhikshith Gajulapalli, Eric Xia
Potential Business Impact:
Makes AI learn new tasks faster and better.
Results in interpretability suggest that large vision and language models learn implicit linear encodings when models are biased by in-context prompting. However, the existence of similar linear representations in more general adaptation regimes has not yet been demonstrated. In this work, we develop the concept of a task matrix, a linear transformation from a base to finetuned embedding state. We demonstrate that for vision and text models and ten different datasets, a base model augmented with a task matrix achieves results surpassing linear probes, sometimes approaching finetuned levels. Our results validate the existence of cross-layer linear encodings between pretrained and finetuned architectures. Moreover, we show that a data-based approximation for such encodings is both efficient and generalizable to multiple domains. We make our implementation publicly available.
Similar Papers
Latent Traits and Cross-Task Transfer: Deconstructing Dataset Interactions in LLM Fine-tuning
Computation and Language
Helps computers learn new tasks better.
Comparing Task-Agnostic Embedding Models for Tabular Data
Machine Learning (CS)
Makes computers understand data much faster.
Provable Benefits of Task-Specific Prompts for In-context Learning
Computation and Language
Teaches AI to learn different tasks faster.