Latent Traits and Cross-Task Transfer: Deconstructing Dataset Interactions in LLM Fine-tuning
By: Shambhavi Krishna , Atharva Naik , Chaitali Agarwal and more
Potential Business Impact:
Helps computers learn new tasks better.
Large language models are increasingly deployed across diverse applications. This often includes tasks LLMs have not encountered during training. This implies that enumerating and obtaining the high-quality training data for all tasks is infeasible. Thus, we often need to rely on transfer learning using datasets with different characteristics, and anticipate out-of-distribution requests. Motivated by this practical need, we propose an analysis framework, building a transfer learning matrix and dimensionality reduction, to dissect these cross-task interactions. We train and analyze 10 models to identify latent abilities (e.g., Reasoning, Sentiment Classification, NLU, Arithmetic) and discover the side effects of the transfer learning. Our findings reveal that performance improvements often defy explanations based on surface-level dataset similarity or source data quality. Instead, hidden statistical factors of the source dataset, such as class distribution and generation length proclivities, alongside specific linguistic features, are actually more influential. This work offers insights into the complex dynamics of transfer learning, paving the way for more predictable and effective LLM adaptation.
Similar Papers
Donors and Recipients: On Asymmetric Transfer Across Tasks and Languages with Parameter-Efficient Fine-Tuning
Computation and Language
Makes AI better at many jobs at once.
Beyond Neural Incompatibility: Easing Cross-Scale Knowledge Transfer in Large Language Models through Latent Semantic Alignment
Computation and Language
Lets smaller AI learn from bigger AI.
Beyond the Rosetta Stone: Unification Forces in Generalization Dynamics
Computation and Language
Helps computers use knowledge across different languages.