Characterization of Transfer Using Multi-task Learning Curves
By: András Millinghoffer, Bence Bolgár, Péter Antal
Transfer effects manifest themselves both during training using a fixed data set and in inductive inference using accumulating data. We hypothesize that perturbing the data set by including more samples, instead of perturbing the model by gradient updates, provides a complementary and more fundamental characterization of transfer effects. To capture this phenomenon, we quantitatively model transfer effects using multi-task learning curves approximating the inductive performance over varying sample sizes. We describe an efficient method to approximate multi-task learning curves analogous to the Task Affinity Grouping method applied during training. We compare the statistical and computational approaches to transfer, which indicates considerably higher compute costs for the previous but better power and broader applicability. Evaluations are performed using a benchmark drug-target interaction data set. Our results show that learning curves can better capture the effects of multi-task learning and their multi-task extensions can delineate pairwise and contextual transfer effects in foundation models.
Similar Papers
Latent Traits and Cross-Task Transfer: Deconstructing Dataset Interactions in LLM Fine-tuning
Computation and Language
Helps computers learn new tasks better.
Advantages and limitations in the use of transfer learning for individual treatment effects in causal machine learning
Machine Learning (Stat)
Helps predict effects in small groups using big group data.
Cross-Learning from Scarce Data via Multi-Task Constrained Optimization
Machine Learning (CS)
Learns better from less information by sharing knowledge.