Efficient Multi-Source Knowledge Transfer by Model Merging
By: Marcin Osial , Bartosz Wójcik , Bartosz Zieliński and more
Potential Business Impact:
Learns faster by combining knowledge from many AI models.
While transfer learning is an advantageous strategy, it overlooks the opportunity to leverage knowledge from numerous available models online. Addressing this multi-source transfer learning problem is a promising path to boost adaptability and cut re-training costs. However, existing approaches are inherently coarse-grained, lacking the necessary precision for granular knowledge extraction and the aggregation efficiency required to fuse knowledge from either a large number of source models or those with high parameter counts. We address these limitations by leveraging Singular Value Decomposition (SVD) to first decompose each source model into its elementary, rank-one components. A subsequent aggregation stage then selects only the most salient components from all sources, thereby overcoming the previous efficiency and precision limitations. To best preserve and leverage the synthesized knowledge base, our method adapts to the target task by fine-tuning only the principal singular values of the merged matrix. In essence, this process only recalibrates the importance of top SVD components. The proposed framework allows for efficient transfer learning, is robust to perturbations both at the input level and in the parameter space (e.g., noisy or pruned sources), and scales well computationally.
Similar Papers
Stay Unique, Stay Efficient: Preserving Model Personality in Multi-Task Merging
Machine Learning (CS)
Combines AI skills without losing what it learned.
Model Recycling Framework for Multi-Source Data-Free Supervised Transfer Learning
Machine Learning (CS)
Reuses old computer brains to learn new things.
Towards Reversible Model Merging For Low-rank Weights
Machine Learning (CS)
Combines AI models without losing their skills.