Score: 0

Efficient Multi-Source Knowledge Transfer by Model Merging

Published: August 26, 2025 | arXiv ID: 2508.19353v1

By: Marcin Osial , Bartosz Wójcik , Bartosz Zieliński and more

Potential Business Impact:

Learns faster by combining knowledge from many AI models.

Business Areas:
Machine Learning Artificial Intelligence, Data and Analytics, Software

While transfer learning is an advantageous strategy, it overlooks the opportunity to leverage knowledge from numerous available models online. Addressing this multi-source transfer learning problem is a promising path to boost adaptability and cut re-training costs. However, existing approaches are inherently coarse-grained, lacking the necessary precision for granular knowledge extraction and the aggregation efficiency required to fuse knowledge from either a large number of source models or those with high parameter counts. We address these limitations by leveraging Singular Value Decomposition (SVD) to first decompose each source model into its elementary, rank-one components. A subsequent aggregation stage then selects only the most salient components from all sources, thereby overcoming the previous efficiency and precision limitations. To best preserve and leverage the synthesized knowledge base, our method adapts to the target task by fine-tuning only the principal singular values of the merged matrix. In essence, this process only recalibrates the importance of top SVD components. The proposed framework allows for efficient transfer learning, is robust to perturbations both at the input level and in the parameter space (e.g., noisy or pruned sources), and scales well computationally.

Page Count
19 pages

Category
Computer Science:
Machine Learning (CS)