FroM: Frobenius Norm-Based Data-Free Adaptive Model Merging
By: Zijian Li , Xiaocheng Feng , Huixin Liu and more
Potential Business Impact:
Combines AI knowledge without messing up tasks.
With the development of large language models, fine-tuning has emerged as an effective method to enhance performance in specific scenarios by injecting domain-specific knowledge. In this context, model merging techniques provide a solution for fusing knowledge from multiple fine-tuning models by combining their parameters. However, traditional methods often encounter task interference when merging full fine-tuning models, and this problem becomes even more evident in parameter-efficient fine-tuning scenarios. In this paper, we introduce an improvement to the RegMean method, which indirectly leverages the training data to approximate the outputs of the linear layers before and after merging. We propose an adaptive merging method called FroM, which directly measures the model parameters using the Frobenius norm, without any training data. By introducing an additional hyperparameter for control, FroM outperforms baseline methods across various fine-tuning scenarios, alleviating the task interference problem.
Similar Papers
Dynamic Fisher-weighted Model Merging via Bayesian Optimization
Computation and Language
Combines AI models to do many jobs better.
FW-Merging: Scaling Model Merging with Frank-Wolfe Optimization
Machine Learning (CS)
Combines AI models to do many jobs better.
Mitigating Parameter Interference in Model Merging via Sharpness-Aware Fine-Tuning
Machine Learning (CS)
Combines AI models without losing their skills.