Elastic Mixture of Rank-Wise Experts for Knowledge Reuse in Federated Fine-Tuning
By: Yebo Wu , Jingguang Li , Zhijiang Guo and more
Potential Business Impact:
Reuses old AI knowledge to train new AI faster.
Federated fine-tuning offers a promising solution for adapting Large Language Models (LLMs) to downstream tasks while safeguarding data privacy. However, its high computational and communication demands hinder its deployment on resource-constrained devices. In this paper, we propose SmartFed, a resource-efficient federated fine-tuning framework. SmartFed intelligently reuses knowledge embedded in existing LoRA modules, eliminating the need for expensive training from scratch when adapting LLMs to new tasks. To effectively exploit this knowledge and ensure scalability, we introduce the Mixture of Rank-Wise Experts (MoRE). MoRE decomposes LoRA modules into fine-grained rank-level experts. These experts are selectively activated and combined based on input semantics and resource budgets. Moreover, to optimize resource utilization, we present the Elastic Expert Quota Allocation (EEQA). EEQA adaptively allocates expert capacity across parameter matrices based on their contribution to model performance, focusing computing resources on the critical experts. Extensive evaluations across multiple benchmarks demonstrate that SmartFed significantly outperforms existing methods in model performance and training efficiency.
Similar Papers
Adaptive LoRA Experts Allocation and Selection for Federated Fine-Tuning
Machine Learning (CS)
Helps AI learn from private data without sharing.
Federated Fine-Tuning of Sparsely-Activated Large Language Models on Resource-Constrained Devices
Distributed, Parallel, and Cluster Computing
Makes smart computer brains learn faster on weak computers.
FFT-MoE: Efficient Federated Fine-Tuning for Foundation Models via Large-scale Sparse MoE under Heterogeneous Edge
Machine Learning (CS)
Teaches AI to learn from many computers without sharing secrets.