EcoLoRA: Communication-Efficient Federated Fine-Tuning of Large Language Models
By: Han Liu , Ruoyao Wen , Srijith Nair and more
Potential Business Impact:
Makes AI learn faster with less data sent.
To address data locality and privacy restrictions, Federated Learning (FL) has recently been adopted to fine-tune large language models (LLMs), enabling improved performance on various downstream tasks without requiring aggregated data. However, the repeated exchange of model updates in FL can result in prohibitively high communication costs, hindering the distributed learning process. To address this challenge, we propose EcoLoRA, a novel communication-efficient federated fine-tuning framework for LLMs. Leveraging the modular structure, we propose a round-robin segment sharing scheme, where each client uploads only a complementary LoRA segment per round to reduce network bandwidth. It is further combined with adaptive sparsification methods tailored to LoRA's training dynamics and lossless encoding techniques. We conduct extensive evaluations on both question-answering and value-alignment tasks across multiple datasets and models. The results show that EcoLoRA significantly reduces communication overhead without compromising performance. For instance, it reduces communication time by up to 79% and total training time by up to 65%.
Similar Papers
Adaptive LoRA Experts Allocation and Selection for Federated Fine-Tuning
Machine Learning (CS)
Helps AI learn from private data without sharing.
HLoRA: Efficient Federated Learning System for LLM Heterogeneous Fine-Tuning
Distributed, Parallel, and Cluster Computing
Teaches AI new things without seeing private data.
Communication-Efficient Wireless Federated Fine-Tuning for Large-Scale AI Models
Machine Learning (CS)
Trains big computer brains with less data sent.