Memory-Efficient Split Federated Learning for LLM Fine-Tuning on Heterogeneous Mobile Devices
By: Xiaopei Chen , Liang Li , Fei Ji and more
Potential Business Impact:
Makes phones train big AI models faster.
In this paper, we propose an edge-assisted split federated learning framework to facilitate large language model (LLM) fine-tuning on heterogeneous mobile devices while alleviating memory pressures on both mobile devices and the edge server. Specifically, mobile devices perform low-rank adaptation (LoRA) fine-tuning on only a subset of lower layers of the pre-trained LLM, tailored to their individual capacities. On the server, a full LLM is maintained, and the corresponding LoRA modules are selectively fine-tuned in a sequential manner for each device. To further enhance training efficiency, we propose a server-side training scheduling method that optimizes the processing order of devices for accelerating fine-tuning. Extensive experiments demonstrate that compared to the baselines, our scheme can reduce 79\% memory footprint and 6\% training time while achieving comparable performance.
Similar Papers
Efficient Split Federated Learning for Large Language Models over Communication Networks
Machine Learning (CS)
Makes smart computer programs train faster on phones.
Fed MobiLLM: Efficient Federated LLM Fine-Tuning over Heterogeneous Mobile Devices via Server Assisted Side-Tuning
Machine Learning (CS)
Makes phones smarter without slowing them down.
HLoRA: Efficient Federated Learning System for LLM Heterogeneous Fine-Tuning
Distributed, Parallel, and Cluster Computing
Teaches AI new things without seeing private data.