MeTA-LoRA: Data-Efficient Multi-Task Fine-Tuning for Large Language Models
By: Bo Cheng , Xu Wang , Jinda Liu and more
Potential Business Impact:
Teaches AI more with less information.
Low-Rank Adaptation (LoRA) has emerged as one of the most widely used parameter-efficient fine-tuning (PEFT) methods for adapting large language models (LLMs) to downstream tasks. While highly effective in single-task settings, it struggles to efficiently leverage inter-task knowledge in complex multi-task learning scenarios, often requiring substantial task-specific data to achieve optimal performance. To address this limitation, we introduce MeTA-LoRA, a two-stage optimization framework that significantly improves data efficiency in multi-task adaptation. In the first stage, task-specific LoRA adapters are learned using only a few samples from each involved dataset, enabling rapid adaptation without large-scale supervision. In the second stage, the shared LoRA adapter is updated by aggregating gradients from multiple tasks to promote knowledge transfer across tasks, further reducing data usage by leveraging common patterns. In both multi-task learning and multilingual learning scenarios, our method matches or surpasses the performance of traditional full-data LoRA fine-tuning approaches, while using significantly less task-specific data.
Similar Papers
MetaLoRA: Tensor-Enhanced Adaptive Low-Rank Fine-tuning
Machine Learning (CS)
Teaches computers to learn new things faster.
Align, Don't Divide: Revisiting the LoRA Architecture in Multi-Task Learning
Computation and Language
Makes AI learn many jobs with less effort.
LoRAFusion: Efficient LoRA Fine-Tuning for LLMs
Machine Learning (CS)
Makes AI learn faster and use less power.