Align, Don't Divide: Revisiting the LoRA Architecture in Multi-Task Learning
By: Jinda Liu , Bo Cheng , Yi Chang and more
Potential Business Impact:
Makes AI learn many jobs with less effort.
Parameter-Efficient Fine-Tuning (PEFT) is essential for adapting Large Language Models (LLMs). In practice, LLMs are often required to handle a diverse set of tasks from multiple domains, a scenario naturally addressed by multi-task learning (MTL). Within this MTL context, a prevailing trend involves LoRA variants with multiple adapters or heads, which advocate for structural diversity to capture task-specific knowledge. Our findings present a direct challenge to this paradigm. We first show that a simplified multi-head architecture with high inter-head similarity substantially outperforms complex multi-adapter and multi-head systems. This leads us to question the multi-component paradigm itself, and we further demonstrate that a standard single-adapter LoRA, with a sufficiently increased rank, also achieves highly competitive performance. These results lead us to a new hypothesis: effective MTL generalization hinges on learning robust shared representations, not isolating task-specific features. To validate this, we propose Align-LoRA, which incorporates an explicit loss to align task representations within the shared adapter space. Experiments confirm that Align-LoRA significantly surpasses all baselines, establishing a simpler yet more effective paradigm for adapting LLMs to multiple tasks. The code is available at https://github.com/jinda-liu/Align-LoRA.
Similar Papers
MeTA-LoRA: Data-Efficient Multi-Task Fine-Tuning for Large Language Models
Computation and Language
Teaches AI more with less information.
LoRI: Reducing Cross-Task Interference in Multi-Task Low-Rank Adaptation
Machine Learning (CS)
Makes AI learn many things without forgetting.
LoRAFusion: Efficient LoRA Fine-Tuning for LLMs
Machine Learning (CS)
Makes AI learn faster and use less power.