How Much is Too Much? Exploring LoRA Rank Trade-offs for Retaining Knowledge and Domain Robustness
By: Darshita Rathore , Vineet Kumar , Chetna Bansal and more
Potential Business Impact:
Makes AI smarter for questions, even new ones.
Large language models are increasingly adapted to downstream tasks through fine-tuning. Full supervised fine-tuning (SFT) and parameter-efficient fine-tuning (PEFT) methods, such as Low-Rank Adaptation (LoRA), are two dominant approaches. While PEFT methods are widely used for their computational efficiency, the implications of their configurations (e.g., rank) remain under-explored in downstream Q&A tasks and generalisation. In this work, we perform a comprehensive evaluation across multiple reasoning and recall datasets, conducting a rank sweep to quantify the trade-off between SFT and PEFT. We also compare the accuracy of PEFT and SFT models across in-domain and out-of-domain adaptation, highlighting distinct generalisation behaviour and task-specific forgetting. We demonstrate that LoRA achieves competitive and in some cases superior performance compared to SFT, particularly on reasoning tasks at specific rank values. Additionally, we analyze the internal representations via spectral features and layer-wise attention structures, offering insights into representational drift and structural changes in attention patterns.
Similar Papers
Towards Higher Effective Rank in Parameter-efficient Fine-tuning using Khatri--Rao Product
Machine Learning (CS)
Makes AI learn better without needing more power.
Sensitivity-LoRA: Low-Load Sensitivity-Based Fine-Tuning for Large Language Models
Machine Learning (CS)
Makes smart computer programs learn faster and better.
Data Efficient Adaptation in Large Language Models via Continuous Low-Rank Fine-Tuning
Artificial Intelligence
Teaches AI new things without forgetting old.