The Effectiveness of Approximate Regularized Replay for Efficient Supervised Fine-Tuning of Large Language Models
By: Matthew Riemer , Erik Miehling , Miao Liu and more
Although parameter-efficient fine-tuning methods, such as LoRA, only modify a small subset of parameters, they can have a significant impact on the model. Our instruction-tuning experiments show that LoRA-based supervised fine-tuning can catastrophically degrade model capabilities, even when trained on very small datasets for relatively few steps. With that said, we demonstrate that while the most straightforward approach (that is likely the most used in practice) fails spectacularly, small tweaks to the training procedure with very little overhead can virtually eliminate the problem. Particularly, in this paper we consider a regularized approximate replay approach which penalizes KL divergence with respect to the initial model and interleaves in data for next token prediction from a different, yet similar, open access corpus to what was used in pre-training. When applied to Qwen instruction-tuned models, we find that this recipe preserves general knowledge in the model without hindering plasticity to new tasks by adding a modest amount of computational overhead.
Similar Papers
Mitigating Forgetting in Low Rank Adaptation
Machine Learning (CS)
Keeps old knowledge when learning new things.
Investigating ReLoRA: Effects on the Learning Dynamics of Small Language Models
Computation and Language
Makes small AI models learn better and faster.
LoRA-Loop: Closing the Synthetic Replay Cycle for Continual VLM Learning
CV and Pattern Recognition
Teaches AI new things without forgetting old ones.