Score: 0

Improved Supervised Fine-Tuning for Large Language Models to Mitigate Catastrophic Forgetting

Published: June 11, 2025 | arXiv ID: 2506.09428v2

By: Fei Ding, Baiqiao Wang

Potential Business Impact:

Keeps AI smart while teaching it new tricks.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Supervised Fine-Tuning (SFT) is a critical step for enhancing the instruction-following capabilities of Large Language Models (LLMs) and adapting them to specialized domains. However, SFT often leads to a degradation of the model's general abilities, a phenomenon known as catastrophic forgetting. This problem is exacerbated when third-party practitioners fine-tune open-source models, as the original SFT data is typically not available. To address this challenge, we propose a novel and cost-effective SFT method that effectively mitigates catastrophic forgetting without requiring access to the original SFT data. Our approach first reconstructs the likely instruction distribution of the base model. It then employs a multi-model generation and filtering pipeline to synthesize a high-quality general-purpose dataset. This synthetic dataset is mixed with new, domain-specific data for fine-tuning. Experimental results show that our method not only preserves the model's capabilities in general domains but also improves task-specific performance, outperforming baselines that use publicly available SFT datasets.

Country of Origin
🇨🇳 China

Page Count
8 pages

Category
Computer Science:
Computation and Language