Scaling Towards the Information Boundary of Instruction Set: InfinityInstruct-Subject Technical Report
By: Li Du , Hanyu Zhao , Yiming Ju and more
Potential Business Impact:
Teaches computers to follow harder instructions better.
Instruction tuning has become a foundation for unlocking the capabilities of large-scale pretrained models and improving their performance on complex tasks. Thus, the construction of high-quality instruction datasets is crucial for enhancing model performance and generalizability. Although current instruction datasets have reached tens of millions of samples, models finetuned on them may still struggle with complex instruction following and tasks in rare domains. This is primarily due to limited expansion in both ``coverage'' (coverage of task types and knowledge areas) and ``depth'' (instruction complexity) of the instruction set. To address this issue, we propose a systematic instruction data construction framework, which integrates a hierarchical labeling system, an informative seed selection algorithm, an evolutionary data synthesis process, and a model deficiency diagnosis with targeted data generation. These components form an iterative closed-loop to continuously enhance the coverage and depth of instruction data. Based on this framework, we construct InfinityInstruct-Subject, a high-quality dataset containing ~1.5 million instructions. Experiments on multiple foundation models and benchmark tasks demonstrate its effectiveness in improving instruction-following capabilities. Further analyses suggest that InfinityInstruct-Subject shows enlarged coverage and depth compared to comparable synthesized instruction datasets. Our work lays a theoretical and practical foundation for the efficient, continuous evolution of instruction datasets, moving from data quantity expansion to qualitative improvement.
Similar Papers
Infinity Instruct: Scaling Instruction Selection and Synthesis to Enhance Language Models
Computation and Language
Makes free AI understand and talk better.
Infinite-Instruct: Synthesizing Scaling Code instruction Data with Bidirectional Synthesis and Static Verification
Computation and Language
Teaches computers to write better code.
Accelerate Scaling of LLM Alignment via Quantifying the Coverage and Depth of Instruction Set
Artificial Intelligence
Makes AI smarter and learn faster.