DSFT: Inspiring Diffusion Large Language Models to Comprehend Mathematical and Logical Patterns
By: Ranfei Chen, Ming Chen
Potential Business Impact:
Teaches computers math and logic better.
Diffusion large language models (dLLMs) have emerged as a new architecture following auto regressive models. Their denoising process offers a powerful generative advantage, but they present significant challenges in learning and understanding numerically sensitive mathematical and order-sensitive logical tasks. Current training methods, including pre-training, fine-tuning, and reinforcement learning, focus primarily on improving general knowledge retention and reasoning abilities, but lack a comprehensive understanding of mathematical and logical patterns. We propose DSFT, a simple yet effective Diffusion SFT strategy, by adjusting the masking strategy and loss function, guiding models to understand mathematical and logical patterns. This strategy can be flexibly combined with pre-training, reinforcement learning, and other training methods. Validated on models such as LLaDA and Dream series, we prove that DSFT on small-scale data can achieve improvements of 5-10% and approximately 2% on mathematical and logical problems, respectively. This inspiring masking approach offers insights for future learning of specific patterns, which can be easily and efficiently combined with other training methods and applied to various dLLMs. Our code is publicly available at https://anonymous.4open.science/r/DSFT-0FFB/
Similar Papers
d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning
Computation and Language
Helps computers solve math and planning problems.
WeFT: Weighted Entropy-driven Fine-Tuning for dLLMs
Computation and Language
Makes AI better at solving puzzles and math.
State Fourier Diffusion Language Model (SFDLM): A Scalable, Novel Iterative Approach to Language Modeling
Machine Learning (CS)
Creates text by fixing jumbled words, no big computers needed.