DIP: Dynamic In-Context Planner For Diffusion Language Models
By: Yang Li , Han Meng , Chenan Wang and more
Potential Business Impact:
Makes AI understand better, faster, with less work.
Diffusion language models (DLMs) have shown strong potential for general natural language tasks with in-context examples. However, due to the bidirectional attention mechanism, DLMs incur substantial computational cost as context length increases. This work addresses this issue with a key discovery: unlike the sequential generation in autoregressive language models (ARLMs), the diffusion generation paradigm in DLMs allows \textit{efficient dynamic adjustment of the context} during generation. Building on this insight, we propose \textbf{D}ynamic \textbf{I}n-Context \textbf{P}lanner (DIP), a context-optimization method that dynamically selects and inserts in-context examples during generation, rather than providing all examples in the prompt upfront. Results show DIP maintains generation quality while achieving up to 12.9$\times$ inference speedup over standard inference and 1.17$\times$ over KV cache-enhanced inference.
Similar Papers
In-Context Iterative Policy Improvement for Dynamic Manipulation
Robotics
Teaches robots to learn new tasks quickly.
Context-Aware Initialization for Reducing Generative Path Length in Diffusion Language Models
Computation and Language
Makes AI write faster by starting with good guesses.
Planner and Executor: Collaboration between Discrete Diffusion And Autoregressive Models in Reasoning
Computation and Language
Makes AI smarter and faster by combining two types.