Activation Steering for Masked Diffusion Language Models
By: Adi Shnaidman , Erin Feiglin , Osher Yaari and more
Masked diffusion language models (MDLMs) generate text through an iterative denoising process. They have recently gained attention due to mask-parallel decoding and competitive performance with autoregressive large language models. However, effective mechanisms for inference-time control and steering in MDLMs remain largely unexplored. We present an activation-steering framework for MDLMs that computes layer-wise steering vectors from a single forward pass using contrastive examples, without simulating the denoising trajectory. These directions are applied at every reverse-diffusion step, yielding an efficient inference-time control mechanism. Experiments on LLaDA-8B-Instruct demonstrate reliable modulation of high-level attributes, with ablations examining the effects of steering across transformer sub-modules and token scope (prompt vs.\ response).
Similar Papers
Diffusion Language Model Inference with Monte Carlo Tree Search
Computation and Language
Makes AI write better by finding best word choices.
Simple Denoising Diffusion Language Models
Machine Learning (CS)
Makes computers write better stories and sentences.
Improving Text Style Transfer using Masked Diffusion Language Models with Inference-time Scaling
Computation and Language
Makes computers write better stories and sentences.