Delay-Aware Diffusion Policy: Bridging the Observation-Execution Gap in Dynamic Tasks
By: Aileen Liao , Dong-Ki Kim , Max Olan Smith and more
Potential Business Impact:
Makes robots react faster to changing worlds.
As a robot senses and selects actions, the world keeps changing. This inference delay creates a gap of tens to hundreds of milliseconds between the observed state and the state at execution. In this work, we take the natural generalization from zero delay to measured delay during training and inference. We introduce Delay-Aware Diffusion Policy (DA-DP), a framework for explicitly incorporating inference delays into policy learning. DA-DP corrects zero-delay trajectories to their delay-compensated counterparts, and augments the policy with delay conditioning. We empirically validate DA-DP on a variety of tasks, robots, and delays and find its success rate more robust to delay than delay-unaware methods. DA-DP is architecture agnostic and transfers beyond diffusion policies, offering a general pattern for delay-aware imitation learning. More broadly, DA-DP encourages evaluation protocols that report performance as a function of measured latency, not just task difficulty.
Similar Papers
Sample from What You See: Visuomotor Policy Learning via Diffusion Bridge with Observation-Embedded Stochastic Differential Equation
Artificial Intelligence
Robots learn to move better by watching and copying.
Dynamic Test-Time Compute Scaling in Control Policy: Difficulty-Aware Stochastic Interpolant Policy
Robotics
Robots solve harder jobs faster by thinking less.
CDP: Towards Robust Autoregressive Visuomotor Policy Learning via Causal Diffusion
CV and Pattern Recognition
Robots learn better by remembering past actions.