Score: 2

Learning what to say and how precisely: Efficient Communication via Differentiable Discrete Communication Learning

Published: November 3, 2025 | arXiv ID: 2511.01554v1

By: Aditya Kapoor , Yash Bhisikar , Benjamin Freed and more

Potential Business Impact:

Agents learn to send smarter, smaller messages.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Effective communication in multi-agent reinforcement learning (MARL) is critical for success but constrained by bandwidth, yet past approaches have been limited to complex gating mechanisms that only decide \textit{whether} to communicate, not \textit{how precisely}. Learning to optimize message precision at the bit-level is fundamentally harder, as the required discretization step breaks gradient flow. We address this by generalizing Differentiable Discrete Communication Learning (DDCL), a framework for end-to-end optimization of discrete messages. Our primary contribution is an extension of DDCL to support unbounded signals, transforming it into a universal, plug-and-play layer for any MARL architecture. We verify our approach with three key results. First, through a qualitative analysis in a controlled environment, we demonstrate \textit{how} agents learn to dynamically modulate message precision according to the informational needs of the task. Second, we integrate our variant of DDCL into four state-of-the-art MARL algorithms, showing it reduces bandwidth by over an order of magnitude while matching or exceeding task performance. Finally, we provide direct evidence for the \enquote{Bitter Lesson} in MARL communication: a simple Transformer-based policy leveraging DDCL matches the performance of complex, specialized architectures, questioning the necessity of bespoke communication designs.

Country of Origin
🇬🇧 United Kingdom


Page Count
30 pages

Category
Computer Science:
Multiagent Systems