Information Gradient for Directed Acyclic Graphs: A Score-based Framework for End-to-End Mutual Information Maximization
By: Tadashi Wadayama
Potential Business Impact:
Helps computers learn to send and get information better.
This paper presents a general framework for end-to-end mutual information maximization in communication and sensing systems represented by stochastic directed acyclic graphs (DAGs). We derive a unified formula for the (mutual) information gradient with respect to arbitrary internal parameters, utilizing marginal and conditional score functions. We demonstrate that this gradient can be efficiently computed using vector-Jacobian products (VJP) within standard automatic differentiation frameworks, enabling the optimization of complex networks under global resource constraints. Numerical experiments on both linear multipath DAGs and nonlinear channels validate the proposed framework; the results confirm that the estimator, utilizing score functions learned via denoising score matching, accurately reproduces ground-truth gradients and successfully maximizes end-to-end mutual information. Beyond maximization, we extend our score-based framework to a novel unsupervised paradigm: digital twin calibration via Fisher divergence minimization.
Similar Papers
Information Gradient for Nonlinear Gaussian Channel with Applications to Task-Oriented Communication
Information Theory
Improves how machines learn from noisy signals.
Transformers Provably Learn Directed Acyclic Graphs via Kernel-Guided Mutual Information
Machine Learning (CS)
Helps computers find hidden connections in data.
Mutual Information Estimation via Score-to-Fisher Bridge for Nonlinear Gaussian Noise Channels
Information Theory
Helps computers understand messy signals better.