Symbol Distributions in Semantic Communications: A Source-Channel Equilibrium Perspective
By: Hanju Yoo , Dongha Choi , Songkuk Kim and more
Semantic communication systems often use an end-to-end neural network to map input data into continuous symbols. These symbols, which are essentially neural network features, usually have fixed dimensions and heavy-tailed distributions. However, due to the end-to-end training nature of the neural network encoder, the underlying reason for the symbol distribution remains underexplored. We propose a new explanation for the semantic symbol distribution: an inherent trade-off between source coding and communications. Specifically, the encoder balances two objectives: allocating power for minimum \emph{effective codelength} (for source coding) and maximizing mutual information (for communications). We formalize this trade-off via an information-theoretic optimization framework, which yields a Student's $t$-distribution as the resulting symbol distribution. Through extensive studies on image-based semantic systems, we find that our formulation models the learned symbols and predicts how the symbol distribution's shape parameter changes with respect to (i) the use of variable-length coding and (ii) the dataset's entropy variability. Furthermore, we demonstrate how introducing a regularizer that enforces a target symbol distribution, which guides the encoder towards a target prior (e.g., Gaussian), improves training convergence and supports our hypothesis.
Similar Papers
Semantic Communication with Distribution Learning through Sequential Observations
Machine Learning (CS)
Helps computers understand messages, not just words.
Secure Semantic Communication over Wiretap Channels: Rate-Distortion-Equivocation Tradeoff
Information Theory
Keeps secret messages safe from spies.
Variable-Length Joint Source-Channel Coding for Semantic Communication
Information Theory
Makes computers understand messages with fewer bits.