What can we learn from signals and systems in a transformer? Insights for probabilistic modeling and inference architecture
By: Heng-Sheng Chang, Prashant G. Mehta
Potential Business Impact:
Lets computers guess words by understanding patterns.
In the 1940s, Wiener introduced a linear predictor, where the future prediction is computed by linearly combining the past data. A transformer generalizes this idea: it is a nonlinear predictor where the next-token prediction is computed by nonlinearly combining the past tokens. In this essay, we present a probabilistic model that interprets transformer signals as surrogates of conditional measures, and layer operations as fixed-point updates. An explicit form of the fixed-point update is described for the special case when the probabilistic model is a hidden Markov model (HMM). In part, this paper is in an attempt to bridge the classical nonlinear filtering theory with modern inference architectures.
Similar Papers
Towards Understanding Transformers in Learning Random Walks
Machine Learning (CS)
Shows how computers learn to predict movement.
Towards Theoretical Understanding of Transformer Test-Time Computing: Investigation on In-Context Linear Regression
Machine Learning (CS)
Makes computer writing smarter by trying many ideas.
Towards Theoretical Understanding of Transformer Test-Time Computing: Investigation on In-Context Linear Regression
Machine Learning (CS)
Makes AI think more to give better answers.