Universal computation is intrinsic to language model decoding
By: Alex Lewandowski, Marlos C. Machado, Dale Schuurmans
Language models now provide an interface to express and often solve general problems in natural language, yet their ultimate computational capabilities remain a major topic of scientific debate. Unlike a formal computer, a language model is trained to autoregressively predict successive elements in human-generated text. We prove that chaining a language model's autoregressive output is sufficient to perform universal computation. That is, a language model can simulate the execution of any algorithm on any input. The challenge of eliciting desired computational behaviour can thus be reframed in terms of programmability: the ease of finding a suitable prompt. Strikingly, we demonstrate that even randomly initialized language models are capable of universal computation before training. This implies that training does not give rise to computational expressiveness -- rather, it improves programmability, enabling a natural language interface for accessing these intrinsic capabilities.
Similar Papers
Training Language Models to Explain Their Own Computations
Computation and Language
Lets computers explain how they think.
Large language models are not about language
Computation and Language
Teaches computers how minds learn language.
What do language models model? Transformers, automata, and the format of thought
Computation and Language
Computers learn language like a machine, not a brain.