Loop as a Bridge: Can Looped Transformers Truly Link Representation Space and Natural Language Outputs?
By: Guanxu Chen, Dongrui Liu, Jing Shao
Large Language Models (LLMs) often exhibit a gap between their internal knowledge and their explicit linguistic outputs. In this report, we empirically investigate whether Looped Transformers (LTs)--architectures that increase computational depth by iterating shared layers--can bridge this gap by utilizing their iterative nature as a form of introspection. Our experiments reveal that while increasing loop iterations narrows the gap, it is partly driven by a degradation of their internal knowledge carried by representations. Moreover, another empirical analysis suggests that current LTs' ability to perceive representations does not improve across loops; it is only present in the final loop. These results suggest that while LTs offer a promising direction for scaling computational depth, they have yet to achieve the introspection required to truly link representation space and natural language.
Similar Papers
Reasoning with Latent Thoughts: On the Power of Looped Transformers
Computation and Language
Makes computers solve hard problems with fewer parts.
Parallel Loop Transformer for Efficient Test-Time Computation Scaling
Computation and Language
Makes smart computer programs faster and cheaper.
A Survey on Large Language Models with some Insights on their Capabilities and Limitations
Computation and Language
Computers learn to think and solve problems.