Far from the Shallow: Brain-Predictive Reasoning Embedding through Residual Disentanglement
By: Linyang He , Tianjun Zhong , Richard Antonello and more
Potential Business Impact:
Finds brain's thinking part separate from words.
Understanding how the human brain progresses from processing simple linguistic inputs to performing high-level reasoning is a fundamental challenge in neuroscience. While modern large language models (LLMs) are increasingly used to model neural responses to language, their internal representations are highly "entangled," mixing information about lexicon, syntax, meaning, and reasoning. This entanglement biases conventional brain encoding analyses toward linguistically shallow features (e.g., lexicon and syntax), making it difficult to isolate the neural substrates of cognitively deeper processes. Here, we introduce a residual disentanglement method that computationally isolates these components. By first probing an LM to identify feature-specific layers, our method iteratively regresses out lower-level representations to produce four nearly orthogonal embeddings for lexicon, syntax, meaning, and, critically, reasoning. We used these disentangled embeddings to model intracranial (ECoG) brain recordings from neurosurgical patients listening to natural speech. We show that: 1) This isolated reasoning embedding exhibits unique predictive power, accounting for variance in neural activity not explained by other linguistic features and even extending to the recruitment of visual regions beyond classical language areas. 2) The neural signature for reasoning is temporally distinct, peaking later (~350-400ms) than signals related to lexicon, syntax, and meaning, consistent with its position atop a processing hierarchy. 3) Standard, non-disentangled LLM embeddings can be misleading, as their predictive success is primarily attributable to linguistically shallow features, masking the more subtle contributions of deeper cognitive processing.
Similar Papers
Large Reasoning Embedding Models: Towards Next-Generation Dense Retrieval Paradigm
Information Retrieval
Helps online shoppers find products even with tricky searches.
Large Reasoning Embedding Models: Towards Next-Generation Dense Retrieval Paradigm
Information Retrieval
Helps online stores find products for tricky searches.
Unsupervised decoding of encoded reasoning using language model interpretability
Artificial Intelligence
Uncovers how AI thinks, even when hidden.