Modeling cognitive processes of natural reading with transformer-based Language Models
By: Bruno Bianchi, Fermín Travi, Juan E. Kamienkowski
Potential Business Impact:
Helps computers understand how people read.
Recent advances in Natural Language Processing (NLP) have led to the development of highly sophisticated language models for text generation. In parallel, neuroscience has increasingly employed these models to explore cognitive processes involved in language comprehension. Previous research has shown that models such as N-grams and LSTM networks can partially account for predictability effects in explaining eye movement behaviors, specifically Gaze Duration, during reading. In this study, we extend these findings by evaluating transformer-based models (GPT2, LLaMA-7B, and LLaMA2-7B) to further investigate this relationship. Our results indicate that these architectures outperform earlier models in explaining the variance in Gaze Durations recorded from Rioplantense Spanish readers. However, similar to previous studies, these models still fail to account for the entirety of the variance captured by human predictability. These findings suggest that, despite their advancements, state-of-the-art language models continue to predict language in ways that differ from human readers.
Similar Papers
Advancements in Natural Language Processing: Exploring Transformer-Based Architectures for Text Understanding
Computation and Language
Computers now understand and write like people.
From Idea to Implementation: Evaluating the Influence of Large Language Models in Software Development -- An Opinion Paper
Artificial Intelligence
Helps programmers code faster and better.
Surprisal from Larger Transformer-based Language Models Predicts fMRI Data More Poorly
Computation and Language
Brain scans show how well computers understand words.