Multilingual Relative Clause Attachment Ambiguity Resolution in Large Language Models
By: So Young Lee , Russell Scheinberg , Amber Shore and more
Potential Business Impact:
Computers understand sentences better, but not all languages.
This study examines how large language models (LLMs) resolve relative clause (RC) attachment ambiguities and compares their performance to human sentence processing. Focusing on two linguistic factors, namely the length of RCs and the syntactic position of complex determiner phrases (DPs), we assess whether LLMs can achieve human-like interpretations amid the complexities of language. In this study, we evaluated several LLMs, including Claude, Gemini and Llama, in multiple languages: English, Spanish, French, German, Japanese, and Korean. While these models performed well in Indo-European languages (English, Spanish, French, and German), they encountered difficulties in Asian languages (Japanese and Korean), often defaulting to incorrect English translations. The findings underscore the variability in LLMs' handling of linguistic ambiguities and highlight the need for model improvements, particularly for non-European languages. This research informs future enhancements in LLM design to improve accuracy and human-like processing in diverse linguistic environments.
Similar Papers
Investigating Syntactic Biases in Multilingual Transformers with RC Attachment Ambiguities in Italian and English
Computation and Language
Computers don't understand sentences like people do.
Who Relies More on World Knowledge and Bias for Syntactic Ambiguity Resolution: Humans or LLMs?
Computation and Language
Computers struggle to understand tricky sentences.
It Depends: Resolving Referential Ambiguity in Minimal Contexts with Commonsense Knowledge
Computation and Language
Helps computers understand confusing words in chats.