SANDWiCH: Semantical Analysis of Neighbours for Disambiguating Words in Context ad Hoc
By: Daniel Guzman-Olivares, Lara Quijano-Sanchez, Federico Liberatore
Potential Business Impact:
Helps computers understand words like people do.
The rise of generative chat-based Large Language Models (LLMs) over the past two years has spurred a race to develop systems that promise near-human conversational and reasoning experiences. However, recent studies indicate that the language understanding offered by these models remains limited and far from human-like performance, particularly in grasping the contextual meanings of words, an essential aspect of reasoning. In this paper, we present a simple yet computationally efficient framework for multilingual Word Sense Disambiguation (WSD). Our approach reframes the WSD task as a cluster discrimination analysis over a semantic network refined from BabelNet using group algebra. We validate our methodology across multiple WSD benchmarks, achieving a new state of the art for all languages and tasks, as well as in individual assessments by part of speech. Notably, our model significantly surpasses the performance of current alternatives, even in low-resource languages, while reducing the parameter count by 72%.
Similar Papers
Exploring the Word Sense Disambiguation Capabilities of Large Language Models
Computation and Language
AI now understands word meanings better than ever.
Solving Word-Sense Disambiguation and Word-Sense Induction with Dictionary Examples
Computation and Language
Helps computers understand rare languages using dictionaries.
Do Large Language Models Understand Word Senses?
Computation and Language
Computers now understand word meanings in sentences.