Explicit Learning and the LLM in Machine Translation
By: Malik Marmonier, Rachel Bawden, Benoît Sagot
Potential Business Impact:
Computers learn new languages from books.
This study explores an LLM's ability to learn new languages using explanations found in a grammar book, a process we term "explicit learning." To rigorously assess this ability, we design controlled translation experiments between English and constructed languages generated, through specific cryptographic means, from Latin or French. Contrary to previous studies, our results demonstrate that LLMs do possess a measurable capacity for explicit learning. This ability, however, diminishes as the complexity of the linguistic phenomena to be learned increases. Supervised fine-tuning on ad hoc chains of thought significantly enhances LLM performance but struggles to generalize to typologically novel or more complex linguistic features. These findings point to the need for more diverse training sets and alternative fine-tuning strategies to further improve explicit learning by LLMs, benefiting low-resource languages typically described in grammar books but lacking extensive corpora.
Similar Papers
ExpeTrans: LLMs Are Experiential Transfer Learners
Computation and Language
Computers learn new tasks without human help.
Translation in the Wild
Computation and Language
Computers learn to translate languages by reading the internet.
LLMs Are Globally Multilingual Yet Locally Monolingual: Exploring Knowledge Transfer via Language and Thought Theory
Computation and Language
Helps computers understand facts in any language.