Facts Do Care About Your Language: Assessing Answer Quality of Multilingual LLMs
By: Yuval Kansal, Shmuel Berman, Lydia Liu
Potential Business Impact:
Makes learning tools more truthful for all languages.
Factuality is a necessary precursor to useful educational tools. As adoption of Large Language Models (LLMs) in education continues of grow, ensuring correctness in all settings is paramount. Despite their strong English capabilities, LLM performance in other languages is largely untested. In this work, we evaluate the correctness of the Llama3.1 family of models in answering factual questions appropriate for middle and high school students. We demonstrate that LLMs not only provide extraneous and less truthful information, but also exacerbate existing biases against rare languages.
Similar Papers
Better To Ask in English? Evaluating Factual Accuracy of Multilingual LLMs in English and Low-Resource Languages
Computation and Language
Computers answer questions better in English than Indian languages.
Facts are Harder Than Opinions -- A Multilingual, Comparative Analysis of LLM-Based Fact-Checking Reliability
Computers and Society
Helps computers spot fake news in many languages.
CodeSimpleQA: Scaling Factuality in Code Large Language Models
Computation and Language
Tests if AI truly knows how to code.