Score: 1

KatotohananQA: Evaluating Truthfulness of Large Language Models in Filipino

Published: September 7, 2025 | arXiv ID: 2509.06065v1

By: Lorenzo Alfred Nery, Ronald Dawson Catignas, Thomas James Tiam-Lee

Potential Business Impact:

Tests if AI tells truth in Filipino language.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Large Language Models (LLMs) achieve remarkable performance across various tasks, but their tendency to produce hallucinations limits reliable adoption. Benchmarks such as TruthfulQA have been developed to measure truthfulness, yet they are primarily available in English, leaving a gap in evaluating LLMs in low-resource languages. To address this, we present KatotohananQA, a Filipino translation of the TruthfulQA benchmark. Seven free-tier proprietary models were assessed using a binary-choice framework. Findings show a significant performance gap between English and Filipino truthfulness, with newer OpenAI models (GPT-5 and GPT-5 mini) demonstrating strong multilingual robustness. Results also reveal disparities across question characteristics, suggesting that some question types, categories, and topics are less robust to multilingual transfer which highlight the need for broader multilingual evaluation to ensure fairness and reliability in LLM usage.

Country of Origin
🇵🇭 Philippines

Repos / Data Links

Page Count
14 pages

Category
Computer Science:
Computation and Language