JBE-QA: Japanese Bar Exam QA Dataset for Assessing Legal Domain Knowledge
By: Zhihan Cao , Fumihito Nishino , Hiroaki Yamada and more
Potential Business Impact:
Tests if computers understand Japanese law.
We introduce JBE-QA, a Japanese Bar Exam Question-Answering dataset to evaluate large language models' legal knowledge. Derived from the multiple-choice (tanto-shiki) section of the Japanese bar exam (2015-2024), JBE-QA provides the first comprehensive benchmark for Japanese legal-domain evaluation of LLMs. It covers the Civil Code, the Penal Code, and the Constitution, extending beyond the Civil Code focus of prior Japanese resources. Each question is decomposed into independent true/false judgments with structured contextual fields. The dataset contains 3,464 items with balanced labels. We evaluate 26 LLMs, including proprietary, open-weight, Japanese-specialised, and reasoning models. Our results show that proprietary models with reasoning enabled perform best, and the Constitution questions are generally easier than the Civil Code or the Penal Code questions.
Similar Papers
MizanQA: Benchmarking Large Language Models on Moroccan Legal Question Answering
Computation and Language
Tests computers on Arabic law questions.
MultiWikiQA: A Reading Comprehension Benchmark in 300+ Languages
Computation and Language
Helps computers understand text in many languages.
MultiWikiQA: A Reading Comprehension Benchmark in 300+ Languages
Computation and Language
Helps computers understand Wikipedia in many languages.