Benchmarking Vision-Language Models on Chinese Ancient Documents: From OCR to Knowledge Reasoning
By: Haiyang Yu , Yuchuan Wu , Fan Shi and more
Potential Business Impact:
Helps computers understand old Chinese writing.
Chinese ancient documents, invaluable carriers of millennia of Chinese history and culture, hold rich knowledge across diverse fields but face challenges in digitization and understanding, i.e., traditional methods only scan images, while current Vision-Language Models (VLMs) struggle with their visual and linguistic complexity. Existing document benchmarks focus on English printed texts or simplified Chinese, leaving a gap for evaluating VLMs on ancient Chinese documents. To address this, we present AncientDoc, the first benchmark for Chinese ancient documents, designed to assess VLMs from OCR to knowledge reasoning. AncientDoc includes five tasks (page-level OCR, vernacular translation, reasoning-based QA, knowledge-based QA, linguistic variant QA) and covers 14 document types, over 100 books, and about 3,000 pages. Based on AncientDoc, we evaluate mainstream VLMs using multiple metrics, supplemented by a human-aligned large language model for scoring.
Similar Papers
VisTW: Benchmarking Vision-Language Models for Traditional Chinese in Taiwan
Computation and Language
Tests computers understanding Chinese pictures and words.
AncientBench: Towards Comprehensive Evaluation on Excavated and Transmitted Chinese Corpora
Computation and Language
Helps computers understand old Chinese writings.
ThaiOCRBench: A Task-Diverse Benchmark for Vision-Language Understanding in Thai
Computation and Language
Helps computers understand Thai documents better.