FlipVQA-Miner: Cross-Page Visual Question-Answer Mining from Textbooks
By: Zhen Hao Wong , Jingwen Deng , Hao Liang and more
Potential Business Impact:
Makes AI smarter using old school books.
The development of Large Language Models (LLMs) increasingly depends on high-quality supervised data, yet existing instruction-tuning and RL datasets remain costly to curate and often rely on synthetic samples that introduce hallucination and limited diversity. At the same time, textbooks and exercise materials contain abundant, high-quality human-authored Question-Answer(QA) content that remains underexploited due to the difficulty of transforming raw PDFs into AI-ready supervision. Although modern OCR and vision-language models can accurately parse document structure, their outputs lack the semantic alignment required for training. We propose an automated pipeline that extracts well-formed QA and visual-QA (VQA) pairs from educational documents by combining layout-aware OCR with LLM-based semantic parsing. Experiments across diverse document types show that the method produces accurate, aligned, and low-noise QA/VQA pairs. This approach enables scalable use of real-world educational content and provides a practical alternative to synthetic data generation for improving reasoning-oriented LLM training. All code and data-processing pipelines are open-sourced at https://github.com/OpenDCAI/DataFlow.
Similar Papers
Text-VQA Aug: Pipelined Harnessing of Large Multimodal Models for Automated Synthesis
CV and Pattern Recognition
Computers can now answer questions about text in pictures.
VQ-VA World: Towards High-Quality Visual Question-Visual Answering
CV and Pattern Recognition
Makes computers draw pictures from questions.
PustakAI: Curriculum-Aligned and Interactive Textbooks Using Large Language Models
Computation and Language
Helps AI answer school questions accurately.