Hierarchical Vision-Language Reasoning for Multimodal Multiple-Choice Question Answering
By: Ao Zhou , Zebo Gu , Tenghao Sun and more
Potential Business Impact:
Helps computers understand Japanese documents better.
Multimodal Large Language Models (MLLMs) have demonstrated remarkable multimodal understanding capabilities in Visual Question Answering (VQA) tasks by integrating visual and textual features. However, under the challenging ten-choice question evaluation paradigm, existing methods still exhibit significant limitations when processing PDF documents with complex layouts and lengthy content. Notably, current mainstream models suffer from a strong bias toward English training data, resulting in suboptimal performance for Japanese and other language scenarios. To address these challenges, this paper proposes a novel Japanese PDF document understanding framework that combines multimodal hierarchical reasoning mechanisms with Colqwen-optimized retrieval methods, while innovatively introducing a semantic verification strategy through sub-question decomposition. Experimental results demonstrate that our framework not only significantly enhances the model's deep semantic parsing capability for complex documents, but also exhibits superior robustness in practical application scenarios.
Similar Papers
Knowledge-based Visual Question Answer with Multimodal Processing, Retrieval and Filtering
CV and Pattern Recognition
Helps computers answer questions using pictures and facts.
Unexplored flaws in multiple-choice VQA evaluations
CV and Pattern Recognition
Makes AI answers change just by changing the question's words.
Elevating Visual Question Answering through Implicitly Learned Reasoning Pathways in LVLMs
CV and Pattern Recognition
Helps computers understand pictures by asking themselves questions.