ChartComplete: A Taxonomy-based Inclusive Chart Dataset
By: Ahmad Mustapha, Charbel Toumieh, Mariette Awad
With advancements in deep learning (DL) and computer vision techniques, the field of chart understanding is evolving rapidly. In particular, multimodal large language models (MLLMs) are proving to be efficient and accurate in understanding charts. To accurately measure the performance of MLLMs, the research community has developed multiple datasets to serve as benchmarks. By examining these datasets, we found that they are all limited to a small set of chart types. To bridge this gap, we propose the ChartComplete dataset. The dataset is based on a chart taxonomy borrowed from the visualization community, and it covers thirty different chart types. The dataset is a collection of classified chart images and does not include a learning signal. We present the ChartComplete dataset as is to the community to build upon it.
Similar Papers
BigCharts-R1: Enhanced Chart Reasoning with Visual Reinforcement Finetuning
Computation and Language
Helps computers understand charts better.
In-Depth and In-Breadth: Pre-training Multimodal Language Models Customized for Comprehensive Chart Understanding
Computation and Language
Helps computers understand any kind of chart.
Chart-to-Experience: Benchmarking Multimodal LLMs for Predicting Experiential Impact of Charts
Human-Computer Interaction
Helps computers understand how charts make people feel.