January Food Benchmark (JFB): A Public Benchmark Dataset and Evaluation Suite for Multimodal Food Analysis
By: Amir Hosseinian , Ashkan Dehghani Zahedani , Umer Mansoor and more
Potential Business Impact:
Helps computers guess food nutrition from pictures.
Progress in AI for automated nutritional analysis is critically hampered by the lack of standardized evaluation methodologies and high-quality, real-world benchmark datasets. To address this, we introduce three primary contributions. First, we present the January Food Benchmark (JFB), a publicly available collection of 1,000 food images with human-validated annotations. Second, we detail a comprehensive benchmarking framework, including robust metrics and a novel, application-oriented overall score designed to assess model performance holistically. Third, we provide baseline results from both general-purpose Vision-Language Models (VLMs) and our own specialized model, january/food-vision-v1. Our evaluation demonstrates that the specialized model achieves an Overall Score of 86.2, a 12.1-point improvement over the best-performing general-purpose configuration. This work offers the research community a valuable new evaluation dataset and a rigorous framework to guide and benchmark future developments in automated nutritional analysis.
Similar Papers
FCMBench: A Comprehensive Financial Credit Multimodal Benchmark for Real-world Applications
CV and Pattern Recognition
Helps AI understand loan papers for better decisions.
FCMBench: A Comprehensive Financial Credit Multimodal Benchmark for Real-world Applications
CV and Pattern Recognition
Tests AI for loan applications.
BenchSeg: A Large-Scale Dataset and Benchmark for Multi-View Food Video Segmentation
CV and Pattern Recognition
Helps computers count food calories from videos.