SuperCap: Multi-resolution Superpixel-based Image Captioning
By: Henry Senior , Luca Rossi , Gregory Slabaugh and more
Potential Business Impact:
Helps computers describe pictures without knowing all objects.
It has been a longstanding goal within image captioning to move beyond a dependence on object detection. We investigate using superpixels coupled with Vision Language Models (VLMs) to bridge the gap between detector-based captioning architectures and those that solely pretrain on large datasets. Our novel superpixel approach ensures that the model receives object-like features whilst the use of VLMs provides our model with open set object understanding. Furthermore, we extend our architecture to make use of multi-resolution inputs, allowing our model to view images in different levels of detail, and use an attention mechanism to determine which parts are most relevant to the caption. We demonstrate our model's performance with multiple VLMs and through a range of ablations detailing the impact of different architectural choices. Our full model achieves a competitive CIDEr score of $136.9$ on the COCO Karpathy split.
Similar Papers
Generating Accurate and Detailed Captions for High-Resolution Images
CV and Pattern Recognition
Makes computer pictures describe more details accurately.
Semantic-Clipping: Efficient Vision-Language Modeling with Semantic-Guidedd Visual Selection
CV and Pattern Recognition
Helps computers understand pictures better by focusing on important parts.
Low-hallucination Synthetic Captions for Large-Scale Vision-Language Model Pre-training
CV and Pattern Recognition
Makes computers understand pictures better with fake text.