Few-Shot Adaptation Benchmark for Remote Sensing Vision-Language Models
By: Karim El Khoury , Maxime Zanella , Christophe De Vleeschouwer and more
Potential Business Impact:
Helps AI learn from few pictures of Earth.
Remote Sensing Vision-Language Models (RSVLMs) have shown remarkable potential thanks to large-scale pretraining, achieving strong zero-shot performance on various tasks. However, their ability to generalize in low-data regimes, such as few-shot learning, remains insufficiently explored. In this work, we present the first structured benchmark for evaluating few-shot adaptation methods on RSVLMs. We conduct comprehensive experiments across ten remote sensing scene classification datasets, applying five widely used few-shot adaptation strategies to three state-of-the-art RSVLMs with varying backbones. Our findings reveal that models with similar zero-shot performance can exhibit markedly different behavior under few-shot adaptation, with some RSVLMs being inherently more amenable to such adaptation than others. The variability of performance and the absence of a clear winner among existing methods highlight the need for the development of more robust methods for few-shot adaptation tailored to RS. To facilitate future research, we provide a reproducible benchmarking framework and open-source code to systematically evaluate RSVLMs under few-shot conditions. The source code is publicly available on Github: https://github.com/elkhouryk/fewshot_RSVLMs
Similar Papers
A Recipe for Improving Remote Sensing VLM Zero Shot Generalization
CV and Pattern Recognition
Teaches computers to understand satellite pictures.
Robust Few-Shot Vision-Language Model Adaptation
CV and Pattern Recognition
Helps AI understand images it hasn't seen.
A Vision Centric Remote Sensing Benchmark
CV and Pattern Recognition
Helps computers understand satellite pictures better.