Score: 1

Effortless Vision-Language Model Specialization in Histopathology without Annotation

Published: August 11, 2025 | arXiv ID: 2508.07835v1

By: Jingna Qiu , Nishanth Jain , Jonas Ammeling and more

Potential Business Impact:

Teaches AI to see tiny body parts better.

Recent advances in Vision-Language Models (VLMs) in histopathology, such as CONCH and QuiltNet, have demonstrated impressive zero-shot classification capabilities across various tasks. However, their general-purpose design may lead to suboptimal performance in specific downstream applications. While supervised fine-tuning methods address this issue, they require manually labeled samples for adaptation. This paper investigates annotation-free adaptation of VLMs through continued pretraining on domain- and task-relevant image-caption pairs extracted from existing databases. Our experiments on two VLMs, CONCH and QuiltNet, across three downstream tasks reveal that these pairs substantially enhance both zero-shot and few-shot performance. Notably, with larger training sizes, continued pretraining matches the performance of few-shot methods while eliminating manual labeling. Its effectiveness, task-agnostic design, and annotation-free workflow make it a promising pathway for adapting VLMs to new histopathology tasks. Code is available at https://github.com/DeepMicroscopy/Annotation-free-VLM-specialization.

Country of Origin
🇩🇪 Germany

Repos / Data Links

Page Count
12 pages

Category
Computer Science:
CV and Pattern Recognition