Slide-Level Prompt Learning with Vision Language Models for Few-Shot Multiple Instance Learning in Histopathology
By: Devavrat Tomar , Guillaume Vray , Dwarikanath Mahapatra and more
Potential Business Impact:
Helps doctors find diseases with fewer patient pictures.
In this paper, we address the challenge of few-shot classification in histopathology whole slide images (WSIs) by utilizing foundational vision-language models (VLMs) and slide-level prompt learning. Given the gigapixel scale of WSIs, conventional multiple instance learning (MIL) methods rely on aggregation functions to derive slide-level (bag-level) predictions from patch representations, which require extensive bag-level labels for training. In contrast, VLM-based approaches excel at aligning visual embeddings of patches with candidate class text prompts but lack essential pathological prior knowledge. Our method distinguishes itself by utilizing pathological prior knowledge from language models to identify crucial local tissue types (patches) for WSI classification, integrating this within a VLM-based MIL framework. Our approach effectively aligns patch images with tissue types, and we fine-tune our model via prompt learning using only a few labeled WSIs per category. Experimentation on real-world pathological WSI datasets and ablation studies highlight our method's superior performance over existing MIL- and VLM-based methods in few-shot WSI classification tasks. Our code is publicly available at https://github.com/LTS5/SLIP.
Similar Papers
Investigating Zero-Shot Diagnostic Pathology in Vision-Language Models with Efficient Prompt Design
CV and Pattern Recognition
Helps doctors find cancer faster using AI pictures.
Leveraging Vision-Language Embeddings for Zero-Shot Learning in Histopathology Images
CV and Pattern Recognition
Helps doctors find diseases in pictures without training.
Multi-Resolution Pathology-Language Pre-training Model with Text-Guided Visual Representation
CV and Pattern Recognition
Helps doctors see cancer details better.