Generative vector search to improve pathology foundation models across multimodal vision-language tasks
By: Markus Ekvall , Ludvig Bergenstråhle , Patrick Truong and more
Retrieval-augmented generation improves large language models by grounding outputs in external knowledge sources, reducing hallucinations and addressing knowledge cutoffs. However, standard embedding-based retrieval fails to capture the complexity of multi-concept queries, particularly in domains like biomedicine, where biological data are inherently high-dimensional. For example,omics datasets, and clinical reports simultaneously exhibit numerous molecular, cellular, and physiological features. We present Stochastic Latent Matching (STHLM), a generative vector search method that samples query-conditioned embeddings from text or image inputs to enhance retrieval performance. Analogous to how Chain-of-Thought reasoning enables language models to "think longer" on complex problems, STHLM allows retrieval systems to "search wider" through iterative sampling. STHLM demonstrates critical improvements over classical vector retrieval across diverse benchmarks, including scientific literature, clinical notes, and tissue images, boosting retrieval performance by 10-30% through test-time compute (trading latency for accuracy), while enabling up to a 10-fold compression of embedding dimensions.
Similar Papers
Enhancing Subsequent Video Retrieval via Vision-Language Models (VLMs)
CV and Pattern Recognition
Find videos faster by understanding their stories.
LLM-guided Hierarchical Retrieval
Information Retrieval
Helps computers find information faster in huge libraries.
A Little More Like This: Text-to-Image Retrieval with Vision-Language Models Using Relevance Feedback
CV and Pattern Recognition
Improves image search by learning from results.