ArchSeek: Retrieving Architectural Case Studies Using Vision-Language Models
By: Danrui Li , Yichao Shi , Yaluo Wang and more
Potential Business Impact:
Finds building ideas using pictures and words.
Efficiently searching for relevant case studies is critical in architectural design, as designers rely on precedent examples to guide or inspire their ongoing projects. However, traditional text-based search tools struggle to capture the inherently visual and complex nature of architectural knowledge, often leading to time-consuming and imprecise exploration. This paper introduces ArchSeek, an innovative case study search system with recommendation capability, tailored for architecture design professionals. Powered by the visual understanding capabilities from vision-language models and cross-modal embeddings, it enables text and image queries with fine-grained control, and interaction-based design case recommendations. It offers architects a more efficient, personalized way to discover design inspirations, with potential applications across other visually driven design fields. The source code is available at https://github.com/danruili/ArchSeek.
Similar Papers
ArchiLense: A Framework for Quantitative Analysis of Architectural Styles Based on Vision Large Language Models
CV and Pattern Recognition
Computer sees building styles like an expert.
SARCH: Multimodal Search for Archaeological Archives
Information Retrieval
Finds old book pictures and words faster.
ArchGPT: Understanding the World's Architectures with Large Multimodal Models
Graphics
Lets computers answer questions about buildings.