KathDB: Explainable Multimodal Database Management System with Human-AI Collaboration
By: Guorui Xiao , Enhao Zhang , Nicole Sullivan and more
Traditional DBMSs execute user- or application-provided SQL queries over relational data with strong semantic guarantees and advanced query optimization, but writing complex SQL is hard and focuses only on structured tables. Contemporary multimodal systems (which operate over relations but also text, images, and even videos) either expose low-level controls that force users to use (and possibly create) machine learning UDFs manually within SQL or offload execution entirely to black-box LLMs, sacrificing usability or explainability. We propose KathDB, a new system that combines relational semantics with the reasoning power of foundation models over multimodal data. Furthermore, KathDB includes human-AI interaction channels during query parsing, execution, and result explanation, such that users can iteratively obtain explainable answers across data modalities.
Similar Papers
Towards a Multimodal Stream Processing System
Databases
Lets computers understand and react to many things at once.
A Multi-Agent System for Semantic Mapping of Relational Data to Knowledge Graphs
Databases
Connects different computer data for businesses.
Continual Learning of Domain Knowledge from Human Feedback in Text-to-SQL
Computation and Language
Teaches computers to answer questions from data better.