Multi-modal Knowledge Graph Generation with Semantics-enriched Prompts
By: Yajing Xu , Zhiqiang Liu , Jiaoyan Chen and more
Potential Business Impact:
Makes smart pictures for computer knowledge.
Multi-modal Knowledge Graphs (MMKGs) have been widely applied across various domains for knowledge representation. However, the existing MMKGs are significantly fewer than required, and their construction faces numerous challenges, particularly in ensuring the selection of high-quality, contextually relevant images for knowledge graph enrichment. To address these challenges, we present a framework for constructing MMKGs from conventional KGs. Furthermore, to generate higher-quality images that are more relevant to the context in the given knowledge graph, we designed a neighbor selection method called Visualizable Structural Neighbor Selection (VSNS). This method consists of two modules: Visualizable Neighbor Selection (VNS) and Structural Neighbor Selection (SNS). The VNS module filters relations that are difficult to visualize, while the SNS module selects neighbors that most effectively capture the structural characteristics of the entity. To evaluate the quality of the generated images, we performed qualitative and quantitative evaluations on two datasets, MKG-Y and DB15K. The experimental results indicate that using the VSNS method to select neighbors results in higher-quality images that are more relevant to the knowledge graph.
Similar Papers
Aligning Vision to Language: Annotation-Free Multimodal Knowledge Graph Construction for Enhanced LLMs Reasoning
CV and Pattern Recognition
Helps computers understand pictures and words together better.
Guided Navigation in Knowledge-Dense Environments: Structured Semantic Exploration with Guidance Graphs
Computation and Language
Helps computers find facts faster and smarter.
Meta-Semantics Augmented Few-Shot Relational Learning
Artificial Intelligence
Teaches computers to learn new facts with few examples.