Mitigating LLM Hallucinations with Knowledge Graphs: A Case Study
By: Harry Li , Gabriel Appleby , Kenneth Alperin and more
Potential Business Impact:
AI learns facts to stop making things up.
High-stakes domains like cyber operations need responsible and trustworthy AI methods. While large language models (LLMs) are becoming increasingly popular in these domains, they still suffer from hallucinations. This research paper provides learning outcomes from a case study with LinkQ, an open-source natural language interface that was developed to combat hallucinations by forcing an LLM to query a knowledge graph (KG) for ground-truth data during question-answering (QA). We conduct a quantitative evaluation of LinkQ using a well-known KGQA dataset, showing that the system outperforms GPT-4 but still struggles with certain question categories - suggesting that alternative query construction strategies will need to be investigated in future LLM querying systems. We discuss a qualitative study of LinkQ with two domain experts using a real-world cybersecurity KG, outlining these experts' feedback, suggestions, perceived limitations, and future opportunities for systems like LinkQ.
Similar Papers
Lie to Me: Knowledge Graphs for Robust Hallucination Self-Detection in LLMs
Computation and Language
Makes AI tell the truth, not make things up.
The Role of Visualization in LLM-Assisted Knowledge Graph Systems: Effects on User Trust, Exploration, and Workflows
Machine Learning (CS)
Helps people understand complex data by asking questions.
Aligning Knowledge Graphs and Language Models for Factual Accuracy
Computation and Language
Makes AI tell the truth, not make things up.