Hide or Highlight: Understanding the Impact of Factuality Expression on User Trust
By: Hyo Jin Do, Werner Geyer
Potential Business Impact:
Hides bad AI answers to build trust.
Large language models are known to produce outputs that are plausible but factually incorrect. To prevent people from making erroneous decisions by blindly trusting AI, researchers have explored various ways of communicating factuality estimates in AI-generated outputs to end-users. However, little is known about whether revealing content estimated to be factually incorrect influences users' trust when compared to hiding it altogether. We tested four different ways of disclosing an AI-generated output with factuality assessments: transparent (highlights less factual content), attention (highlights factual content), opaque (removes less factual content), ambiguity (makes less factual content vague), and compared them with a baseline response without factuality information. We conducted a human subjects research (N = 148) using the strategies in question-answering scenarios. We found that the opaque and ambiguity strategies led to higher trust while maintaining perceived answer quality, compared to the other strategies. We discuss the efficacy of hiding presumably less factual content to build end-user trust.
Similar Papers
Highlight All the Phrases: Enhancing LLM Transparency through Visual Factuality Indicators
Human-Computer Interaction
Colors show if AI is telling the truth.
Self-Transparency Failures in Expert-Persona LLMs: How Instruction-Following Overrides Honesty
Artificial Intelligence
AI tells you when it's pretending to be a doctor.
Users Favor LLM-Generated Content -- Until They Know It's AI
Human-Computer Interaction
People like AI answers more when they don't know it's AI.