XChoice: Explainable Evaluation of AI-Human Alignment in LLM-based Constrained Choice Decision Making
By: Weihong Qi , Fan Huang , Rasika Muralidharan and more
Potential Business Impact:
Helps AI understand why people make choices.
We present XChoice, an explainable framework for evaluating AI-human alignment in constrained decision making. Moving beyond outcome agreement such as accuracy and F1 score, XChoice fits a mechanism-based decision model to human data and LLM-generated decisions, recovering interpretable parameters that capture the relative importance of decision factors, constraint sensitivity, and implied trade-offs. Alignment is assessed by comparing these parameter vectors across models, options, and subgroups. We demonstrate XChoice on Americans' daily time allocation using the American Time Use Survey (ATUS) as human ground truth, revealing heterogeneous alignment across models and activities and salient misalignment concentrated in Black and married groups. We further validate robustness of XChoice via an invariance analysis and evaluate targeted mitigation with a retrieval augmented generation (RAG) intervention. Overall, XChoice provides mechanism-based metrics that diagnose misalignment and support informed improvements beyond surface outcome matching.
Similar Papers
A Survey on Human-Centered Evaluation of Explainable AI Methods in Clinical Decision Support Systems
Machine Learning (CS)
Makes doctors trust computer health advice.
Mind the XAI Gap: A Human-Centered LLM Framework for Democratizing Explainable AI
Machine Learning (CS)
Explains AI decisions for everyone, not just experts.
Towards Cognitively-Faithful Decision-Making Models to Improve AI Alignment
Machine Learning (CS)
Teaches computers to think like people.