Interpretable and Robust Dialogue State Tracking via Natural Language Summarization with LLMs
By: Rafael Carranza, Mateo Alejandro Rojas
Potential Business Impact:
Helps chatbots understand what you're saying better.
This paper introduces a novel approach to Dialogue State Tracking (DST) that leverages Large Language Models (LLMs) to generate natural language descriptions of dialogue states, moving beyond traditional slot-value representations. Conventional DST methods struggle with open-domain dialogues and noisy inputs. Motivated by the generative capabilities of LLMs, our Natural Language DST (NL-DST) framework trains an LLM to directly synthesize human-readable state descriptions. We demonstrate through extensive experiments on MultiWOZ 2.1 and Taskmaster-1 datasets that NL-DST significantly outperforms rule-based and discriminative BERT-based DST baselines, as well as generative slot-filling GPT-2 DST models, in both Joint Goal Accuracy and Slot Accuracy. Ablation studies and human evaluations further validate the effectiveness of natural language state generation, highlighting its robustness to noise and enhanced interpretability. Our findings suggest that NL-DST offers a more flexible, accurate, and human-understandable approach to dialogue state tracking, paving the way for more robust and adaptable task-oriented dialogue systems.
Similar Papers
Beyond Single-User Dialogue: Assessing Multi-User Dialogue State Tracking Capabilities of Large Language Models
Computation and Language
Makes computers understand many people talking at once.
Approaching Dialogue State Tracking via Aligning Speech Encoders and LLMs
Audio and Speech Processing
Helps computers understand what people say.
Joint Speech and Text Training for LLM-Based End-to-End Spoken Dialogue State Tracking
Computation and Language
Lets computers understand spoken words in new situations.