Beyond Single-User Dialogue: Assessing Multi-User Dialogue State Tracking Capabilities of Large Language Models
By: Sangmin Song , Juhwan Choi , JungMin Yun and more
Potential Business Impact:
Makes computers understand many people talking at once.
Large language models (LLMs) have demonstrated remarkable performance in zero-shot dialogue state tracking (DST), reducing the need for task-specific training. However, conventional DST benchmarks primarily focus on structured user-agent conversations, failing to capture the complexities of real-world multi-user interactions. In this study, we assess the robustness of LLMs in multi-user DST while minimizing dataset construction costs. Inspired by recent advances in LLM-based data annotation, we extend an existing DST dataset by generating utterances of a second user based on speech act theory. Our methodology systematically incorporates a second user's utterances into conversations, enabling a controlled evaluation of LLMs in multi-user settings. Experimental results reveal a significant performance drop compared to single-user DST, highlighting the limitations of current LLMs in extracting and tracking dialogue states amidst multiple speakers. Our findings emphasize the need for future research to enhance LLMs for multi-user DST scenarios, paving the way for more realistic and robust DST models.
Similar Papers
Interpretable and Robust Dialogue State Tracking via Natural Language Summarization with LLMs
Computation and Language
Helps chatbots understand what you're saying better.
Joint Speech and Text Training for LLM-Based End-to-End Spoken Dialogue State Tracking
Computation and Language
Lets computers understand spoken words in new situations.
Neural Models and Language Model Prompting for the Multidimensional Evaluation of Open-Ended Conversations
Computation and Language
Helps AI chatbots talk better.