Mitigating Strategy Preference Bias in Emotional Support Conversation via Uncertainty Estimations
By: Yougen Zhou , Qin Chen , Ningning Zhou and more
Potential Business Impact:
Helps computers give better emotional support talks.
Emotional support conversation (ESC) aims to alleviate distress through empathetic dialogue, yet large language models (LLMs) face persistent challenges in delivering effective ESC due to low accuracy in strategy planning. Moreover, there is a considerable preference bias towards specific strategies. Prior methods using fine-tuned strategy planners have shown potential in reducing such bias, while the underlying causes of the preference bias in LLMs have not well been studied. To address these issues, we first reveal the fundamental causes of the bias by identifying the knowledge boundaries of LLMs in strategy planning. Then, we propose an approach to mitigate the bias by reinforcement learning with a dual reward function, which optimizes strategy planning via both accuracy and entropy-based confidence for each region according to the knowledge boundaries. Experiments on the ESCov and ExTES datasets with multiple LLM backbones show that our approach outperforms the baselines, confirming the effectiveness of our approach.
Similar Papers
Convert Language Model into a Value-based Strategic Planner
Computation and Language
Helps computers give better emotional support talks.
Towards Open-Ended Emotional Support Conversations in LLMs via Reinforcement Learning with Future-Oriented Rewards
Artificial Intelligence
Helps computers give better emotional support.
Emotional Support with LLM-based Empathetic Dialogue Generation
Artificial Intelligence
Helps computers give comforting and helpful advice.