Evaluating Contrastive Feedback for Effective User Simulations
By: Andreas Konstantin Kruff, Timo Breuer, Philipp Schaer
Potential Business Impact:
Teaches computers to act like real people searching.
The use of Large Language Models (LLMs) for simulating user behavior in the domain of Interactive Information Retrieval has recently gained significant popularity. However, their application and capabilities remain highly debated and understudied. This study explores whether the underlying principles of contrastive training techniques, which have been effective for fine-tuning LLMs, can also be applied beneficially in the area of prompt engineering for user simulations. Previous research has shown that LLMs possess comprehensive world knowledge, which can be leveraged to provide accurate estimates of relevant documents. This study attempts to simulate a knowledge state by enhancing the model with additional implicit contextual information gained during the simulation. This approach enables the model to refine the scope of desired documents further. The primary objective of this study is to analyze how different modalities of contextual information influence the effectiveness of user simulations. Various user configurations were tested, where models are provided with summaries of already judged relevant, irrelevant, or both types of documents in a contrastive manner. The focus of this study is the assessment of the impact of the prompting techniques on the simulated user agent performance. We hereby lay the foundations for leveraging LLMs as part of more realistic simulated users.
Similar Papers
An Analysis of Large Language Models for Simulating User Responses in Surveys
Computation and Language
Helps computers understand many different opinions.
Understanding Learner-LLM Chatbot Interactions and the Impact of Prompting Guidelines
Human-Computer Interaction
Teaches people to ask AI better questions.
Simulating LLM-to-LLM Tutoring for Multilingual Math Feedback
Computation and Language
Teaches math better in any language.