Validating Search Query Simulations: A Taxonomy of Measures
By: Andreas Konstantin Kruff, Nolwenn Bernard, Philipp Schaer
Potential Business Impact:
Helps computers pretend to search like people.
Assessing the validity of user simulators when used for the evaluation of information retrieval systems remains an open question, constraining their effective use and the reliability of simulation-based results. To address this issue, we conduct a comprehensive literature review with a particular focus on methods for the validation of simulated user queries with regard to real queries. Based on the review, we develop a taxonomy that structures the current landscape of available measures. We empirically corroborate the taxonomy by analyzing the relationships between the different measures applied to four different datasets representing diverse search scenarios. Finally, we provide concrete recommendations on which measures or combinations of measures should be considered when validating user simulation in different contexts. Furthermore, we release a dedicated library with the most commonly used measures to facilitate future research.
Similar Papers
Sim4IA-Bench: A User Simulation Benchmark Suite for Next Query and Utterance Prediction
Information Retrieval
Tests if computer searches act like real people.
Experimental Evaluation of Dynamic Topic Modeling Algorithms
Information Retrieval
Tracks how online topics change over time.
Enhancing Information Retrieval in Digital Libraries through Unit Harmonisation in Scholarly Knowledge Graphs
Digital Libraries
Finds and compares science data across studies.