Modeling shopper interest broadness with entropy-driven dialogue policy in the context of arbitrarily large product catalogs
By: Firas Jarboui, Issa Memari
Potential Business Impact:
Helps online shoppers find what they want faster.
Conversational recommender systems promise rich interactions for e-commerce, but balancing exploration (clarifying user needs) and exploitation (making recommendations) remains challenging, especially when deploying large language models (LLMs) with vast product catalogs. We address this challenge by modeling the breadth of user interest via the entropy of retrieval score distributions. Our method uses a neural retriever to fetch relevant items for a user query and computes the entropy of the re-ranked scores to dynamically route the dialogue policy: low-entropy (specific) queries trigger direct recommendations, whereas high-entropy (ambiguous) queries prompt exploratory questions. This simple yet effective strategy allows an LLM-driven agent to remain aware of an arbitrarily large catalog in real-time without bloating its context window.
Similar Papers
Research on E-Commerce Long-Tail Product Recommendation Mechanism Based on Large-Scale Language Models
Information Retrieval
Helps online stores show you more unique items.
Enhancing Serendipity Recommendation System by Constructing Dynamic User Knowledge Graphs with Large Language Models
Information Retrieval
Finds you cooler stuff you might like.
LLMs as Sparse Retrievers:A Framework for First-Stage Product Search
Information Retrieval
Helps online stores find products faster and better.