Personalized Product Search Ranking: A Multi-Task Learning Approach with Tabular and Non-Tabular Data
By: Lalitesh Morishetti , Abhay Kumar , Jonathan Scott and more
Potential Business Impact:
Finds you better stuff when you shop online.
In this paper, we present a novel model architecture for optimizing personalized product search ranking using a multi-task learning (MTL) framework. Our approach uniquely integrates tabular and non-tabular data, leveraging a pre-trained TinyBERT model for semantic embeddings and a novel sampling technique to capture diverse customer behaviors. We evaluate our model against several baselines, including XGBoost, TabNet, FT-Transformer, DCN-V2, and MMoE, focusing on their ability to handle mixed data types and optimize personalized ranking. Additionally, we propose a scalable relevance labeling mechanism based on click-through rates, click positions, and semantic similarity, offering an alternative to traditional human-annotated labels. Experimental results show that combining non-tabular data with advanced embedding techniques in multi-task learning paradigm significantly enhances model performance. Ablation studies further underscore the benefits of incorporating relevance labels, fine-tuning TinyBERT layers, and TinyBERT query-product embedding interactions. These results demonstrate the effectiveness of our approach in achieving improved personalized product search ranking.
Similar Papers
MultiTab: A Scalable Foundation for Multitask Learning on Tabular Data
Machine Learning (CS)
Helps computers learn many things from tables faster.
Tensorized Multi-Task Learning for Personalized Modeling of Heterogeneous Individuals with High-Dimensional Data
Machine Learning (CS)
Helps computers learn about different groups of people.
iLTM: Integrated Large Tabular Model
Machine Learning (CS)
Makes computers learn from data much better.