Efficiency and Effectiveness of SPLADE Models on Billion-Scale Web Document Title
By: Taeryun Won , Tae Kwan Lee , Hiun Kim and more
Potential Business Impact:
Finds web pages faster and better.
This paper presents a comprehensive comparison of BM25, SPLADE, and Expanded-SPLADE models in the context of large-scale web document retrieval. We evaluate the effectiveness and efficiency of these models on datasets spanning from tens of millions to billions of web document titles. SPLADE and Expanded-SPLADE, which utilize sparse lexical representations, demonstrate superior retrieval performance compared to BM25, especially for complex queries. However, these models incur higher computational costs. We introduce pruning strategies, including document-centric pruning and top-k query term selection, boolean query with term threshold to mitigate these costs and improve the models' efficiency without significantly sacrificing retrieval performance. The results show that Expanded-SPLADE strikes the best balance between effectiveness and efficiency, particularly when handling large datasets. Our findings offer valuable insights for deploying sparse retrieval models in large-scale search engines.
Similar Papers
The Role of Vocabularies in Learning Sparse Representations for Ranking
Information Retrieval
Helps computers find information faster and better.
Semantic Search for Information Retrieval
Information Retrieval
Helps computers find information by understanding meaning.
Practical Code RAG at Scale: Task-Aware Retrieval Design Choices under Compute Budgets
Machine Learning (CS)
Helps computers write better code faster.