Evaluating Recabilities of Foundation Models: A Multi-Domain, Multi-Dataset Benchmark
By: Qijiong Liu , Jieming Zhu , Yingxin Lai and more
Potential Business Impact:
Tests AI to recommend things better.
Comprehensive evaluation of the recommendation capabilities of existing foundation models across diverse datasets and domains is essential for advancing the development of recommendation foundation models. In this study, we introduce RecBench-MD, a novel and comprehensive benchmark designed to assess the recommendation abilities of foundation models from a zero-resource, multi-dataset, and multi-domain perspective. Through extensive evaluations of 19 foundation models across 15 datasets spanning 10 diverse domains -- including e-commerce, entertainment, and social media -- we identify key characteristics of these models in recommendation tasks. Our findings suggest that in-domain fine-tuning achieves optimal performance, while cross-dataset transfer learning provides effective practical support for new recommendation scenarios. Additionally, we observe that multi-domain training significantly enhances the adaptability of foundation models. All code and data have been publicly released to facilitate future research.
Similar Papers
Unifying Inductive, Cross-Domain, and Multimodal Learning for Robust and Generalizable Recommendation
Information Retrieval
Recommends better things by learning from many sources.
RecBase: Generative Foundation Model Pretraining for Zero-Shot Recommendation
Information Retrieval
Recommends items better across different apps.
Generative Representational Learning of Foundation Models for Recommendation
Information Retrieval
Helps computers recommend things better for everyone.