Score: 1

A Remarkably Efficient Paradigm to Multimodal Large Language Models for Sequential Recommendation

Published: November 8, 2025 | arXiv ID: 2511.05885v1

By: Qiyong Zhong , Jiajie Su , Ming Yang and more

Potential Business Impact:

Makes online shopping suggestions much faster.

Business Areas:
Semantic Search Internet Services

In this paper, we proposed Speeder, a remarkably efficient paradigm to multimodal large language models for sequential recommendation. Speeder introduces 3 key components: (1) Multimodal Representation Compression (MRC), which efficiently reduces redundancy in item descriptions; (2) Sequential Position Awareness Enhancement (SPAE), which strengthens the model's ability to capture complex sequential dependencies; (3) Modality-aware Progressive Optimization (MPO), which progressively integrates different modalities to improve the model's understanding and reduce cognitive biases. Through extensive experiments, Speeder demonstrates superior performance over baselines in terms of VHR@1 and computational efficiency. Specifically, Speeder achieved 250% of the training speed and 400% of the inference speed compared to the state-of-the-art MLLM-based SR models. Future work could focus on incorporating real-time feedback from real-world systems.

Country of Origin
πŸ‡ΈπŸ‡¬ πŸ‡¨πŸ‡³ China, Singapore

Page Count
12 pages

Category
Computer Science:
Information Retrieval