Causal Inspired Multi Modal Recommendation
By: Jie Yang, Chenyang Gu, Zixuan Liu
Potential Business Impact:
Fixes online shopping picks by ignoring fake trends.
Multimodal recommender systems enhance personalized recommendations in e-commerce and online advertising by integrating visual, textual, and user-item interaction data. However, existing methods often overlook two critical biases: (i) modal confounding, where latent factors (e.g., brand style or product category) simultaneously drive multiple modalities and influence user preference, leading to spurious feature-preference associations; (ii) interaction bias, where genuine user preferences are mixed with noise from exposure effects and accidental clicks. To address these challenges, we propose a Causal-inspired multimodal Recommendation framework. Specifically, we introduce a dual-channel cross-modal diffusion module to identify hidden modal confounders, utilize back-door adjustment with hierarchical matching and vector-quantized codebooks to block confounding paths, and apply front-door adjustment combined with causal topology reconstruction to build a deconfounded causal subgraph. Extensive experiments on three real-world e-commerce datasets demonstrate that our method significantly outperforms state-of-the-art baselines while maintaining strong interpretability.
Similar Papers
Does Multimodality Improve Recommender Systems as Expected? A Critical Analysis and Future Directions
Information Retrieval
Makes online suggestions better by using different info.
Knowledge graph-based personalized multimodal recommendation fusion framework
Information Retrieval
Helps computers understand what you like better.
Semantic Item Graph Enhancement for Multimodal Recommendation
Information Retrieval
Helps online stores show you better stuff.