MIDG: Mixture of Invariant Experts with knowledge injection for Domain Generalization in Multimodal Sentiment Analysis
By: Yangle Li, Danli Luo, Haifeng Hu
Potential Business Impact:
Helps computers understand feelings from text and pictures.
Existing methods in domain generalization for Multimodal Sentiment Analysis (MSA) often overlook inter-modal synergies during invariant features extraction, which prevents the accurate capture of the rich semantic information within multimodal data. Additionally, while knowledge injection techniques have been explored in MSA, they often suffer from fragmented cross-modal knowledge, overlooking specific representations that exist beyond the confines of unimodal. To address these limitations, we propose a novel MSA framework designed for domain generalization. Firstly, the framework incorporates a Mixture of Invariant Experts model to extract domain-invariant features, thereby enhancing the model's capacity to learn synergistic relationships between modalities. Secondly, we design a Cross-Modal Adapter to augment the semantic richness of multimodal representations through cross-modal knowledge injection. Extensive domain experiments conducted on three datasets demonstrate that the proposed MIDG achieves superior performance.
Similar Papers
FINE: Factorized multimodal sentiment analysis via mutual INformation Estimation
Multimedia
Helps computers understand feelings from text and pictures.
Improving Multimodal Sentiment Analysis via Modality Optimization and Dynamic Primary Modality Selection
CV and Pattern Recognition
Makes computers understand feelings from videos better.
Graph-based Interaction Augmentation Network for Robust Multimodal Sentiment Analysis
Multimedia
Helps computers understand feelings from messy videos.