SentiMM: A Multimodal Multi-Agent Framework for Sentiment Analysis in Social Media
By: Xilai Xu , Zilin Zhao , Chengye Song and more
Potential Business Impact:
Helps computers understand feelings in pictures and words.
With the increasing prevalence of multimodal content on social media, sentiment analysis faces significant challenges in effectively processing heterogeneous data and recognizing multi-label emotions. Existing methods often lack effective cross-modal fusion and external knowledge integration. We propose SentiMM, a novel multi-agent framework designed to systematically address these challenges. SentiMM processes text and visual inputs through specialized agents, fuses multimodal features, enriches context via knowledge retrieval, and aggregates results for final sentiment classification. We also introduce SentiMMD, a large-scale multimodal dataset with seven fine-grained sentiment categories. Extensive experiments demonstrate that SentiMM achieves superior performance compared to state-of-the-art baselines, validating the effectiveness of our structured approach.
Similar Papers
Multimodal Emotion Recognition and Sentiment Analysis in Multi-Party Conversation Contexts
CV and Pattern Recognition
Lets computers understand feelings from talking, faces, and videos.
Do Multimodal LLMs See Sentiment?
CV and Pattern Recognition
Helps computers understand feelings in pictures.
Senti-iFusion: An Integrity-centered Hierarchical Fusion Framework for Multimodal Sentiment Analysis under Uncertain Modality Missingness
Human-Computer Interaction
Helps computers understand feelings even with missing info.