Rethinking Multimodal Sentiment Analysis: A High-Accuracy, Simplified Fusion Architecture
By: Nischal Mandal, Yang Li
Potential Business Impact:
Helps computers understand feelings from talking, seeing, and hearing.
Multimodal sentiment analysis, a pivotal task in affective computing, seeks to understand human emotions by integrating cues from language, audio, and visual signals. While many recent approaches leverage complex attention mechanisms and hierarchical architectures, we propose a lightweight, yet effective fusion-based deep learning model tailored for utterance-level emotion classification. Using the benchmark IEMOCAP dataset, which includes aligned text, audio-derived numeric features, and visual descriptors, we design a modality-specific encoder using fully connected layers followed by dropout regularization. The modality-specific representations are then fused using simple concatenation and passed through a dense fusion layer to capture cross-modal interactions. This streamlined architecture avoids computational overhead while preserving performance, achieving a classification accuracy of 92% across six emotion categories. Our approach demonstrates that with careful feature engineering and modular design, simpler fusion strategies can outperform or match more complex models, particularly in resource-constrained environments.
Similar Papers
Dynamic Multimodal Sentiment Analysis: Leveraging Cross-Modal Attention for Enabled Classification
Computation and Language
Helps computers understand feelings from voice, face, and words.
Multimodal Emotion Recognition and Sentiment Analysis in Multi-Party Conversation Contexts
CV and Pattern Recognition
Lets computers understand feelings from talking, faces, and videos.
Representation Decomposition for Learning Similarity and Contrastness Across Modalities for Affective Computing
Computation and Language
Helps computers understand feelings from pictures and words.