Explainable Multimodal Aspect-Based Sentiment Analysis with Dependency-guided Large Language Model
By: Zhongzheng Wang , Yuanhe Tian , Hongzhi Wang and more
Multimodal aspect-based sentiment analysis (MABSA) aims to identify aspect-level sentiments by jointly modeling textual and visual information, which is essential for fine-grained opinion understanding in social media. Existing approaches mainly rely on discriminative classification with complex multimodal fusion, yet lacking explicit sentiment explainability. In this paper, we reformulate MABSA as a generative and explainable task, proposing a unified framework that simultaneously predicts aspect-level sentiment and generates natural language explanations. Based on multimodal large language models (MLLMs), our approach employs a prompt-based generative paradigm, jointly producing sentiment and explanation. To further enhance aspect-oriented reasoning capabilities, we propose a dependency-syntax-guided sentiment cue strategy. This strategy prunes and textualizes the aspect-centered dependency syntax tree, guiding the model to distinguish different sentiment aspects and enhancing its explainability. To enable explainability, we use MLLMs to construct new datasets with sentiment explanations to fine-tune. Experiments show that our approach not only achieves consistent gains in sentiment classification accuracy, but also produces faithful, aspect-grounded explanations.
Similar Papers
Learning to Extract Cross-Domain Aspects and Understanding Sentiments Using Large Language Models
Computation and Language
Finds what people like or dislike about products.
Structured Prompting and LLM Ensembling for Multimodal Conversational Aspect-based Sentiment Analysis
Computation and Language
Helps computers understand feelings in talking.
Large-Scale Aspect-Based Sentiment Analysis with Reasoning-Infused LLMs
Computation and Language
Helps computers understand feelings in reviews better.