Towards LLM-Empowered Fine-Grained Speech Descriptors for Explainable Emotion Recognition
By: Youjun Chen , Xurong Xie , Haoning Xu and more
Potential Business Impact:
Helps computers understand how you feel when you talk.
This paper presents a novel end-to-end LLM-empowered explainable speech emotion recognition (SER) approach. Fine-grained speech emotion descriptor (SED) features, e.g., pitch, tone and emphasis, are disentangled from HuBERT SSL representations via alternating LLM fine-tuning to joint SER-SED prediction and ASR tasks. VAE compressed HuBERT features derived via Information Bottleneck (IB) are used to adjust feature granularity. Experiments on the IEMOCAP and MELD benchmarks demonstrate that our approach consistently outperforms comparable LLaMA-based SER baselines, including those using either (a) alternating multi-task fine-tuning alone or (b) feature disentanglement only. Statistically significant increase of SER unweighted accuracy by up to 4.0% and 3.7% absolute (5.4% and 6.6% relative) are obtained. More importantly, emotion descriptors offer further explainability for SER.
Similar Papers
MSF-SER: Enriching Acoustic Modeling with Multi-Granularity Semantics for Speech Emotion Recognition
Sound
Helps computers understand feelings in voices better.
Toward Efficient Speech Emotion Recognition via Spectral Learning and Attention
Sound
Helps computers understand your feelings from your voice.
Mobile-Efficient Speech Emotion Recognition Using DistilHuBERT: A Cross-Corpus Validation Study
Sound
Makes phones understand your feelings from your voice.