AFD-SLU: Adaptive Feature Distillation for Spoken Language Understanding
By: Yan Xie , Yibo Cui , Liang Xie and more
Potential Business Impact:
Helps computers understand what you say better.
Spoken Language Understanding (SLU) is a core component of conversational systems, enabling machines to interpret user utterances. Despite its importance, developing effective SLU systems remains challenging due to the scarcity of labeled training data and the computational burden of deploying Large Language Models (LLMs) in real-world applications. To further alleviate these issues, we propose an Adaptive Feature Distillation framework that transfers rich semantic representations from a General Text Embeddings (GTE)-based teacher model to a lightweight student model. Our method introduces a dynamic adapter equipped with a Residual Projection Neural Network (RPNN) to align heterogeneous feature spaces, and a Dynamic Distillation Coefficient (DDC) that adaptively modulates the distillation strength based on real-time feedback from intent and slot prediction performance. Experiments on the Chinese profile-based ProSLU benchmark demonstrate that AFD-SLU achieves state-of-the-art results, with 95.67% intent accuracy, 92.02% slot F1 score, and 85.50% overall accuracy.
Similar Papers
DIFFA: Large Language Diffusion Models Can Listen and Understand
Sound
Lets computers understand spoken words better.
LLMCARE: Alzheimer's Detection via Transformer Models Enhanced by LLM-Generated Synthetic Data
Computation and Language
Finds early signs of memory loss in voices.
Adaptive Federated Distillation for Multi-Domain Non-IID Textual Data
Computation and Language
Helps AI learn from many different kinds of text.