SpeechIQ: Speech Intelligence Quotient Across Cognitive Levels in Voice Understanding Large Language Models
By: Zhen Wan , Chao-Han Huck Yang , Yahan Yu and more
Potential Business Impact:
Tests how well computers understand your voice.
We introduce Speech-based Intelligence Quotient (SIQ) as a new form of human cognition-inspired evaluation pipeline for voice understanding large language models, LLM Voice, designed to assess their voice understanding ability. Moving beyond popular voice understanding metrics such as word error rate (WER), SIQ examines LLM Voice across three cognitive levels motivated by Bloom's Taxonomy: (1) Remembering (i.e., WER for verbatim accuracy); (2) Understanding (i.e., similarity of LLM's interpretations); and (3) Application (i.e., QA accuracy for simulating downstream tasks). We demonstrate that SIQ not only quantifies voice understanding abilities but also provides unified comparisons between cascaded methods (e.g., ASR LLM) and end-to-end models, identifies annotation errors in existing benchmarks, and detects hallucinations in LLM Voice. Our framework represents a first-of-its-kind intelligence examination that bridges cognitive principles with voice-oriented benchmarks, while exposing overlooked challenges in multi-modal training.
Similar Papers
SpeechIQ: Speech-Agentic Intelligence Quotient Across Cognitive Levels in Voice Understanding by Large Language Models
Computation and Language
Tests how well computers understand spoken words.
Artificial Intelligence Quotient (AIQ): A Novel Framework for Measuring Human-AI Collaborative Intelligence
Human-Computer Interaction
Measures how well people work with AI.
SpeechQualityLLM: LLM-Based Multimodal Assessment of Speech Quality
Sound
Lets computers explain why audio sounds bad.