S2SBench: A Benchmark for Quantifying Intelligence Degradation in Speech-to-Speech Large Language Models
By: Yuanbo Fang , Haoze Sun , Jun Liu and more
Potential Business Impact:
Makes talking computers think better.
End-to-end speech large language models ((LLMs)) extend the capabilities of text-based models to directly process and generate audio tokens. However, this often leads to a decline in reasoning and generation performance compared to text input, a phenomenon referred to as intelligence degradation. To systematically evaluate this gap, we propose S2SBench, a benchmark designed to quantify performance degradation in Speech LLMs. It includes diagnostic datasets targeting sentence continuation and commonsense reasoning under audio input. We further introduce a pairwise evaluation protocol based on perplexity differences between plausible and implausible samples to measure degradation relative to text input. We apply S2SBench to analyze the training process of Baichuan-Audio, which further demonstrates the benchmark's effectiveness. All datasets and evaluation code are available at https://github.com/undobug/S2SBench.
Similar Papers
SI-Bench: Benchmarking Social Intelligence of Large Language Models in Human-to-Human Conversations
Computation and Language
Tests how well AI understands people talking.
URO-Bench: Towards Comprehensive Evaluation for End-to-End Spoken Dialogue Models
Computation and Language
Tests talking computers on understanding, thinking, and speaking.
S2S-Arena, Evaluating Speech2Speech Protocols on Instruction Following with Paralinguistic Information
Computation and Language
Tests how well talking computers understand and speak.