AudioEval: Automatic Dual-Perspective and Multi-Dimensional Evaluation of Text-to-Audio-Generation
By: Hui Wang , Jinghua Zhao , Cheng Liu and more
Potential Business Impact:
Helps computers judge how good spoken words sound.
Text-to-audio (TTA) is rapidly advancing, with broad potential in virtual reality, accessibility, and creative media. However, evaluating TTA quality remains difficult: human ratings are costly and limited, while existing objective metrics capture only partial aspects of perceptual quality. To address this gap, we introduce AudioEval, the first large-scale TTA evaluation dataset, containing 4,200 audio samples from 24 systems with 126,000 ratings across five perceptual dimensions, annotated by both experts and non-experts. Based on this resource, we propose Qwen-DisQA, a multimodal scoring model that jointly processes text prompts and generated audio to predict human-like quality ratings. Experiments show its effectiveness in providing reliable and scalable evaluation. The dataset will be made publicly available to accelerate future research.
Similar Papers
TTA-Bench: A Comprehensive Benchmark for Evaluating Text-to-Audio Models
Sound
Tests AI that makes sounds from words.
Improving Perceptual Audio Aesthetic Assessment via Triplet Loss and Self-Supervised Embeddings
Audio and Speech Processing
Rates how good computer-made sounds are.
Audio-Based Crowd-Sourced Evaluation of Machine Translation Quality
Computation and Language
Listens to translated speech to check its quality.