Score: 0

Improving Noise Robust Audio-Visual Speech Recognition via Router-Gated Cross-Modal Feature Fusion

Published: August 26, 2025 | arXiv ID: 2508.18734v1

By: DongHoon Lim , YoungChae Kim , Dong-Hyun Kim and more

Potential Business Impact:

Helps computers understand speech better in noisy places.

Business Areas:
Speech Recognition Data and Analytics, Software

Robust audio-visual speech recognition (AVSR) in noisy environments remains challenging, as existing systems struggle to estimate audio reliability and dynamically adjust modality reliance. We propose router-gated cross-modal feature fusion, a novel AVSR framework that adaptively reweights audio and visual features based on token-level acoustic corruption scores. Using an audio-visual feature fusion-based router, our method down-weights unreliable audio tokens and reinforces visual cues through gated cross-attention in each decoder layer. This enables the model to pivot toward the visual modality when audio quality deteriorates. Experiments on LRS3 demonstrate that our approach achieves an 16.51-42.67% relative reduction in word error rate compared to AV-HuBERT. Ablation studies confirm that both the router and gating mechanism contribute to improved robustness under real-world acoustic noise.

Country of Origin
🇰🇷 Korea, Republic of

Page Count
7 pages

Category
Computer Science:
CV and Pattern Recognition