The TEA-ASLP System for Multilingual Conversational Speech Recognition and Speech Diarization in MLC-SLM 2025 Challenge
By: Hongfei Xue , Kaixun Huang , Zhikai Zhou and more
Potential Business Impact:
Makes computers understand many languages spoken.
This paper presents the TEA-ASLP's system submitted to the MLC-SLM 2025 Challenge, addressing multilingual conversational automatic speech recognition (ASR) in Task I and speech diarization ASR in Task II. For Task I, we enhance Ideal-LLM model by integrating known language identification and a multilingual MOE LoRA structure, along with using CTC-predicted tokens as prompts to improve autoregressive generation. The model is trained on approximately 180k hours of multilingual ASR data. In Task II, we replace the baseline English-Chinese speaker diarization model with a more suitable English-only version. Our approach achieves a 30.8% reduction in word error rate (WER) compared to the baseline speech language model, resulting in a final WER of 9.60% in Task I and a time-constrained minimum-permutation WER of 17.49% in Task II, earning first and second place in the respective challenge tasks.
Similar Papers
Transsion Multilingual Speech Recognition System for MLC-SLM 2025 Challenge
Audio and Speech Processing
Lets computers understand many languages spoken.
The Eloquence team submission for task 1 of MLC-SLM challenge
Sound
Helps computers understand many languages spoken.
NTU Speechlab LLM-Based Multilingual ASR System for Interspeech MLC-SLM Challenge 2025
Computation and Language
Makes computers understand many languages better.