Efficient Speech Translation through Model Compression and Knowledge Distillation
By: Yasmin Moslem
Potential Business Impact:
Makes translation apps smaller and faster.
Efficient deployment of large audio-language models for speech translation remains challenging due to their significant computational requirements. In this paper, we address this challenge through our system submissions to the "Model Compression" track at the International Conference on Spoken Language Translation (IWSLT 2025). We experiment with a combination of approaches including iterative layer pruning based on layer importance evaluation, low-rank adaptation with 4-bit quantization (QLoRA), and knowledge distillation. In our experiments, we use Qwen2-Audio-7B-Instruct for speech translation into German and Chinese. Our pruned (student) models achieve up to a 50% reduction in both model parameters and storage footprint, while retaining 97-100% of the translation quality of the in-domain (teacher) models.
Similar Papers
Exploring the Limits of Model Compression in LLMs: A Knowledge Distillation Study on QA Tasks
Computation and Language
Makes smart computer programs smaller and faster.
On Multilingual Encoder Language Model Compression for Low-Resource Languages
Computation and Language
Makes computer language programs much smaller.
Efficient AI in Practice: Training and Deployment of Efficient LLMs for Industry Applications
Information Retrieval
Makes small AI models as smart as big ones.