On Multilingual Encoder Language Model Compression for Low-Resource Languages
By: Daniil Gurgurov , Michal Gregor , Josef van Genabith and more
Potential Business Impact:
Makes computer language programs much smaller.
In this paper, we combine two-step knowledge distillation, structured pruning, truncation, and vocabulary trimming for extremely compressing multilingual encoder-only language models for low-resource languages. Our novel approach systematically combines existing techniques and takes them to the extreme, reducing layer depth, feed-forward hidden size, and intermediate layer embedding size to create significantly smaller monolingual models while retaining essential language-specific knowledge. We achieve compression rates of up to 92% with only a marginal performance drop of 2-10% in four downstream tasks, including sentiment analysis, topic classification, named entity recognition, and part-of-speech tagging, across three low-resource languages. Notably, the performance degradation correlates with the amount of language-specific data in the teacher model, with larger datasets resulting in smaller performance losses. Additionally, we conduct extensive ablation studies to identify best practices for multilingual model compression using these techniques.
Similar Papers
Efficient Speech Translation through Model Compression and Knowledge Distillation
Computation and Language
Makes translation apps smaller and faster.
Towards Inclusive NLP: Assessing Compressed Multilingual Transformers across Diverse Language Benchmarks
Computation and Language
Makes AI understand many languages, even rare ones.
Optimizing LLMs for Resource-Constrained Environments: A Survey of Model Compression Techniques
Machine Learning (CS)
Shrinks big AI so phones can use it.