Governance-Aware Hybrid Fine-Tuning for Multilingual Large Language Models
By: Haomin Qi , Chengbo Huang , Zihan Dai and more
Potential Business Impact:
Makes AI understand many languages better, even with less data.
We present a governance-aware hybrid fine-tuning framework for multilingual, low-resource adaptation of large language models. The core algorithm combines gradient-aligned low-rank updates with structured orthogonal transformations through layer-wise mixing and introduces unitary constraints in selected sub-layers to stabilize deep optimization. In tandem with lightweight, label-free data governance steps, including language identification, near-duplicate removal, and quality filtering, the framework targets accuracy, calibration, and cross-language parity under tight compute budgets. Across XNLI and FLORES, the hybrid approach delivers consistent gains over strong PEFT baselines while maintaining directional balance and improving probability calibration, as shown in Tables II and III. It is more resilient to lightweight orthographic variants, as shown in Table IV, and benefits additively from simple governance steps, as shown in Table V. Training footprint measurements indicate modest overhead and a favorable cost-quality frontier, as shown in Table VI and Figure 2. Together, these results show that hybrid and unitary PEFT provide a stable and accessible path to resource-efficient multilingual adaptation when paired with practical data governance.
Similar Papers
A Bayesian Hybrid Parameter-Efficient Fine-Tuning Method for Large Language Models
Machine Learning (CS)
Helps AI learn better from new information.
Enhancing Non-English Capabilities of English-Centric Large Language Models through Deep Supervision Fine-Tuning
Computation and Language
Makes computers understand and speak all languages better.
Optimizing Language Models for Grammatical Acceptability: A Comparative Study of Fine-Tuning Techniques
Computation and Language
Makes smart computer programs learn faster, cheaper.