Score: 1

Governance-Aware Hybrid Fine-Tuning for Multilingual Large Language Models

Published: December 19, 2025 | arXiv ID: 2512.17344v1

By: Haomin Qi , Chengbo Huang , Zihan Dai and more

Potential Business Impact:

Makes AI understand many languages better, even with less data.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

We present a governance-aware hybrid fine-tuning framework for multilingual, low-resource adaptation of large language models. The core algorithm combines gradient-aligned low-rank updates with structured orthogonal transformations through layer-wise mixing and introduces unitary constraints in selected sub-layers to stabilize deep optimization. In tandem with lightweight, label-free data governance steps, including language identification, near-duplicate removal, and quality filtering, the framework targets accuracy, calibration, and cross-language parity under tight compute budgets. Across XNLI and FLORES, the hybrid approach delivers consistent gains over strong PEFT baselines while maintaining directional balance and improving probability calibration, as shown in Tables II and III. It is more resilient to lightweight orthographic variants, as shown in Table IV, and benefits additively from simple governance steps, as shown in Table V. Training footprint measurements indicate modest overhead and a favorable cost-quality frontier, as shown in Table VI and Figure 2. Together, these results show that hybrid and unitary PEFT provide a stable and accessible path to resource-efficient multilingual adaptation when paired with practical data governance.

Country of Origin
πŸ‡ΊπŸ‡Έ πŸ‡©πŸ‡° Denmark, United States

Page Count
11 pages

Category
Computer Science:
Computation and Language