Generalized Fisher-Weighted SVD: Scalable Kronecker-Factored Fisher Approximation for Compressing Large Language Models
By: Viktoriia Chekalina , Daniil Moskovskiy , Daria Cherniuk and more
Potential Business Impact:
Makes big computer brains smaller, smarter.
The Fisher information is a fundamental concept for characterizing the sensitivity of parameters in neural networks. However, leveraging the full observed Fisher information is too expensive for large models, so most methods rely on simple diagonal approximations. While efficient, this approach ignores parameter correlations, often resulting in reduced performance on downstream tasks. In this work, we mitigate these limitations and propose Generalized Fisher-Weighted SVD (GFWSVD), a post-training LLM compression technique that accounts for both diagonal and off-diagonal elements of the Fisher information matrix, providing a more accurate reflection of parameter importance. To make the method tractable, we introduce a scalable adaptation of the Kronecker-factored approximation algorithm for the observed Fisher information. We demonstrate the effectiveness of our method on LLM compression, showing improvements over existing compression baselines. For example, at a 20 compression rate on the MMLU benchmark, our method outperforms FWSVD, which is based on a diagonal approximation of the Fisher information, by 5 percent, SVD-LLM by 3 percent, and ASVD by 6 percent compression rate.
Similar Papers
Globally optimized SVD compression of LLMs via Fermi-function-based rank selection and gauge fixing
Machine Learning (CS)
Makes big computer brains smaller and faster.
Beyond Variance: Knowledge-Aware LLM Compression via Fisher-Aligned Subspace Diagnostics
Machine Learning (CS)
Makes big AI models fit on small devices.
AdaSVD: Adaptive Singular Value Decomposition for Large Language Models
CV and Pattern Recognition
Makes smart computer brains smaller, work better.