An Efficient Compression of Deep Neural Network Checkpoints Based on Prediction and Context Modeling
By: Yuriy Kim, Evgeny Belyaev
Potential Business Impact:
Shrinks computer learning files to save space.
This paper is dedicated to an efficient compression of weights and optimizer states (called checkpoints) obtained at different stages during a neural network training process. First, we propose a prediction-based compression approach, where values from the previously saved checkpoint are used for context modeling in arithmetic coding. Second, in order to enhance the compression performance, we also propose to apply pruning and quantization of the checkpoint values. Experimental results show that our approach achieves substantial bit size reduction, while enabling near-lossless training recovery from restored checkpoints, preserving the model's performance and making it suitable for storage-limited environments.
Similar Papers
Coding for Computation: Efficient Compression of Neural Networks for Reconfigurable Hardware
Machine Learning (CS)
Makes smart computer programs run much faster.
Optimizing Deep Neural Networks using Safety-Guided Self Compression
Machine Learning (CS)
Shrinks smart computer programs without losing smarts.
Lossless Compression of Neural Network Components: Weights, Checkpoints, and K/V Caches in Low-Precision Formats
Machine Learning (CS)
Shrinks AI models to save space and speed.