Progressive Learned Image Compression for Machine Perception
By: Jungwoo Kim, Jun-Hyuk Kim, Jong-Seok Lee
Recent advances in learned image codecs have been extended from human perception toward machine perception. However, progressive image compression with fine granular scalability (FGS)-which enables decoding a single bitstream at multiple quality levels-remains unexplored for machine-oriented codecs. In this work, we propose a novel progressive learned image compression codec for machine perception, PICM-Net, based on trit-plane coding. By analyzing the difference between human- and machine-oriented rate-distortion priorities, we systematically examine the latent prioritization strategies in terms of machine-oriented codecs. To further enhance real-world adaptability, we design an adaptive decoding controller, which dynamically determines the necessary decoding level during inference time to maintain the desired confidence of downstream machine prediction. Extensive experiments demonstrate that our approach enables efficient and adaptive progressive transmission while maintaining high performance in the downstream classification task, establishing a new paradigm for machine-aware progressive image compression.
Similar Papers
Unified Coding for Both Human Perception and Generalized Machine Analytics with CLIP Supervision
CV and Pattern Recognition
Makes pictures useful for people and computers.
Machines Serve Human: A Novel Variable Human-machine Collaborative Compression Framework
CV and Pattern Recognition
Makes pictures smaller for people and computers.
Diff-ICMH: Harmonizing Machine and Human Vision in Image Compression with Generative Prior
CV and Pattern Recognition
Makes pictures look good and computers understand them.