Adaptive Bitrate Quantization Scheme Without Codebook for Learned Image Compression

Abstract

We propose a generic approach to quantization without codebook in learned image compression called onehot max (OHM, Ω) quantization. It reorganizes the feature space resulting in an additional dimension, along which vector quantization yields one-hot vectors by comparing activations. Furthermore, we show how to integrate Ω quantization into a compression system with bitrate adaptation, i.e., full control over bitrate during inference. We perform experiments on both MNIST and Kodak and report on rate-distortion trade-offs comparing with the integer rounding reference. For low bitrates (< 0.4 bpp), our proposed quantizer yields better performance while exhibiting also other advantageous training and inference properties. Code is available at https://github.com/ifnspaml/OHMQ.

Publication
In Proc. of CVF/IEEE Conference on Computer Vision and Pattern Recognition - Workshops
Andreas Bär
Andreas Bär
PhD Student / Research Associate