Three Gaps for Quantisation in Learned Image Compression

Shi Pan, Chris Finlay, Chri Besenbruch, William Knottenbelt; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2021, pp. 720-726

Abstract


Learned lossy image compression has demonstrated impressive progress via end-to-end neural network training. However, this end-to-end training belies the fact that lossy compression is inherently not differentiable, due to the necessity of quantisation. To overcome this difficulty in training, researchers have used various approximations to the quantisation step. However, little work has studied the mechanism of quantisation approximation itself. We address this issue, identifying three gaps arising in the quantisation approximation problem. These gaps are visualised, and show the effect of applying different quantisation approximation methods. Following this analysis, we propose a Soft-STE quantisation approximation method, which closes these gaps and demonstrates better performance than other quantisation approaches on the Kodak dataset.

Related Material


[pdf]
[bibtex]
@InProceedings{Pan_2021_CVPR, author = {Pan, Shi and Finlay, Chris and Besenbruch, Chri and Knottenbelt, William}, title = {Three Gaps for Quantisation in Learned Image Compression}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = {June}, year = {2021}, pages = {720-726} }