Improving Latent Quantization of Learned Image Compression with Gradient Scaling

Heming Sun*, Lu Yu, Jiro Katto

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Learned image compression (LIC) has shown its superior compression ability. Quantization is an inevitable stage to generate quantized latent for the entropy coding. To solve the non-differentiable problem of quantization in the training phase, many differentiable approximated quantization methods have been proposed. However, the derivative of quantized latent to non-quantized latent are set as one in most of the previous methods. As a result, the quantization error between non-quantized and quantized latent is not taken into consideration in the gradient descent. To address this issue, we exploit the gradient scaling method to scale the gradient of non-quantized latent in the back-propagation. The experimental results show that we can outperform the recent LIC quantization methods.

Original languageEnglish
Title of host publication2022 IEEE International Conference on Visual Communications and Image Processing, VCIP 2022
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781665475921
DOIs
Publication statusPublished - 2022
Event2022 IEEE International Conference on Visual Communications and Image Processing, VCIP 2022 - Suzhou, China
Duration: 2022 Dec 132022 Dec 16

Publication series

Name2022 IEEE International Conference on Visual Communications and Image Processing, VCIP 2022

Conference

Conference2022 IEEE International Conference on Visual Communications and Image Processing, VCIP 2022
Country/TerritoryChina
CitySuzhou
Period22/12/1322/12/16

Keywords

  • Gradient scaling
  • Learned image compression
  • Quantization

ASJC Scopus subject areas

  • Computer Networks and Communications
  • Computer Science Applications
  • Computer Vision and Pattern Recognition
  • Signal Processing

Fingerprint

Dive into the research topics of 'Improving Latent Quantization of Learned Image Compression with Gradient Scaling'. Together they form a unique fingerprint.

Cite this