Sorry, you need to enable JavaScript to visit this website.

Interpretable Learned Image Compression: A Frequency Transform Decomposition Perspective

Citation Author(s):
Chuanmin Jia, Siwei Ma
Submitted by:
Yuefeng Zhang
Last updated:
3 March 2022 - 8:58am
Document Type:
Presentation Slides
Document Year:
Yuefeng Zhang
Paper Code:

Image compression is a key problem in this age of information explosion. With the help of machine learning, recent studies have shown that learning-based image compression methods tend to surpass traditional codecs. Image compression can be split into three steps: transform, quantization, and entropy estimation.
However, the transform step in traditional codecs lacks flexibility because of the strict mathematical premise while the transform in most learning-based codecs neglects its intrinsic interpretation. After observing compression degradation degree varies on different frequency bands, we propose an end-to-end compression model from the frequency perspective with a frequency-pyramid transform and a frequency-aware fusion module. Intuitively, we can infer that the low-frequency part contains the global structure while the high-frequency part gets finer details, satisfying the feature of human visual system (HVS). In our proposed model, independent probability estimation models are set for each frequency split. Extensive experiments are conducted to demonstrate that our model outperforms all traditional codecs (e.g., JPEG, JPEG2000, HEVC, and VVC) on MS-SSIM metric on both Kodak and CLIC2020 professional test datasets. Taking BPG-4:4:4 as the anchor, our proposed model achieves 11.6% BD-rate reduction under PSNR measurement, which is evaluated on the Kodak dataset.

1 user has voted: Yuefeng Zhang


Welcome to contact us if you have any questions & suggestions!