Sorry, you need to enable JavaScript to visit this website.

Multi-Density Convolutional Neural Network for In-Loop Filter in Video Coding

Citation Author(s):
Changyue Ma, Ru-Ling Liao, Yan Ye
Submitted by:
Zhao Wang
Last updated:
28 February 2021 - 3:57am
Document Type:
Presentation Slides
Document Year:
Presenters Name:
Zhao Wang
Paper Code:



As the latest video coding standard, Versatile Video Coding (VVC) achieves up to 40% Bjøntegaard delta bit-rate (BD-rate) reduction compared with High Efficiency Video Coding (HEVC). Recently, Convolutional Neural Network (CNN) has attracted tremendous attention and shows great potential in video coding. In this paper, we design a Multi-Density Convolutional Neural Network (MDCNN) as an integrated in-loop filter to improve the quality of the reconstructed frames. The core of our approach is the multi-density block (MDB), which contains two branches: (a) the basic branch maintaining full resolution for capturing spatially-precise representations, (b) the density branch for learning rich spatial correlation with larger receptive field through down-sampling and up-sampling. The feature maps of two branches will be fused into one stream repeatedly. Benefiting from this architecture, spatially-precise representations and density correlations in larger receptive field are utilized to improve the model performance and promote the model’s robustness to different input resolutions. Experimental results show that, in terms of BD-rate savings for the (Y, U, V) components compared to the state-of-the-art VVC standard, the proposed MDCNN filter can achieve (5.06%, 13.86%, 13.76%) and (4.36%, 10.85%, 10.91%) coding gain for the Random Access (RA) configuration and the All Intra (AI) configuration, respectively.

0 users have voted:

Dataset Files