Abstract

As the upcoming video coding standard, Versatile Video Coding (i.e., VVC) achieves up to 30% Bjontegaard delta bit-rate (BD-rate) reduction compared with High Efficiency Video Coding (H.265/HEVC). To eliminate or alleviate different kinds of compression artifacts like blocking, ringing, blurring and contouring effects, three in-loop filters, i.e. de-blocking filter (DBF), sample adaptive offset (SAO) and adaptive loop filter (ALF), have been involved in VVC. Recently, Convolutional Neural Network (CNN) has attracted tremendous attention and shows great potential in many tasks in image processing. In this work, we design a CNN-based in-loop filter as an integrated single-model solution which is adaptive to almost any scenarios in video coding. An architecture named as ADCNN (i.e., Attention based Dual-scale CNN) with an attention based processing block is proposed to reduce artifacts of I frames and B frames, which take advantage of informative priors such as the quantization parameter (QP) and partitioning information. Different from existing CNN-based filtering methods, which are mainly designed for the luma component and may need to train different models for different QPs, the proposed filter is adapted to different QPs and different frame types, and all the components (i.e., both luma and chroma) are processed simultaneously with feature exchange and fusion between components for information supplementary. Experimental results show that the proposed ADCNN filter can achieve 6.54%, 13.27%, 15.72% BD-rate savings for Y, U, V respectively under the all intra configuration and 2.81%, 7.86%, 8.60% BD-rate savings under the random access configuration. It can be used to replace all the conventional in-loop filters and also outperforms them without increase in encoding time.

Highlights

  • In recent years, the video business has rapidly developed while the demand for high resolution and high definition has continuously improved

  • DIV2K dataset [19] which is consist of 900 2K resolution PNG pictures (800 images for training, and 100 images for validation) is employed to derive training and validation data through versatile video coding (VVC) reference software VTM-4.0 [22] under the All Intra configuration using quantization parameter (QP) values of {22, 27, 32, 37}

  • In this work, a Convolutional Neural Network (CNN)-based in-loop filter is proposed as an integrated solution to replace all the conventional filters in video coding

Read more

Summary

INTRODUCTION

The video business has rapidly developed while the demand for high resolution and high definition has continuously improved. Inspired by the Wiener filter theory, Adaptive Loop Filter (ALF) [7] trains filter coefficients to minimize the distortion between reconstructed frames and original frames, where the coefficients will be transmitted to decoder These in-loop filters are stacked to alleviate different kinds of distortions and to improve visual quality while saving coding bit-rate. The proposed model can reduce different kinds of artifacts such as blocking, ringing and fuzzy distortion. As the result, it can replace all the current in-loop filters in VVC.

RELATED WORK
LOSS FUNCTION
CTU ADAPTIVE CONTROL
Findings
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call