Abstract

Lossy video compression achieves coding gains at the expense of the quality loss of the decoded images. Owing to the success of deep learning techniques, especially convolutional neural networks (CNNs), many compression artifacts reduction (CAR) techniques have been used to significantly improve the quality of decoded images by applying CNNs which are trained to predict the original artifact-free images from the decoded images. Most existing video compression standards control the compression ratio using a quantization parameter (QP), so the quality of the decoded images is strongly QP-dependent. Training individual CNNs for predetermined QPs is one of the common approaches to dealing with different levels of compression artifacts. However, compression artifacts are also dependent on the local characteristics of an image. Therefore, a CNN trained for specific QP cannot fully remove the compression artifacts of all images, even those encoded using the same QP. In this paper, we introduce a pixel-precise network selection network (PNSNet). From multiple reconstructed images obtained using multiple QP-specific CAR networks, PNSNet is trained to find the best CAR network for each pixel. The output of PNSNet is then used as an explicit spatial attention channel for an image fusion network that combines multiple reconstructed images. Experimental results demonstrated that the quality of decoded images can be significantly improved by the proposed multiple CAR network fusion method.

Highlights

  • Lossy compression of images is inevitable in most practical situations

  • To save network bandwidth and storage space, existing image and video compression standards, such as the joint photographic experts group (JPEG) and high efficiency video coding (HEVC) [1], apply quantization taking into account the trade-off between quality and bitrate

  • PIXEL-PRECISE NETWORK SELECTION NETWORK So far, we have explained the necessity of pixel-adaptive compression artifacts reduction (CAR), it is infeasible to apply different CAR networks to every pixel

Read more

Summary

Introduction

Lossy compression of images is inevitable in most practical situations. To save network bandwidth and storage space, existing image and video compression standards, such as the joint photographic experts group (JPEG) and high efficiency video coding (HEVC) [1], apply quantization taking into account the trade-off between quality and bitrate. An early approach, called deblocking [2], applies handcrafted filters to alleviate blocking artifacts caused by block-based compression. To this end, one of the advanced CAR methods estimates Wiener-based adaptive filters [3] by minimizing the mean squared error between the original and the decoded images. The Wiener filter can improve the quality of decoded images, its coefficients must be estimated for each frame and transmitted to the decoder side. These issues can be addressed by introducing machine learning frameworks

Objectives
Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.