Abstract

Multiview video plus depth is one of the mainstream representations of 3D scenes in emerging free viewpoint video, which generates virtual 3D synthesized images through a depth-image-based-rendering (DIBR) technique. However, the inaccuracy of depth maps and imperfect DIBR techniques result in different geometric distortions that seriously deteriorate the users’ visual perception. An effective 3D synthesized image quality assessment (IQA) metric can simulate human visual perception and determine the application feasibility of the synthesized content. In this paper, a no-reference IQA metric based on visual-entropy-guided multi-layer features analysis for 3D synthesized images is proposed. According to the energy entropy, the geometric distortions are divided into two visual attention layers, namely, bottom-up layer and top-down layer. The feature of salient distortion is measured by regional proportion plus transition threshold on a bottom-up layer. In parallel, the key distribution regions of insignificant geometric distortion are extracted by a relative total variation model, and the features of these distortions are measured by the interaction of decentralized attention and concentrated attention on top-down layers. By integrating the features of both bottom-up and top-down layers, a more visually perceptive quality evaluation model is built. Experimental results show that the proposed method is superior to the state-of-the-art in assessing the quality of 3D synthesized images.

Highlights

  • With the advancement of video technologies, a free viewpoint video (FVV) system is gradually applied to various fields, such as distance education, medical service, and entertainment [1]

  • The hole problem can be alleviated in synthesized images, but inaccurate depth information brings the geometric distortions, curving, and object shifting, which are respectively visible in the chair and face of Figure 1b,c

  • This section mainly evaluates the performance of the visual-entropy-guided multi-layer features analysis (MLFA) method

Read more

Summary

Introduction

With the advancement of video technologies, a free viewpoint video (FVV) system is gradually applied to various fields, such as distance education, medical service, and entertainment [1]. Due to the particularity of synthetic distortions, the existing IQA methods for 2D traditional distortions, like [10,11,12,13,14], cannot measure the 3D synthesized distortions effectively With this concern, some researchers have proposed IQA metrics targeting 3D synthesized images. Wang et al extracted features of geometric distortion, global sharpness, and image complexity in a wavelet transform domain to evaluate the quality of 3D synthesized images [29]. These transform-domain-based metrics eliminate uninterested information of synthesized image and save calculation time but are still sensitive to limited geometric distortion types.

Motivation
The Proposed Visual-Entropy-Guided MLFA Method
Feature Extraction of the Bottom-Up Layer
Feature Extraction of a Top-Down Layer
Experimental Results and Analysis
Databases and Evaluation Criteria
Parameters Determination
Impact of Training Percentages
Performance Analysis of a Multi-Layer Strategy
Performance Analysis of Key Region Extraction
Feature Ablation Experiments
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call