In this paper, we introduce a completely blind and unsupervised no-reference model that performs the quality assessment of stereoscopic 3D videos by using joint motion and depth visibility. In order to achieve such objective metric, first we compute the correlation map between the motion vector map and the depth map, and then we study the natural scene statistics of the estimated correlation map. We empirically model these statistics with Univariate Generalized Gaussian Distribution (UGGD) and compute the UGGD parameters at multi-scale and multi-orient steerable subband decomposition. It is shown that the estimated UGGD features are capable of discriminating distortions. We then compute the mean vector and the covariance matrix from the estimated UGGD features of the correlation maps of motion vector and the depth maps of undistorted and distorted S3D videos. This is followed by the computation of the Wave Hedges distance between the mean vectors of the undistorted and distorted contents, and the Bhattacharyya distance is measured between the corresponding covariance matrices. We pool these distance measures to achieve the overall joint motion and depth quality of S3D videos. The efficiency of the proposed algorithm is evaluated on the well-known IRCCYN, LFOVIAS3DPh1 and LFOVIAS3DPh2 S3D video datasets. Our model demonstrates robust and consistent performance across all distortion types and shows competitive performance against other 2D and 3D image and video quality assessment algorithms. Furthermore, since the proposed algorithm is completely blind, it does not require any training and testing analysis on the content features and the subjective scores.
Read full abstract