Abstract

Blind binocular visual quality prediction (BVQP) is more challenging than blind monocular visual quality prediction (MVQP). Recently, the application of convolutional neural networks (CNNs) to blind MVQP has resulted in significant progress in that area. In contrast, the adoption of deep learning for blind BVQP has received scant attention. In this study, we devised an end-to-end deep fusion network (DFNet) model trained in a unified framework for blind BVQP. This core prediction engine comprises monocular feature encoding networks and binocular feature fusion networks, followed by a quality prediction layer. The monocular feature encoding networks are first established to capture the low- and high-level monocular features of the left and right retinal views, respectively. Subsequently, these monocular features are integrated by the binocular feature fusion networks to obtain binocular deep features. Finally, the final binocular visual quality is predicted by quality prediction networks. Comparisons via experiments using two standard subject-rated BVQP datasets indicate that the proposed DFNet architecture achieves highly consistent alignment with human assessment and outperforms most relevant existing models.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call