Abstract

This article presents a novel method to obtain a sparse representation of multiview images. The method is based on the fact that multiview data is composed of epipolar-plane image lines which are highly redundant. We extend this principle to obtain the layer-based representation, which partitions a multiview image dataset into redundant regions (which we call layers) each related to a constant depth in the observed scene. The layers are extracted using a general segmentation framework which takes into account the camera setup and occlusion constraints. To obtain a sparse representation, the extracted layers are further decomposed using a multidimensional discrete wavelet transform (DWT), first across the view domain followed by a two-dimensional (2D) DWT applied to the image dimensions. We modify the viewpoint DWT to take into account occlusions and scene depth variations. Simulation results based on nonlinear approximation show that the sparsity of our representation is superior to the multi-dimensional DWT without disparity compensation. In addition we demonstrate that the constant depth model of the representation can be used to synthesise novel viewpoints for immersive viewing applications and also de-noise multiview images.

Highlights

  • The notion of sparsity, namely the idea that the essential information contained in a signal can be represented with a small number of significant components, is widespread in signal processing and data analysis in general

  • Our results show that the proposed layer-based representation offers superior approximation properties compared to a typical multi-dimensional DWTh

  • 6 Conclusion We presented a novel method to obtain a sparse representation of multiview images

Read more

Summary

Introduction

The notion of sparsity, namely the idea that the essential information contained in a signal can be represented with a small number of significant components, is widespread in signal processing and data analysis in general. Model must take into account appearing (disocclusions) and disappearing (occlusions) objects This nonlinear property means that finding a sparse representation is inherently more difficult than in the two-dimensional (2D) case. For this reason, in this article we propose a hybrid method to obtain a sparse representation of multiview images. The layer-based representation partitions the multiview images into a set of layers each related to a constant depth in the observed scene. In the case of the 4D light field, the sparse representation of the data is obtained by taking a 4D discrete wavelet transform (DWT) of each depth layer. We review the structure of multiview data, discuss the layer-based representation and present a high-level overview of our proposed method.

Plenoptic function
Layer-based representation
Multiview image segmentation
Imposing camera setup and occlusion constraints
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call