Abstract

The problem of reconstructing a depth map from a sequence of differently focused images (focal stack) is called Depth from focus. The core idea of this method is to analyze the sharpness of each pixel and compare it along the z axis of the focal stack to estimate the true depth value. This approach has two main drawbacks: it depends on the optics of the camera and on the focus measure operator. Recent advances in deep learning techniques show promising results in this way, however, still have problems generalizing to different scenes, cameras, optics and focal stack focus positions. In this paper we propose a novel deep learning based method to approach this problem. Firstly, we propose to estimate distances relative to the focal stack focus position instead of estimating true or invariant depths, allowing us to generalize different scenes and optical setups without losing the possibility to extract real distances. Secondly, we present our novel architecture: a 2D siamese encoder–3D decoder with a differentiable argmax regression that is able to compute depth from stacks of variable sizes. Finally, we compare our method with 2 other depth from focus algorithms and with a monocular depth estimation method.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.