Abstract
Separating an RGB image into its albedo and shading maps abstracts the idea behind intrinsic image decomposition. It is a prolonged and underdetermined problem in computer vision, and we aim to tackle this problem using an efficient and optimized method. This work is an attempt to decompose the input image into its albedo and shading component using a Deep Convolution Neural Network (DCNN) model considering an encoder-decoder based architecture. Our model uses a pre-trained DenseNet-169 as an encoder for the input image and two decoders for the two output images viz. albedo and shading. The synthetic ground truth images from MPI-Sintel dataset helps in the dense supervision of our model. We show results for both synthetic images and real images from the IIW dataset. Our architecture aims to reduce the number of parameters used while training and has produced better results when compared with previous approaches on the MPI-Sintel dataset.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.