Abstract
For the imaging signal processing (ISP) pipeline of digital image devices, it is of high significance to remove undesirable illuminant effects and obtain color invariance, commonly known as ‘computational color constancy’. Achieving the computational color constancy requires going through two phases: the illumination estimation, which will be the primary focus of this work, and the human visual perception-based chromatic adaptation. At the first phase, illumination estimation is to predict RGB triplets, the numeric representations of incident illuminant colors, by calculating the values of image pixels. How much the network can increase its estimation accuracy is a key to realizing computational color constancy. With recent advances in deep learning (DL), a lot of deep learning-based approaches have been suggested, bringing higher accuracy to computer vision applications, but there are still quite a few obstacles to overcome such as instability of learning. In an attempt to address this ill-posed problem in the illumination estimation space, this article presents a novel deep learning-based approach, the Cascading Residual Network Architecture (CRNA), which incorporates the ResNet and cascading mechanism into the deep convolutional neural network (DCNN). The cascading mechanism enables the proposed network to restrain from suddenly varying in size, serves to mitigate learning instability, and accordingly reduces the quality degradation. This is attributed to the ability of the cascading mechanism that fine-tunes the pre-trained DCNN. Considerable amounts of datasets and comparative experiments highlight that the proposed approach delivers more stable and robust results and imply the potential for generalization of the proposed approach across deep learning applications.
Highlights
In digital photography, digital images may carry undesired color casts due to an unintended source illuminant in a scene
Achieving the computational color constancy requires going through two phases: the illumination estimation, which will be the primary focus of this work, and the human visual perception-based (HVP)
This article presents a novel approach to more accurate illuminant estimation by embedding the ResNet and cascading mechanism into the deep convolutional neural network (DCNN)
Summary
Digital images may carry undesired color casts due to an unintended source illuminant in a scene. The imaging model [1] is used to calculate the pixel values based on three factors: the spectrum of the source illuminant, the reflectance of the object surface and the spectral sensitivity of the camera sensor. The latter two factors: the reflectance of the object surface and the spectral sensitivity of the camera sensor, are kept constant by shooting the same scene with the same camera whereas the spectrum of the source illuminant is varied by taking the photos under varying illuminant conditions This points to a fact that a digital camera can capture any incident source illuminant but cannot detect the illuminant itself. Achieving the computational color constancy requires going through two phases: the illumination estimation, which will be the primary focus of this work, and the HVP-
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.