Abstract

For the imaging signal processing (ISP) pipeline of digital image devices, it is of high significance to remove undesirable illuminant effects and obtain color invariance, commonly known as ‘computational color constancy’. Achieving the computational color constancy requires going through two phases: the illumination estimation, which will be the primary focus of this work, and the human visual perception-based chromatic adaptation. At the first phase, illumination estimation is to predict RGB triplets, the numeric representations of incident illuminant colors, by calculating the values of image pixels. How much the network can increase its estimation accuracy is a key to realizing computational color constancy. With recent advances in deep learning (DL), a lot of deep learning-based approaches have been suggested, bringing higher accuracy to computer vision applications, but there are still quite a few obstacles to overcome such as instability of learning. In an attempt to address this ill-posed problem in the illumination estimation space, this article presents a novel deep learning-based approach, the Cascading Residual Network Architecture (CRNA), which incorporates the ResNet and cascading mechanism into the deep convolutional neural network (DCNN). The cascading mechanism enables the proposed network to restrain from suddenly varying in size, serves to mitigate learning instability, and accordingly reduces the quality degradation. This is attributed to the ability of the cascading mechanism that fine-tunes the pre-trained DCNN. Considerable amounts of datasets and comparative experiments highlight that the proposed approach delivers more stable and robust results and imply the potential for generalization of the proposed approach across deep learning applications.

Highlights

  • In digital photography, digital images may carry undesired color casts due to an unintended source illuminant in a scene

  • Achieving the computational color constancy requires going through two phases: the illumination estimation, which will be the primary focus of this work, and the human visual perception-based (HVP)

  • This article presents a novel approach to more accurate illuminant estimation by embedding the ResNet and cascading mechanism into the deep convolutional neural network (DCNN)

Read more

Summary

Introduction

Digital images may carry undesired color casts due to an unintended source illuminant in a scene. The imaging model [1] is used to calculate the pixel values based on three factors: the spectrum of the source illuminant, the reflectance of the object surface and the spectral sensitivity of the camera sensor. The latter two factors: the reflectance of the object surface and the spectral sensitivity of the camera sensor, are kept constant by shooting the same scene with the same camera whereas the spectrum of the source illuminant is varied by taking the photos under varying illuminant conditions This points to a fact that a digital camera can capture any incident source illuminant but cannot detect the illuminant itself. Achieving the computational color constancy requires going through two phases: the illumination estimation, which will be the primary focus of this work, and the HVP-

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call