Most existing low-light image enhancement approaches predominantly focus on directly modeling the mapping between low-light images and normal-light images, but models trained on the single fixed data set often fail to achieve satisfactory results in complex real-world scenes. To address the limitations posed by limited data, this paper proposes a joint framework to learn spatially adaptive degradation and enhancement representations of bright-to-dark images and dark-to-bright images. Firstly, we design an illumination degradation model based on cycle consistent loss to learn how to degrade normal-light images into low-light reference images. The degradation model can synthesize a large number of normal/degraded-light pairs, which is beneficial for the training of the second-stage enhancement model. Subsequently, leveraging the self-luminance information of images, we construct a self-attention-based light enhancement model. The enhancement model is trained on mixed data, which combines both real normal/low-light image pairs and synthetic normal/degraded-light image pairs. In this way, the enhancement model can learn more features from a diverse set of images, enhancing its robustness. Extensive experiments on low-light image enhancement, low-light object detection, and nighttime image segmentation have demonstrated that our proposed method outperforms existing supervised and unsupervised methods across a variety of real-world scenes.
Read full abstract