Abstract

Low-light remote sensing image enhancement (LLIE) for unmanned aerial vehicles (UAVs) has significant scientific and practical value because unfavorable lighting conditions make capture more difficult, resulting in undesired images. As acquiring real-world low-light/normal-light image pairs in the field of remote sensing is almost infeasible, performing LLIE in an unpaired manner is practical and valuable. However, without the paired data as the supervision, learning an LLIE network is challenging. To address the challenges, the paper proposes a novel yet effective method to unpaired LLIE, which maximizes the mutual information between low-light and restored images through self-similarity contrastive learning (SSCL) in a fully unsupervised fashion within a single deep GAN framework, named CLEGAN. Instead of supervising the learning using ground truth data, we propose to regularize the unpaired training using the information extracted from the input itself. The non-local patch sampling strategy in SSCL naturally makes the negative samples differ from the positive samples for discriminative representation. Moreover, the single GAN embeds the dual illumination perception module (DIPM) to handle the internal recurrence of information and overall uneven illumination distribution in remote sensing images. DIPM mainly consists of two cooperative blocks: spatial adaptive light adjustment module (SALAM) and global adaptive light adjustment module (GALAM). Specifically, SALAM exploits the internal recurrence of information in remote sensing images to encode a wider range of contextual information into local features and make proper light estimation. Simultaneously, GALAM enhances the most valuable illumination-related channels in the feature map to achieve better light estimation. The experiments on several datasets including low-light remote sensing image dataset and public low-light image datasets show that CLEGAN performs favorably against the existing unpaired LLIE approaches, and even outperforms several fully-supervised methods.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.