Abstract

Object detection has broad applications in areas such as autonomous driving, security surveillance, and deep-sea exploration. However, the performance of current detection algorithms significantly decreases due to the loss of detail, increased noise, and color distortion in images under low-light or nighttime conditions. To address this problem, we propose a plug-and-play multiscale pyramid enhancement network (MPENet), which elegantly cascades with RT-DETR to establish an end-to-end framework for low-light object detection, named MPE-DETR. First, MPENet utilizes Gaussian blur to decompose images into Gaussian pyramids and Laplacian pyramids at different resolutions. Specifically, we designed a high-frequency texture enhancement (HTE) module to capture the edge and texture information of images, and a low-frequency noise smoothing (LNS) module to better understand the overall structure of images and capture global-scale features. Additionally, by concatenating the output features of the HTE and LNS modules along the channel dimension, feature fusion across different scales is realized. We conducted experiments on the ExDark and ExDark + LOD datasets, which are designed for low-light object detection. The results indicate that the proposed method achieved an improvement of 2.1% in mAP@0.5 compared to that of existing SOTA models on the ExDark dataset, and demonstrated strong generalizability and robustness on the ExDark + LOD dataset. The code and results are available at https://github.com/PZDJL/MPENet.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call