Mapping water bodies from remotely sensed imagery is crucial for understanding hydrological and biogeochemical processes. The identification of water extent is mainly dependent on optical and synthetic aperture radar (SAR) images. However, the use of remote sensing for water body mapping is often undermined by the mixed pixel dilemma inherent to traditional hard classification approaches. At the same time, the presence of clouds in optical imagery and speckle noise in SAR imagery, coupled with the difficulty in differentiating between water-like surfaces and actual water bodies, significantly compromise the accuracy of water body identification. This paper proposes a DEEP feature collaborative convolutional neural network (CNN) for Water Super-Resolution Mapping based on Optical and SAR images (DeepOSWSRM), which collaboratively leverages Sentinel-1 and Sentinel-2 imagery to address the challenges of missing data and mixed pixels. The Sentinel-1 image provides complementary water distribution information for the cloudy areas of the Sentinel-2 image, while the Sentinel-2 image enhances the perception capabilities for small water bodies in the Sentinel-1 image. Using PlanetScope imagery as the true reference data, the effectiveness of the proposed method was assessed through two experimental scenarios: one utilizing synthetic coarse-resolution imagery degraded from Sentinel-1 and Sentinel-2 data and another using actual Sentinel-1 and Sentinel-2 data, encompassing both simulated and real cloud conditions. A comparative analysis was conducted against three state-of-the-art CNN-based water mapping methods and two CNN SRM methods. The findings demonstrate that the proposed DeepOSWSRM method successfully produces accurate, fine-resolution water body maps, with its performance mainly benefiting from the fusion of SAR and optical images.
Read full abstract