Extracting fragmented cropland is essential for effective cropland management and sustainable agricultural development. However, extracting fragmented cropland presents significant challenges due to its irregular and blurred boundaries, as well as the diversity in crop types and distribution. Deep learning methods are widely used for land cover classification. This paper proposes ConvNeXt-U, a lightweight deep learning network that efficiently extracts fragmented cropland while reducing computational requirements and saving costs. ConvNeXt-U retains the U-shaped structure of U-Net but replaces the encoder with a simplified ConvNeXt architecture. The decoder remains unchanged from U-Net, and the lightweight CBAM (Convolutional Block Attention Module) is integrated. This module adaptively adjusts the channel and spatial dimensions of feature maps, emphasizing key features and suppressing redundant information, which enhances the capture of edge features and improves extraction accuracy. The case study area is Hengyang County, Hunan Province, China, using GF-2 remote sensing imagery. The results show that ConvNeXt-U outperforms existing methods, such as Swin Transformer (Acc = 85.1%, IoU = 79.1%), MobileNetV3 (Acc = 83.4%, IoU = 77.6%), VGG16 (Acc = 80.5%, IoU = 74.6%), and ResUnet (Acc = 81.8%, IoU = 76.1%), achieving an IoU of 79.5% and Acc of 85.2%. Under the same conditions, ConvNeXt-U has a faster inference speed of 37 images/s, compared to 28 images/s for Swin Transformer, 35 images/s for MobileNetV3, and 0.43 and 0.44 images/s for VGG16 and ResUnet, respectively. Moreover, ConvNeXt-U outperforms other methods in processing the boundaries of fragmented cropland, producing clearer and more complete boundaries. The results indicate that the ConvNeXt and CBAM modules significantly enhance the accuracy of fragmented cropland extraction. ConvNeXt-U is also an effective method for extracting fragmented cropland from remote sensing imagery.
Read full abstract