ABSTRACT Land cover classification, or semantic segmentation, has been one of the most critical research areas in remote sensing (RS) and remains crucial for many downstream applications. Although deep learning (DL) based models have recently dominated this field, models trained using the dataset of one region generally cannot predict reliable classification results in other regions. Despite the abundance of high-resolution (HR) satellite images around the globe, not enough HR labels are available for building a one-for-all-purpose model. This study utilizes widely available low-resolution (LR) labels with weakly supervised methods to obtain land cover maps worldwide. Previous methods designed new learning components, such as loss functions, to fully use LR labels or trained DL models with LR labels to generate intermediate labels for further training and processing. Since this information is directly used or extracted from the original LR labels, methods without additional robustness are sensitive to noise in the labels, which causes error propagation in the results. On this basis, a novel label refinement approach is proposed that transforms noisy original LR labels into refined HR labels using two steps of noise filtering. First, based on spectral indices from the HR images, we select relatively confident labels from the LR labels through a Markov Random Field optimization framework. Second, a shallow classifier such as random forest (RF) is trained using the selected pixels to supplement previously unselected labels and refine low-confidence labels with new and high-confidence labels. The results showed that in the experiment on the Data Fusion Contest 2020 dataset, the semantic segmentation models trained using our refined HR labels had a 2–14% higher average accuracy than those trained using the original LR labels. They also outperformed other weakly supervised methods directly using original LR labels and had a 7.5% higher average accuracy than the winning method.
Read full abstract