Abstract
With the advancement of remote sensing technology, the acquisition of ultra-high-resolution remote sensing imagery has become a reality, opening up new possibilities for detailed research and applications of Earth’s surface. These ultra-high-resolution images, with spatial resolutions at the meter or sub-meter level and pixel counts exceeding 4 million, contain rich geometric and attribute details of surface objects. Their use significantly improves the accuracy of surface feature analysis. However, this also increases the computational resource demands of deep learning-driven semantic segmentation tasks. Therefore, we propose the Transform Dual-Branch Attention Net (TDBAN), which effectively integrates global and local information through a dual-branch design, enhancing image segmentation performance and reducing memory consumption. TDBAN leverages a cross-collaborative module (CCM) based on the Transform mechanism and a data-related learnable fusion module (DRLF) to achieve adaptive content processing. Experimental results show that TDBAN achieves mean intersection over union (mIoU) of 73.6% and 72.7% on DeepGlobe and Inria Aerial datasets, respectively, and surpasses existing models in memory efficiency, highlighting its superiority in handling ultra-high-resolution remote sensing images. This study not only advances the development of ultra-high-resolution remote sensing image segmentation technology, but also lays a solid foundation for further research in this field.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have