Abstract

Unmanned aerial vehicle remote sensing images need to be precisely and efficiently classified. However, complex ground scenes produced by ultra-high ground resolution, data uniqueness caused by multi-perspective observations, and need for manual labelling make it difficult for current popular deep learning networks to obtain reliable references from heterogeneous samples. To address these problems, this paper proposes an optic nerve microsaccade (ONMS) classification network, developed based on multiple dilated convolution. ONMS first applies a Laplacian of Gaussian filter to find typical features of ground objects and establishes class labels using adaptive clustering. Then, using an image pyramid, multi-scale image data are mapped to the class labels adaptively to generate homologous reliable samples. Finally, an end-to-end multi-scale neural network is applied for classification. Experimental results show that ONMS significantly reduces sample labelling costs while retaining high cognitive performance, classification accuracy, and noise resistance—indicating that it has significant application advantages.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call