Abstract
With the increased spatial resolution of current sensor constellations, more details are captured about our changing planet, enabling the recognition of a greater range of land use/land cover classes. While pixeland object-based classification approaches are widely used for extracting information from imagery, recent studies have shown the importance of spatial contexts for discriminating more specific and challenging classes. This paper proposes a new compact representation for the fast query/classification of compound structures from very high resolution optical remote sensing imagery. This bag-of-features representation relies on the multiscale segmentation of the input image and the quantization of image structures pooled into visual word distributions for the characterization of compound structures. A compressed form of the visual word distributions is described, allowing adaptive and fast queries/classification of image patterns. The proposed representation and the query methodology are evaluated for the classification of the UC Merced 21-class data set, for the detection of informal settlements and for the discrimination of challenging agricultural classes. The results show that the proposed representation competes with state-of-the-art techniques. In addition, the complexity analysis demonstrates that the representation requires about 5% of the image storage space while allowing us to perform queries at a speed down to 1 s/ 1000 km 2 /CPU for 2-m multispectral data.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have