Abstract

The classification of remote sensing scene objects has been the subject of extensive studies in recent years due to the quick advancement of earth observation and remote sensing technology. More concentration, Hyperspectral images, and LiDAR data are complementary, and their combined use for data fusion can better mine the multi-dimensional features of ground objects in remote sensing scenes, which can effectively improve the classification accuracy and reliability of ground objects in remote sensing scenes. Single-modal remote sensing data frequently cannot fully meet the needs of ground feature classification due to the increasingly complex types of ground features. In order to solve this problem, we have developed two distinct multi-source fusion classification approaches using LiDAR and hyperspectral data and deep learning techniques. They are the reconstructed multi-layer perceptron network based on the variational auto-Encoder (encoding-decoding form) and the two-stream input convolutional neural network based on the cross-channel reconstruction mechanism. These two approaches can help us find more effective and deeper feature extraction methods and feature fusion methods in this research direction and design training. We adopted the network architecture model used in this area of research and used experimental data to demonstrate the effectiveness and superiority of the suggested network model.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.