Abstract

The classification of remote sensing scene objects has been the subject of extensive studies in recent years due to the quick advancement of earth observation and remote sensing technology. More concentration, Hyperspectral images, and LiDAR data are complementary, and their combined use for data fusion can better mine the multi-dimensional features of ground objects in remote sensing scenes, which can effectively improve the classification accuracy and reliability of ground objects in remote sensing scenes. Single-modal remote sensing data frequently cannot fully meet the needs of ground feature classification due to the increasingly complex types of ground features. In order to solve this problem, we have developed two distinct multi-source fusion classification approaches using LiDAR and hyperspectral data and deep learning techniques. They are the reconstructed multi-layer perceptron network based on the variational auto-Encoder (encoding-decoding form) and the two-stream input convolutional neural network based on the cross-channel reconstruction mechanism. These two approaches can help us find more effective and deeper feature extraction methods and feature fusion methods in this research direction and design training. We adopted the network architecture model used in this area of research and used experimental data to demonstrate the effectiveness and superiority of the suggested network model.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call