With the decline in the protective function for agricultural ecosystems of farmland shelterbelts due to tree withering and dying caused by pest and disease, quickly and accurately identifying the distribution of canopy damage is of great significance for forestry management departments to implement dynamic monitoring. This study focused on Populus bolleana and utilized an unmanned aerial vehicle (UAV) multispectral camera to acquire red–green–blue (RGB) images and multispectral images (MSIs), which were fused with a digital surface model (DSM) generated by UAV LiDAR for feature fusion to obtain DSM + RGB and DSM + MSI images, and random forest (RF), support vector machine (SVM), maximum likelihood classification (MLC), and a deep learning U-Net model were employed to build classification models for forest stand canopy recognition for the four image types. The model results indicate that the recognition performance of RF is superior to that of U-Net, and U-Net performs better overall than SVM and MLC. The classification accuracy of different feature fusion images shows a trend of DSM + MSI images (Kappa = 0.8656, OA = 91.55%) > MSI images > DSM + RGB images > RGB images. DSM + MSI images exhibit the highest producer’s accuracy for identifying healthy and withered canopies, with values of 95.91% and 91.15%, respectively, while RGB images show the lowest accuracy, with producer’s accuracy values of 79.3% and 78.91% for healthy and withered canopies, respectively. This study presents a method for identifying the distribution of Populus bolleana canopies damaged by Anoplophora glabripennis and healthy canopies using the feature fusion of multi-source remote sensing data, providing a valuable data reference for the precise monitoring and management of farmland shelterbelts.
Read full abstract