Abstract
Accurate individual tree detection by unmanned aerial vehicles (UAVs) is a critical technique for smart forest management and serves as the foundation for evaluating ecological functions. Existing object detection and segmentation methods, on the other hand, have reduced accuracy when detecting and segmenting individual trees in complicated urban forest landscapes, as well as poor mask segmentation quality. This study proposes a novel Mask-CSP-attention-coupled network (MCAN) based on the Mask R-CNN algorithm. MCAN uses the Cross Stage Partial Net (CSPNet) framework with the Sigmoid Linear Unit (SiLU) activation function in the backbone network to form a new Cross Stage Partial Residual Net (CSPResNet) and employs a convolutional block attention module (CBAM) mechanism to the feature pyramid network (FPN) for feature fusion and multiscale segmentation to further improve the feature extraction ability of the model, enhance its detail information detection ability, and improve its individual tree detection accuracy. In this study, aerial photography of the study area was conducted by UAVs, and the acquired images were used to produce a dataset for training and validation. The method was compared with the Mask Region-based Convolutional Neural Network (Mask R-CNN), Faster Region-based Convolutional Neural Network (Faster R-CNN), and You Only Look Once v5 (YOLOv5) on the test set. In addition, four scenes—namely, a dense forest distribution, building forest intersection, street trees, and active plaza vegetation—were set up, and the improved segmentation network was used to perform individual tree segmentation on these scenes to test the large-scale segmentation ability of the model. MCAN’s average precision (AP) value for individual tree identification is 92.40%, which is 3.7%, 3.84%, and 12.53% better than that of Mask R-CNN, Faster R-CNN, and YOLOv5, respectively. In comparison to Mask R-CNN, the segmentation AP value is 97.70%, an increase of 8.9%. The segmentation network’s precision for the four scenes in multi-scene segmentation ranges from 95.55% to 92.33%, showing that the proposed network performs high-precision segmentation in many contexts.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.