Abstract
In recent decades, remote sensing object counting has attracted increasing attention from academia and industry due to its potential benefits in urban traffic, public safety, and road planning. However, this issue is becoming a challenge for computer vision because of various technical barriers, such as large-scale variation, complex background interference, and nonuniform density distribution. Recent results show hopeful prospects for object counting using convolutional neural networks (CNNs), but most existing CNN-based methods draw on larger and more complex architectures, which leads to a huge computational and storage burdens, severely limiting their application in real-world scenarios. In this article, a lightweight multiscale feature fusion network for remote sensing object counting, named LMSFFNet, is presented to achieve a better balance between the running speed of the network and the counting accuracy. Specifically, in the encoding process, we select a MobileViT module as the backbone of the network to reduce the numbers of network parameters and computing cost. In return, a cascade structure of the channel–spatial attention mechanisms compensates for the weaker feature extraction ability of the lightweight network. In the decoding process, a lightweight multiscale context fusion module (LMCFM) as a multiscale feature fusion module is developed to solve the problem that the number of parameters increases with the expansion of the object scale when extracting multiscale features. In addition, a lightweight counting scale pooling module (LCSPM) is used to mine the subtle features of the target object. Two kinds of typical object counting experiments, namely, experiments on remote sensing benchmarks (RSOC dataset) and crowd benchmarks (ShanghaiTech, UCF-QNRF, and UCF_CC_50 datasets), show the effectiveness of the proposed method.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: IEEE Transactions on Geoscience and Remote Sensing
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.