Abstract
In the realm of remote sensing image object detection compression distillation, establishing an efficient method for instance feature knowledge transfer between teacher and student models holds paramount importance. To this end, this paper introduces an innovative deep structured instance graph distillation method that endeavors to delve into the underlying information between instance features, thereby optimizing detection performance. Specifically, our proposed method incorporates feature instances and their relations into a graph-based structure (SIG). In this graph, feature instances serve as the nodes, while the relations between them serve as the edges. This structure enables us to capture both the individual significance of each feature instance and their collective influence within the context. Furthermore, in the experiment, we found that the index of some dense and small-target objects did not improve much because the edge assembly generated by a large number of background feature nodes in the SIG module inhibited the loss. To address the perennial imbalance between foreground and background features, we introduce an adaptive background feature mining strategy. Through carefully calibrated weights, this strategy effectively extracts and integrates background information, thereby minimizing noise interference in detection results and augmenting the expressive capacity of foreground features. We achieved state-of-the-art results on both the challenging DIOR and DOTA datasets, with the two-stage Oriented RCNN-based student Resnet18 model achieving a 73.23 mAP on the DOTA benchmark, close to the teacher Resnet101’s 76.16. In addition, on the DIOR dataset, the student Resnet18 based on the two-stage Faster RCNN achieved 70.13 mAP, higher than the baseline 66.31, and the student Resnet50 achieved 72.28, higher than the teacher’s 72.25.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.