Abstract
Sufficient synthetic aperture radar (SAR) data is the key element in achieving excellent target recognition performance for most deep learning algorithms. It is unrealistic to obtain sufficient SAR data from the actual measurements, so SAR simulation based on electromagnetic scattering modeling has become an effective way to obtain sufficient samples. Simulated and measured SAR images are nonhomologous data. Due to the fact that the target geometric model of SAR simulation is not inevitably consistent with the real object, the SAR sensor model in SAR simulation may be different from the actual sensor, the background environment of the object is also inevitably different from that of SAR simulation, the error of electromagnetic modeling method itself, and so on. There are inevitable differences between the simulated and measured SAR images, which will affect the recognition performance. To address this problem, an SAR simulation method based on a high-frequency asymptotic technique and a discrete ray tracing technique is proposed in this paper to obtain SAR simulation images of ground vehicle targets. Next, various convolutional neural networks (CNNs) and AugMix data augmentation methods are proposed to train only on simulated data, and then target recognition on MSTAR measured data is performed. The experiments show that all the CNNs can achieve incredible recognition performance on the nonhomologous SAR data, and the RegNetX-3.2GF achieves state-of-the-art performance, up to 84.81%.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.