Abstract
We obtain a novel analytical derivation for distortion-related constraints in a neural network- (NN)-based automatic target recognition (ATR) system. We obtain two types of constraints for a realistic ATR system implementation involving 4-f correlator architecture. The first constraint determines the relative size between the input objects and input correlation filters. The second constraint dictates the limits on amount of rotation, translation, and scale of input objects for system implementation. We exploit these constraints in recognition of targets varying in rotation, translation, scale, occlusion, and the combination of all of these distortions using a learning vector quantization (LVQ) NN. We present the simulation verification of the constraints using both the gray-scale images and Defense Advanced Research Projects Agency's (DARPA's) Moving and Stationary Target Recognition (MSTAR) synthetic aperture radar (SAR) images with different depression and pose angles.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.