Abstract

Automatic modulation recognition (AMR) is widely employed in communication systems. However, under conditions of low signal-to-noise ratio (SNR), recent studies reveal limitations in achieving high AMR accuracy. In this work, we introduce a novel network architecture that leverages a transformer-inspired approach tailored for AMR, called Feature-Enhanced Transformer with skip-attention (FE-SKViT). This innovative design adeptly harnesses the advantages of translation variant convolution and the Transformer framework, handling intra-signal variance and small cross-signal variance to achieve enhanced recognition accuracy. Experimental results on RadioML2016.10a, RadioML2016.10b, and RML22 datasets demonstrate that the Feature-Enhanced Transformer with skip-attention (FE-SKViT) excels over other methods, particularly under low SNR conditions ranging from −4 to 6 dB.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.