Abstract

AbstractImage steganography is the art of concealing secret information within images to prevent detection. In deep‐learning‐based image steganography, a common practice is to fuse the secret image with the cover image to directly generate the stego image. However, not all features are equally critical for data hiding, and some insignificant ones may lead to the appearance of residual artifacts in the stego image. In this article, a novel network architecture for image steganography with hybrid attention mechanism based on generative adversarial network is introduced. This model consists of three subnetworks: a generator for generate stego images, an extractor for extracting the secret images, and a discriminator to simulate the detection process, which aids the generator in producing more realistic stego images. A specific hybrid attention mechanism (HAM) module is designed that effectively fuses information across channel and spatial domains, facilitating adaptive feature refinement within deep image representations. The experimental results suggest that the HAM module not only enhances the image quality during both the steganography and extraction processes but also improves the model's undetectability. Stego images are mixed with varying levels of noise in the training process, which can further improve robustness. Finally, it is verified that the model outperforms current steganography approaches on three datasets and exhibits good undetectability.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.