As machine learning continues to evolve, traditional image steganography techniques find themselves increasingly unable to meet the dual challenge of deceiving both the human visual system and machine learning models. In response, this paper introduces the Visually Robust Image Steganography (VRIS) model, specifically tailored for this dual deception task. The VRIS model elevates visual security through the use of specialized feature extraction and processing methodologies. It meticulously conducts feature-level fusion between secret images and cover images, ensuring a high level of visual similarity between them. To effectively mislead machine learning models, the VRIS model incorporates a sophisticated strategy utilizing random noise factors and discriminators. This involves adding controlled amounts of random Gaussian noise to the encrypted image, thereby enhancing the difficulty for machine learning frameworks to recognize it. Furthermore, the discriminator is trained to discern between the noise-infused encrypted image and the original cover image. Through adversarial training, the discriminator and VRIS model refine each other, successfully deceiving the machine learning systems. Additionally, the VRIS model presents an innovative method for extracting and reconstructing secret images. This approach safeguards secret information from unauthorized access while enabling legitimate users to non-destructively extract and reconstruct images by leveraging multi-scale features from the encrypted image, combined with advanced feature fusion and reconstruction techniques. Experimental results validate the effectiveness of the VRIS model, achieving high PSNR and SSIM scores on the LFW dataset and demonstrating significant deception capabilities against the ResNet50 model on the Mini-ImageNet dataset, with an impressive misclassification rate of 99.24%.
Read full abstract