Abstract
Current image tampering detection methods rely on various forgery footprints, such as JPEG artifacts and edge inconsistencies, and use algorithms related to image segmentation. However, these methods have several issues, including over-fitting, focusing on only a few specific forgery footprints, and emphasizing semantically relevant information while ignoring tampering traces. This paper proposes a model for image manipulation detection and localization based on multi-scale contrast learning(MSCL-Net). The model utilizes the differences in feature distributions between tampered and untampered regions to extract a comprehensive tamper trace. It uses a dual-stream structured encoder that incorporates both RGB raw images and SRM noise features. A Feature Cross-Fusion Module (FCFM) is proposed to fuse features for improving feature representation of tampered information. The decoding process involves the use of an Adaptive Self-Attention Module (ASAM) to filter and aggregate relevant context from coarse feature maps. Additionally, a Supervised Contrastive Learning Module (SCLM) is used to expand the difference between tampered and untampered areas. The loss function for multi-loss fusion comprises classification loss, segmentation loss, and multi-scale supervised contrastive loss. This improves the network's understanding of global differences, reduces false positives, weakens semantic information, and enhances the model's ability to locate tampered regions of varying sizes. Extensive experiments are conducted across multiple datasets, demonstrating that our model is robust against attacks and resilient to false-positive predictions at both the image-level and pixel-level. Furthermore, its overall performance exceeds that of state-of-the-art alternatives in reliably detecting and localizing tampered images.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.