Abstract

An efficient steganographic method is proposed for embedding secret messages into gray scale and RGB colored images. In this paper, a wavelet domain steganography model is adopted for hiding a large amount of data with high security, good invisibility and no loss of secret message. The information is embedded in those areas of the host image that contains high texture to reduce visibility of the embedded information in the host image. The qualified significant wavelet coefficients and their texture and sensitivity to gray value variations across different coarse scales (level 1, level 2 and level 3 wavelet decompositions) are utilized to determine the positions and the magnitudes to adaptively embed the secret message. The highly textured areas near the edges in high frequency subbands in the wavelet domain are considered. The transformed cover image is divided into number of joint non-overlapping N × N blocks in each layer (R, G, and B layers). A difference value is calculated from the values of the adjacent pixels in each block, producing the partitioned difference image (PDI). The selection of the range intervals is based on the characteristics of human vision's sensitivity to gray value variations from smoothness to contrast. The number of bits which can be embedded in each block varies and is decided by the width of the range to which the difference value of the two pixels and block histogram peak (Pk) belongs to. The embedded secret message can be extracted from the resulting stegno-image and the original cover image is reversed. The proposed model will produce a high-capacity image steganography technique with acceptable level of imperceptibility and distortion in the cover image and high level of overall security. This solution is independent of the nature of the data to be hidden and produces a stego image with minimum degradation. Experimental results show the feasibility of the proposed methods. Various statistics attacks were also conducted to collect related data to show the security of the method.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.