Next Article in Journal
Effect of Time and Temperature on Physicochemical and Microbiological Properties of Sous Vide Chicken Breast Fillets
Previous Article in Journal
Cloud-Edge Collaboration-Based Knowledge Sharing Mechanism for Manufacturing Resources
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Color Image Self-Recovery and Tampering Detection Scheme Based on Fragile Watermarking with High Recovery Capability

by
Rogelio Reyes-Reyes
,
Clara Cruz-Ramos
,
Volodymyr Ponomaryov
*,
Beatriz P. Garcia-Salgado
and
Javier Molina-Garcia
Instituto Politecnico Nacional, ESIME Culhuacan, Santa Ana 1000, Mexico-City 04440, Mexico
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(7), 3187; https://doi.org/10.3390/app11073187
Submission received: 4 March 2021 / Revised: 28 March 2021 / Accepted: 30 March 2021 / Published: 2 April 2021
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
In this paper, a fragile watermarking scheme for color image authentication and self-recovery with high tampering rates is proposed. The original image is sub-sampled and divided into non-overlapping blocks, where a watermark used for recovery purposes is generated for each one of them. Additionally, for each recovery watermark, the bitwise exclusive OR (XOR) operation is applied to obtain a single bit for the block authentication procedure. The embedding and extraction process can be implemented in three variants (1-LSB, 2-LSB or 3-LSB) to solve the tampering coincidence problem (TCP). Three, six or nine copies of the generated watermarks can be embedded according to the variant process. Additionally, the embedding stage is implemented in a bit adjustment phase, increasing the watermarked image quality. A particular procedure is applied during a post-processing step to detect the regions affected by the TCP in each recovery watermark, where a single faithful image used for recovery is generated. In addition, we involve an inpainting algorithm to fill the blocks that have been tampered with, significantly increasing the recovery image quality. Simulation results show that the proposed framework demonstrates higher quality for the watermarked images and an efficient ability to reconstruct tampered image regions with extremely high rates (up to 90%). The novel self-recovery scheme has confirmed superior performance in reconstructing altered image regions in terms of objective criteria values and subjective visual perception via the human visual system against other state-of-the-art approaches.

1. Introduction

Currently, the development of authentication and reconstruction techniques for digital images has been the focus of extensive research due to the accelerated growth of image editing software, which can be used to tamper with digital images in multiple ways. These authentication and reconstruction techniques are used to detect tampered regions in images where, in the case of alteration, a recovery process should be applied to retrieve the original content. These schemes are helpful in different applications, in which undetected modifications of digital images may have serious consequences, e.g., legal proceedings, where a digital image can be used as legal evidence. Therefore, detection and recovery of tampered content in digital images have become issues of outstanding importance.
In recent years, watermarking techniques have been used to authenticate and recover tampered information in digital images [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24]. Watermarking techniques can be classified into three types [1,2]: fragile, semi-fragile and robust. Fragile watermarking [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,23,24] does not support intentional and unintentional attacks; so, in case of any modification, the watermark would be destroyed. In contrast, these techniques offer a high payload capacity and are mainly used for authentication [3,4,5,6,7] that justifies several frameworks, which recently appeared in the self-recovery of tampered image regions [1,2,8,9,10,11,12,13,14,15,16,17,18,19,23,24]. Semi-fragile watermarking techniques are commonly used for copyright protection [20,21,22] and recovery schemes [25,26,27,28,29,30]. These techniques are designed to resist non-intentional manipulations caused by traditional image processing operations such as JPEG compression and scaling. These methods are fragile against intentional manipulations like tampering, resulting in a lower recovery rate compared to strategies based on fragile watermarking. Finally, robust watermarking techniques [31,32,33] are mainly used for copyright protection because they support intentional and non-intentional attacks. Their main disadvantage is a reduced payload capacity in comparison with fragile and semi-fragile watermarking techniques.
Self-recovery techniques based on watermarking consist of initially using small blocks of an image; subsequently, a watermark should be generated and embedded into a different block for content recovery for each self-recovery block. During the recovery process, all tampered blocks are reconstructed by the recovered watermark. This step could fail when a block containing the recovery watermark has been tampered with. In this way, it is impossible to reconstruct a concrete block, generating the so-called tampering coincidence problem.
Considering the approaches mentioned above for authentication and tamper detection, which are based on watermarking, the following properties are required for efficient implementation:
(a)
A minimum number of bits used for recovery and tamper detection: the recovery bits should be embedded redundantly, thus avoiding the tampering coincidence problem.
(b)
Watermark imperceptibility: the embedded recovery and authentication bits must not affect the visual quality of a watermarked image.
(c)
Precise tamper detection: a majority of intentional modifications should be accurately detected.
(d)
Precise recovery of tampered regions: the reconstructed image must demonstrate acceptable visual and objective quality in the reconstructed areas.
In this paper, a self-recovery of high tampering rate framework (denoted as SR-HTR) is designed according to the previously presented properties required for an efficient authentication scheme and tamper detection in color images. This novel fragile scheme appears to demonstrate a high payload capacity that can be used for authentication and recovery processes. To minimize the negative influence of the tampering coincidence problem, the proposed algorithm generates 15 A B / 16 bits in total, A × B   being the size of the image. Additionally, A B / 16 bits should be produced to detect tampered pixels in any area of an image. Thus, the designed framework can embed three, six or nine copies of recovery and authentication watermarks, achieving a high recovery and tamper detection capability. The novel framework can be implemented using several variants for embedding the recovery and authentication watermarks, such as the least-significant bit (LSB) methods: 1-LSB, 2-LSB and 3-LSB, where each one of them provides different advantages that are analyzed below.
For the 2-LSB and 3-LSB embedding processes, a bit adjustment stage is performed [34] on the watermarked pixels, increasing the protected images’ objective quality. A hierarchical algorithm in tamper detection is employed to achieve higher tamper detection accuracy. Additionally, an inpainting process is used to resolve the tampering coincidence problem by regenerating the eliminated blocks.
For evaluating the quality of the results obtained in the numerous experiments, the peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) criteria are employed in this study. Moreover, we use a variation of the PSNR criterion, denoted as PSNR-HVS-M, which considers the human visual system (HVS) and visual masking (M). This criterion employs the contrast sensitivity function (CSF) and maintains a close relationship with discrete cosine transform (DCT) basis functions [35]. Additionally, it has demonstrated good correspondence with human subjective visual perception. Consequently, it could be useful for the justification of the good performance of our novel system.
The rest of this paper is organized as follows. Firstly, Section 2 presents a review of related works. Secondly, Section 3 describes the proposed SR-HTR, followed by Section 4, which explains the experimental setup. Section 5 shows the proposed method’s analysis when the embedding and extraction processes in 1-LSB, 2-LSB and 3-LSB are used. Section 6 presents the experimental results obtained by the proposed framework and their performance comparison against state-of-the-art techniques. Finally, the study’s conclusion is drawn in Section 7.

2. Related Works

In this section, previously reported watermarking algorithms for authentication and self-recovery in digital images are discussed.
Singh et al. [2] divided the image to be protected into non-overlapping blocks of 2 × 2 pixels where, for each block, two flags are generated for tamper detection. In addition, the DCT and a quantization procedure are applied to each DCT block. Next, the most important coefficients are identified and, finally, they are transformed to 10 bits. The LSB method is used to embed the recovery and authentication bits.
Zhang et al. [8] proposed an algorithm for authentication and self-recovery based on a fragile watermarking scheme, in which the recovery bits should be obtained from the original DCT coefficients and a compressive sensing approach. The watermark does not contain additional redundancy, and the embedding process is based on the LSB method, in which the watermark data replace the three LSBs of each pixel.
He et al. [9] generated the recovery bits using non-overlapping blocks of 2 × 2 pixels. Six recovery bits are generated for each block using six most significant bits (MSBs) by averaging and normalization processes. The final watermark is obtained by an encryption process of the six recovery bits and a secret key, where the recovery bits are used to locate the tampered regions. Then, two LSBs of each pixel are used for watermark embedding.
Tong et al. [10] proposed a scheme where two flags are obtained for each non-overlapping block of 2 × 2 pixels to generate tamper detection bits. In addition, the recovery bits are generated according to the average intensity of each block. The embedding technique uses the LSB method, where the three LSBs are utilized for watermark embedding.
Qian et al. [11] performed non-overlapping block divisions on an image. These blocks are classified into four types according to their characteristics. For each classified block, a quantization process is applied to the DCT coefficients. These quantized coefficients are embedded as a recovery watermark, and a hash function MD5 obtains the authentication bits. The embedding process is performed by the LSB method. Finally, to recover the corrupted regions after the recovery process, an inpainting technique is applied.
Qin et al. [12] divided the image into non-overlapping 8 × 8 pixel blocks, classified into textured and smooth blocks. The recovery bits are obtained using vector quantization (VQ) indexing and inpainting processes, where VQ indexing is used for each complex block. Finally, both VQ indexing and inpainting are used for each smooth block.
Li et al. [13] proposed a scheme for the protection of biometric images based on salient region detection, where each biometric image is divided into salient regions (ROIs) and background regions (ROBs). Additionally, each image is divided into blocks of 4 × 4 pixels, and for each block, 16-bit authentication is performed using MD5. Eigenface coefficients obtain the self-recovery watermark, and the embedding process is based on the spatial domain, where operations between the LSB bits and watermark bits are performed.
Jie et al. [14] divided an image into non-overlapping blocks of 2 × 2 pixels, where six recovery bits and two key-based data bits are produced using an averaging process of the MSB pixels. Additionally, to improve the recovered image quality, a 3 × 3 block neighborhood is used to recover each tampered block whose feature hidden in another block is corrupted. The number of blocks in a 3 × 3 block neighborhood of the test block that is inconsistent with their mapping block is obtained to validate an image block. The embedding process is performed by the LSB method.
Dadkhah et al. [15] used singular value decomposition (SVD) to generate bits for tamper detection and self-recovery. The image is divided into non-overlapped blocks of 4 × 4 pixels, and these blocks are divided into blocks of 2 × 2 pixels. For each block of 4 × 4 pixels, 12 bits of authentication are generated, and for each block of 2 × 2 pixels, 20 bits are generated for the recovery process. The embedding scheme is based on the LSB method, where watermarks are embedded into the two LSB bits of each pixel.
Fan [23] proposed an improvement of the reconstruction scheme based on watermarking and the set partitioning in hierarchical trees (SPIHT) transform, where this scheme uses blocks of 32 × 32 pixels to generate the reconstruction bits. Additionally, this algorithm incorporates two versions of the recovery bits to obtain better robustness against the tampering coincidence problem. This scheme’s main disadvantage is that the SPIHT transform can result in wrong recovery bits extracted for a block. Therefore, a complete block can be recovered with inferior quality, resulting in poor objective quality of the reconstructed image.
Tai et al. [24] proposed a scheme based on the integer wavelet transform (IWT), where the recovery and authentication bits are generated using this transformation, and the embedding scheme is based on LSB. This scheme does not embed the redundancy in data, and as a result, suffers the drawback that it cannot avoid the tampering coincidence problem.
Chamlawi et al. [25] proposed a system based on a semi-fragile watermark in the discrete wavelet transform (DWT) domain. The generation of watermarks is performed by DWT–DCT, where the low frequency (LL) sub-band is obtained, followed by the application of DCT for non-overlapped blocks. Finally, a quantization step to get the self-recovery watermark is used. The authentication watermark is a binary watermark correlated with the LL sub-band of the first DWT decomposition level. This method’s principal drawback is its low self-recovery rate because the embedding stage for the recovery bits is performed at the second decomposition level of the DWT in the middle-frequency sub-bands.
In [26], a semi-fragile watermark was employed using the IWT domain, where self-recovery bit generation was performed utilizing integer discrete cosine transformation (IDCT), Huffman coding and an error-correcting procedure based on the Bose–Chaudhuri–Hocquenghem (BCH) encoder. The authentication bits were obtained using the exclusive addition operation between the LL frequency sub-band and a pseudo-random sequence.
Chamlawi et al. [27] used a semi-fragile watermark employing the IWT, where the authentication bits are embedded in the LL sub-band; in the opposite, the self-recovery bits are embedded in the high-low (HL) and low-high (LH) frequency sub-bands. The self-recovery bits are generated by IWT, DCT and a quantization process to obtain the DCT coefficients. Additionally, the tamper detection bits are caused by a bitwise exclusive OR operation between a pseudo-random sequence and a binary matrix.
In a recent study [36], we have proposed a novel method performed on an image in luma and chroma (YCbCr) color space, implementing the halftoning algorithm in the luminance channels, thus obtaining recovery bits. Additionally, this scheme uses the procedure proposed in [15], where for chrominance channels, the watermark bits are also generated for recovery purposes. Additionally, three copies of these bits were embedded using the 2-LSB method, incrementing the robustness, and in such a way resolving the tampering coincidence problem. Contrarily, this study’s proposition consists of the SR-HTR method, which uses a block-based method for the generation of recovery and authentication watermarks employing an RGB color image. The novel framework is designed in three variants (1-LSB, 2-LSB and 3-LSB), where for each variant, there can be embedded three, six and nine copies of the recovery watermark, respectively. In opposition to the scheme proposed in this study, the framework designed in [36] generates a bit rate of 1.75 bits per pixel (bpp) for recovery, and 0.25 bpp for authentication purposes that, as can be seen below, permit one to obtain a sufficiently good performance in the reconstruction of image areas with tampering rates from 10% to 40%. The principal limitation of the framework [36] consists of embedding up to three copies of the recovery watermark because it generates 4 bits of authentication for each block of 4 × 4 pixels. Therefore, it cannot obtain good objective and subjective performance in color image reconstruction for high tampering rates (from 40% to 90%). On the contrary, the novel framework, which generates 0.9375 bpp for recovery and 0.0625 bpp for authentication purposes (one bit for each block of 4 × 4 pixels), can obtain much more redundancy in watermarks used during the recovery process, increasing the robustness of the scheme. The novel SR-HTR framework appears to demonstrate excellent objective performance and subjective visual perception via the HVS in color image reconstruction for the highest tampering rates (from 40% to 90%).
Chia-Chen Lin et al. [37] proposed a self-embedding fragile watermarking scheme for grayscale images that uses a reference matrix as an embedding method. Each non-overlapping pixel pair is used as a coordinate in a reference matrix to embed recovery information, and each provides a 4-bit capacity as hidden space. Tampered regions are detected and restored by comparing the embedded recovery information in two LSBs of the original image with the recovery information generated from suspicious images.
In [38], a self-embedding watermarking method based on absolute moment block truncation coding (AMBTC) for grayscale images is presented. A checksum was introduced for accurate block authentication. They use the optimal pixel adjustment process (OPAP) method for embedding the recovery bits in block units in the second and third LSB and the checksum bits in the first LSB. The proposed method does not support irregular area alterations and has the limitation that if the modified region of the image is large (more than 45%), the restoration information is also removed, so that area cannot be restored.
Chin-Feng Lee et al. [39] performed a self-recovery fragile watermarking authentication scheme for grayscale images, based on two authentication methods: blockwise and pixelwise. In the first one, authentication data are generated from each block. The average block value is then used to produce recovery data; if the block’s size is small, the false positive rate (FPR) will be reduced. For the second one, the authentication data are generated from each pixel, and the recovery data are obtained from the mean value of the block. When tampered pixels are detected, the corresponding pixel area is marked. Consequently, the FPR will be lower than blockwise detection. Since the proposed method uses blocks to detect tampered blocks, when multiple areas of the watermarked image are modified, the tampering coincidence problem significantly reduces the proposal’s performance.
The majority of the previously revised methods have the principal disadvantage that the reconstructed image is low quality when significant tampering has occurred. Therefore, as mentioned above, to avoid the tampering coincidence problem, it is necessary to embed the recovery watermark on more than one occasion, as has been performed in several schemes, where two copies [10,17,23] or three copies [36] of the recovery watermark are embedded; therefore, additional chances for recovery are provided.
The main contributions of the proposed SR-HTR can be summarized as follows:
1. High quality of the watermarked image. A bit adjustment procedure is performed after the embedding process, resulting in an increased quality of a watermarking image. Several state-of-the-art methods show slightly better objective criteria values in comparison with the novel SR-HTR framework, but the detailed analysis confirms that the novel scheme appears to demonstrate acceptable objective perception quality in watermarked images as well as imperceptibility in visual subjective analysis via the HVS.
2. Better quality and robustness against the tampering coincidence problem in comparison with state-of-the-art schemes. The proposed method generates a highly compressed digest watermark, which can be embedded using three, six or nine copies of recovery watermarks, achieving higher tamper detection capability as opposed to other methods that can embed up to three times the digest image generated. The novel SR-HTR framework demonstrates the ability to reconstruct color images with high tampering rates, resulting in excellent objective and subjective performances that can be appreciated in the visual results presented below.
3. Tampering detection accuracy. During the authentication and reconstruction stage, the employed redundancy results in better tampering detection accuracy. The hierarchical tampering detection process, employed in the novel scheme, results in a decrease in false negatives during authentication and a higher manipulation detection. Better quality in the recovered image demonstrated by the proposed SR-HTR framework compared with state-of-the-art schemes is achieved via the implementation of a particular phase where the regions affected by the tampering coincidence problem should be detected. Afterward, such detected areas are processed using the inpainting method, demonstrating better quality of the recovered image in objective criteria values as well as in subjective visual perception via the HVS.

3. Designed Scheme

The designed method is divided into two stages. The first stage consists of the image protection process, which allows the insertion of multiple copies (three, six or nine) of the reconstruction watermark. This algorithm is presented in Figure 1. The second stage contains the authentication and reconstruction process, which uses a hierarchical authentication and an inpainting method to enhance the reconstruction’s performance. The diagram of this stage can be observed in Figure 2.

3.1. Image Protection

The image protection stage is described in this section. This process follows two steps: watermarking generation and insertion in the carrier image. In order to explain these steps, let us denote the original image as I h of size A × B in the color space RGB.

3.1.1. Watermarking Generation for Reconstruction and Authentication Purposes

The original image has three channels according to its color space. Consequently, Pseudocode 1 is applied on each one of the channels, generating three watermarks for reconstruction and three for authentication. The watermarks used for reconstruction were designated as w r , w g and w b , and the ones utilized for authentication were named a u t r , a u t g and a u t b for the channels R, G and B, correspondingly.
Pseudocode 1 Recovery and authentication watermark generation
Input: Image to be processed Ih
[ A , B ] = s i z e ( I h )
i R e f e r e n c e = i m a g e R e s i z e ( I h , 0.25 ) i m a g e   s u b s a m p l i n g
i R e f e r e n c e = b i t A N D ( i R e f e r e n c e , 248 )   r e p l a c e   3   L S B s   b y   0
r e c o v e r y W = [ ]   r e c o v e r y   w a t e r m a r k
a u t e n t W = [ ] a u t h e n t i c a t i o n   w a t e r m a r k
For i = 1   to A / 4   do
For j = 1   to B / 4   do
t m p W = g e t 5 M S B ( i R e f e r e n c e i , j )   e x t r a c t   5   M S B s
r e c o v e r y W = c o n c a t ( r e c o v e r y W , t m p W )   c o n c a t e n a t i o n   p r o c e s s
a u t   =   X O R ( X O R ( t m p W 1 , t m p W 2 ) , X O R ( t m p W 3 , t m p W 4 ) )
a u t   =   X O R ( a u t   , t m p W 5 )
a u t e n t W = c o n c a t ( a u t e n t W , a u t )   c o n c a t e n a t i o n   p r o c e s s
End for
End for
Output: Recovery watermark recoveryW; Authentication watermark autentW
Once the watermarks are obtained, a subsampled process with a factor of 0.25 is performed to significantly reduce the number of bits representing each channel of the original image. When 5 MSB of each pixel are extracted, a total number of 5 A B / 16 bits are available for reconstruction watermark embedding, and A B / 16 bits can be used for authentication watermarks. The authentication bit generation process consists of applying the bitwise XOR operation in the 5 MSBs, where a single bit is generated for each pixel block of size 4 × 4.

3.1.2. Watermark Embedding

The embedding process is described in Pseudocode 2, where w r , w g , w b and an authentication watermark a u t r , a u t g or a u t b are embedded in a selected channel of I h . Firstly, a random permutation of the reconstruction watermarks utilizes a seed of the user key, which must be different for each processed RGB channel or bit plane.
The watermark embedding process employs the 1-LSB method to embed the watermarks for reconstruction and a single watermark for authentication. This is possible due to the size of each one of the reconstruction watermarks, which is 5AB/16, and because a single authentication watermark is composed of AB/16 bits.
Finally, a bit adjustment process is applied, where each pixel in the i,j-th position of each RGB channel of the host image, denoted as I h w , is compared with the pixel in the same position of I h . This comparison has the objective to modify the intensity value of the pixels in I h w to enhance its objective quality. This process depends on the total number of marked bit planes. For each watermarked pixel, the following equation is used:
I h w i , j = { I h i , j 1 , i f   v   i s   ( 2 N 1 )   a n d   LSB N + 1 ( I h i , j )   i s   1 I h i , j + 1 ,   i f   v   i s   ( 2 N 1 )   a n d   LSB N + 1 ( I h i , j )   i s   0   ,
where N represents the watermarked bit plane and v = I h w i , j I h i , j for 1 i A , and 1 j B . This equation can be used only for values N > 2 .
Pseudocode 2 Watermark embedding
Input: Image (single channel) to process Ih; Seed S; Bit plane bit Plane; Recovery watermarks wr, wg and wb; Authentication watermark aut
[ A , B ] = s i z e ( I h )
r n g ( S )   C o n t r o l   r a n d o m   n u m b e r   g e n e r a t o r
n u m R a n d = r a n d p e r m ( A B / 16 ) R a n d o m   p e r m u t a t i o n   o f   i n t e g e r s   i n   r a n g e   [ 1 , A B / 16 ]
n u m = 1
For i = 1   step 4 to A 3   do
For j = 1   step 4 to B 3   do
s u b i n d e x = 5 * n u m R a n d [ n u m ]
i n d e x = s u b i n d e x 4 : s u b i n d e x n u m b e r s   f r o m   ( s u b i n d e x 4 )   t o   s u b i n d e x
t m p W = c o n c a t ( w r i n d e x , w g i n d e x , w b i n d e x , a u t n u m )   c o n c a t e n a t i o n   t m p W   c o n t a i n s   16   b i t s
I h w i : i + 3 , j : j + 3 = e m b e d ( I h i : i + 3 , j : j + 3 , b i t P l a n e , t m p W ) L S B   e m b e d d i n g   o f   t m p W   i n   I h   b y   b i t P l a n e
n u m =   n u m + 1
End for
End for
Equation   ( 1 ) ,   w h e r e   N = b i t   P l a n e
Output: Watermarked image Ihw

3.2. Authentication and Reconstruction

This section describes the authentication and reconstruction process of a given tampered image I h w of size A × B in RGB color space. To accomplish this reconstruction, four steps are used: extraction of the watermarks from the image, authentication of the content, post-processing to indicate the blocks affected by the tampering coincidence problem via an inpainting process to fill these regions and reconstruction of the tampered image.

3.2.1. Watermark Extraction

The extraction of the three reconstruction watermarks is performed as follows:
a u x V a l i   =   j = 1 5 2 8 j w t m 5 ( i 1 ) + j ,     i   s . t .     1 i A B / 16 ,  
where w t m is a vector form of a watermark w r , w g or w b . The process to acquire all the watermarks is detailed in Pseudocode 3, which employs the function v e c 2 m a t ( a u x V a l , A / 4 , B / 4 ) to transform a u x V a l in a matrix of size A / 4   × B / 4 .
This pseudocode is applied to each channel image, where three reconstruction images in the RGB color space ( i R G B R , i R G B G , i R G B B ) and three authentication images ( a u t R , a u t G , a u t B ) to authenticate each RGB channel of I h w are obtained.
Pseudocode 3 Extraction of image watermarks
Input: Watermarked image channel Ihw; Seed S; Bit plane bit Plane
[ A , B ] = s i z e ( I h w )
r n g ( S )   C o n t r o l   r a n d o m   n u m b e r   g e n e r a t o r
n u m R a n d = r a n d p e r m ( A B / 16 )   R a n d o m   p e r m u t a t i o n   o f   i n t e g e r s   i n   r a n g e   [ 1 , A B / 16 ]
w r = [   ] E m p t y   l i s t   o f   s i z e   5 A B / 16
w g = [   ] E m p t y   l i s t   o f   s i z e   5 A B / 16
w b = [   ] E m p t y   l i s t   o f   s i z e   5 A B / 16
a u t = [   ] E m p t y   l i s t   o f   s i z e   A B / 16
n u m = 1
For i = 1   step 4 to   A 3 do
For j = 1   step 4 to B 3   do
s u b i n d e x = 5 * n u m R a n d [ n u m ]
i n d e x = s u b i n d e x 4 : s u b i n d e x   n u m b e r s   f r o m   ( s u b i n d e x 4 )   t o   s u b i n d e x
t m p W = e x t r a c t ( I h w i : i + 3 , j : j + 3 , b i t P l a n e )   e x t r a c t   16   b i t s   f r o m   t h e   b i t P l a n e L S B
n u m = n u m + 1
End for
End for
Equation   ( 2 ) ,   w h e r e   w t m = w r
i m g W t m = v e c 2 m a t ( a u x V a l , A / 4 , B / 4 )   v e c t o r   t o   m a t r i x   c o n v e r s i o n
i R G B ( : , : , 1 ) = i m r e s i z e ( i m g W t m , 4 )
Equation   ( 2 ) ,   w h e r e   w t m = w g
i m g W t m = v e c 2 m a t ( a u x V a l , A / 4 , B / 4 )   v e c t o r   t o   m a t r i x   c o n v e r s i o n
i R G B ( : , : , 2 ) = i m r e s i z e ( i m g W t m , 4 )
Equation   ( 2 ) ,   w h e r e   w t m = w b
i m g W t m = v e c 2 m a t ( a u x V a l , A / 4 , B / 4 )   v e c t o r   t o   m a t r i x   c o n v e r s i o n
i R G B ( : , : , 3 ) = i m r e s i z e ( i m g W t m , 4 )
Output: Recovery image iRGB; Authentication watermark aut

3.2.2. Authentication

This step uses the previously described Pseudocode 1 to generate the bit sequence a u t e n t W   from each channel of   I h w , where a u t e n t W R , a u t e n t W G and a u t e n t W B are obtained. Each bit sequence is compared with each authentication watermark   a u t R , a u t G and a u t B resulting from Pseudocode 3 using the following equation:
a u t e n t I m g i , j = { 255 ,   i f   a u t e n t W j + ( i 1 ) B 4   i s   n o t   a u t j + ( i 1 ) B 4 0 , o t h e r w i s e ,
for all i and j subject to 1 i A / 4 and 1 j B / 4 . Each   a u t e n t I m g is then interpolated to size A × B .
Once the previous steps have been performed, three reconstruction images ( i R G B R , i R G B G and i R G B B ) and three authentication images ( a u t e n t I m g R , a u t e n t I m g G and a u t e n t I m g B ) are generated for each RGB channel and each LSB plane. A general authentication image is then computed by applying the bitwise OR operand to each of the authentication images, i.e.,   i A u t e n t i , j   =   a u t e n t I m g R i , j   +   a u t e n t I m g G i , j   +   a u t e n t I m g B i , j ,   i , j ,   s . t .   1     i     A ,   1     j     B . A point worth mentioning is that this operation is valid only for the 1-LSB embedding method. If the 2-LSB method is utilized, six watermarks for authentication and six for reconstruction should be generated. Analogously, the 3-LSB process requires nine watermarks for each case. Therefore, i A u t e n t is obtained by applying the bitwise OR operation to the six or nine authentication images for the 2-LSB or 3-LSB method, respectively. Finally, the first level hierarchical authentication is performed on i A u t e n t , improving the tamper detection recognition.

3.2.3. Post-Processing and Recovery

A binary image is generated at this stage, representing the blocks affected by the tampering coincidence problem in each reconstruction image generated by Pseudocode 3 ( i R G B R , i R G B G , i R G B B ). Each authentication image generated in Section 3.2.2 ( a u t e n t I m g R , a u t e n t I m g G , a u t e n t I m g B ) is used for this process.
Pseudocode 4 presents the image generation process as follows:
a u t e n t n u m r a n d [ j + ( i 1 ) B / 4 ] = { 255 ,   i f   a u t e n t I m g i , j   i s   255 0 , o t h e r w i s e ,  
for all i and j subject to 1 i A / 4 , and 1 j B / 4 . Finally, the generated image is interpolated to   A × B   size. Pseudocode 4 presents the processes, resulting in three images being obtained, denoted as   T C P R , T C P G and T C P B .
If a large amount of the original image information is altered, the information used for reconstruction could be overwritten, although it is redundant. Therefore, maps to point to the tampering coincidence problem are computed using binary images (Pseudocode 4). Then, an AND operand between each one of these maps’ bits is applied, resulting in a binary image denoted as i T C P , i.e., i T C P i , j   =   T C P R i , j * T C P G i , j * T C P B i , j ,     i , j ,   s . t .   1     i     A ,   1     j     B . This image marks the regions affected by the tampering coincidence problem.
Pseudocode 4 Detection of tampering coincidence problem
Input: Authentication image autent Img; Seed S
[ A , B ] = s i z e ( a u t e n t I m g )
r n g ( S )   C o n t r o l   r a n d o m   n u m b e r   g e n e r a t o r
n u m R a n d = r a n d p e r m ( A B / 16 )   R a n d o m   p e r m u t a t i o n   o f   i n t e g e r s   i n   r a n g e   [ 1 , A B / 16 ]
a u t e n t I m g = i m r e s i z e ( a u t e n t I m g , 0.25 )
Equation   ( 4 )
T C P = v e c 2 m a t ( a u t e n t , A / 4 , B / 4 )   v e c t o r   t o   m a t r i x   c o n v e r s i o n
T C P = i m r e s i z e ( T C P , 4 )
Output: Tampering coincidence problem image TCP
Subsequently, Equation (5) is applied to the reconstruction images obtained by Pseudocode 3 and their corresponding binary images given by the Pseudocode 4, resulting in a single reconstruction image named i R :
i R i , j = α = 1 n ( 1 T C P α i , j ) I w α i , j α = 1 n ( 1 T C P α i , j ) ,   i , j   :   1 i A ,   1 j B ,
where α represents the i-th processed copy, I w represents a recovery image, T C P represents a binary image, which indicates the regions affected by the tampering coincidence problem of I w , where the values “1” indicate that the i, j-th position is authentic and a value of “0” indicates that this position was affected by this problem. Implementation of Equation (5) is given for T C P = [ T C P R , T C P G , T C P B ] and I w = [ i R G B R , i R G B G , i R G B B ] .
The previous sequence is valid when the 1-LSB method is selected for embedding. If the scheme is presented in their variants 2-LSB or 3-LSB, further copies of the watermarks should be computed from the bit panels, and the reconstruction method requires single instances of i R and i T C P . Consequently, the following equations for the N -LSB method must be applied:
i T C P   = { i T C P 1 , i f   N   i s   1 i T C P 1   *   i T C P 2 , i f   N   i s   2 i T C P 1   *   i T C P 2 *   i T C P 3 , i f   N   i s   3 ,
i R   = { i R 1 , i f   N   i s   1 Equation   ( 5 ) ,   where   I w = [ i R 1 , i R 2 , i R 2 ]   and   T C P = [ i T C P 1 , i T C P 2 , i T C P 2 ] ,   i f   N   i s   2 Equation   ( 5 ) ,   where   I w = [ i R 1 , i R 2 , i R 3 ]   and   T C P = [ i T C P 1 , i T C P 2 , i T C P 3 ] ,   i f   N   i s   3 ,
Finally, Pseudocode 5 performs an inpainting method for the given i R and i T C P images. The output image is named as i R e c o v e r y . The inpainting method divides the i R and i T C P images into overlapping blocks of 3 × 3 pixels, and the following equations are used for each block:
i R i , j = a = 1 3 b = 1 3 ( w I o . * ( 1 w T C P ) ) a , b a = 1 3 b = 1 3 ( 1 w T C P a , b ) , i f   w T C P 2 , 2   i s   1   a n d   i = 1 3 j = 1 3 ( 1 w T C P i , j ) >   1 ,
i T C P i , j = 0 ,   i f   w T C P 2 , 2   i s   1   a n d   i = 1 3 j = 1 3 ( 1 w T C P i , j ) >   1 ,
where w T C P represents a block of the i T C P image, which is subject to w T C P   = i T C P i 1 : i + 1 , j 1 : j + 1 , and w I o is a block of the i R image, given w I o   = i R i 1 : i + 1 , j 1 : j + 1 , for all i, j, such that 2 i A + 1 , 2 j B + 1 .
Finally, the tampered zones of the I h w image can be reconstructed by means of the following equation:
I h w i , j = I h w i , j ( 1 i A u t e n t i , j ) + (   i R e c o v e r y i , j )   ( i A u t e n t i , j )   ,   1 i A ,   1 j B .  
Pseudocode 5 Inpainting application
Input: Image to be processed iR; Binary image iTCP
[ A , B ] = s i z e ( i R )
W h i l e   i = 1 A j = 1 B i T C P i , j   ! =   0   d o
i R   = [ i R 1 , 1 i R 1 , : i R 1 , B i R : , 1 i R : , : i R : , B i R A , 1 i R A , : i R A , B ] M a t r i x   o f   s i z e   A + 2 , B + 2
i T C P = [ i T C P 1 , 1 i T C P 1 , : i T C P 1 , B i T C P : , 1 i T C P : , : i T C P : , B i T C P A , 1 i T C P A , : i T C P A , B ] M a t r i x   o f   s i z e   A + 2 , B + 2
i T C P =   i T C P > 127
Equations   ( 8 )   and   ( 9 )
i T C P =   i T C P * 255
i R   = i R   2 : A 1 , 2 : B 1 M a t r i x   o f   s i z e   A , B
i T C P = i T C P 2 : A 1 , 2 : B 1 M a t r i x   o f   s i z e   A , B
End While
i R e c o v e r y =   i R
Output: Image after inpainting process iRecovery

3.3. Implementation of the Algorithms

The proposed SR-HTR method allows the insertion of N copies of the three reconstruction watermarks for the N -LSB method limited by 1 N 3 . However, the algorithm implementation has to be changed depending on the parameter N , as can be observed in Pseudocodes 6 and 7.
Pseudocode 6 Image protection
Input: Image (RGB) to be protected Ih; LSB plane N
w r ,   a u t r   =   Pseudocode   1   ( I h ( : , : , 1 ) )
w g ,   a u t g   =   Pseudocode   1   ( I h ( : , : , 2 ) )
w b ,   a u t b   =   Pseudocode   1   ( I h ( : , : , 3 ) )
I h w   =   I h
F o r   i   = 1   s t e p   1   t o   N   d o
I h w ( : , : , 1 ) =   Pseudocode   2   ( I h w ( : , : , 1 ) , 10 + i , i , w r , w g , w b , a u t r )
I h w ( : , : , 2 ) =   Pseudocode   2   ( I h w ( : , : , 2 ) , 20 + i , i , w r , w g , w b , a u t g )
I h w ( : , : , 3 ) =   Pseudocode   2   ( I h w ( : , : , 3 ) , 30 + i , i , w r , w g , w b , a u t b )
E n d   F o r
Output: Watermarked image Ihw
Pseudocode 7 Image authentication and recovery
Input: Suspicious image (RGB) Ihw; LSB plane N
F o r   i   = 1   s t e p   1   t o   N   d o
i R G B R i ,   a u t R i   =   Pseudocode   3   ( I h w ( : , : , 1 ) , 10 + i , i )
i R G B G i ,   a u t G i   =   Pseudocode   3   ( I h w ( : , : , 2 ) , 20 + i , i )
i R G B B i ,   a u t B i   =   Pseudocode   3   ( I h w ( : , : , 3 ) , 30 + i , i )
E n d   F o r
~ ,   a u t e n t W R   =   Pseudocode   1   ( I h w ( : , : , 1 ) )
~ ,   a u t e n t W G   =   Pseudocode   1   ( I h w ( : , : , 2 ) )
~ ,   a u t e n t W B   =   Pseudocode   1   ( I h w ( : , : , 3 ) )
i A u t e n t   =   z e r o s ( A , B )
F o r   i   = 1   s t e p   1   t o   N   d o
Equation   ( 3 ) ,   w h e r e   a u t e n t W = a u t e n t W R   a n d   a u t =   a u t R i
a u t e n t I m g R i = i m r e s i z e ( a u t e n t I m g , 4 )
Equation   ( 3 ) ,   w h e r e   a u t e n t W = a u t e n t W G   a n d   a u t =   a u t G i
a u t e n t I m g G i = i m r e s i z e ( a u t e n t I m g , 4 )
Equation   ( 3 ) ,   w h e r e   a u t e n t W = a u t e n t W B   a n d   a u t =   a u t B i
a u t e n t I m g B i = i m r e s i z e ( a u t e n t I m g , 4 )
  A i   =   a u t e n t I m g R i + a u t e n t I m g G i + a u t e n t I m g B i OR   operation
i A u t e n t   =   i A u t e n t   +   A i OR   operation
E n d   F o r
i A u t e n t   =   h i e r a r c h i c a l _ a u t h e n t i c a t i o n ( i A u t e n t )
F o r   i   = 1   s t e p   1   t o   N   d o
T C P R I =   Pseudocode   4   ( a u t e n t I m g R i ,   10 + i )
T C P G I =   Pseudocode   4   ( a u t e n t I m g G i ,   20 + i )
T C P B I =   Pseudocode   4   ( a u t e n t I m g B i ,   30 + i )
Equation   ( 5 ) ,   w h e r e   I w = [ i R G B R i , i R G B G i ,   i R G B B i ]   a n d   T C P = [ T C P R I , T C P G I , T C P B I ]  
   i T C P i   =   T C P R I   * T C P G I   *   T C P B I AND   operation
E n d   F o r
Equations   ( 6 )   and   ( 7 )
i R e c o v e r y   =   Pseudocode   5   ( i R ,   i T C P )
Equation   ( 10 )
Output: Restored image Ihw

4. Experimental Setup

Images of sizes 512 × 768 and 768 × 512 from the Kodak database [40], which consists of 24 images, were used for experimentation. These images are labeled as Kodak-n, where n is the image identifier. Some of these images are shown in Figure 3.
The performance criteria to evaluate the quality of watermarked and recovered images obtained by the proposed framework are: PSNR, SSIM and PSNR-HVS-M [35]. The PSNR metric is defined as follows:
P S N R = 10   l o g 10 M A X 2 M S E ,
M S E = 1 X Y i = 0 X 1 j = 0 Y 1 [ I ( i , j ) I ( i , j ) ] 2 ,
where I represents the original image, I corresponds to the modified image, the arguments i , j are used for the pixel position and X ,   Y are the number of rows and columns, correspondingly. The SSIM metric is computed using the following equation:
S S I M ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 ) ,
where μ and σ denote the mean and variance of images x and y , σ x y is the covariance between x and y , constants C 1 , C 2   are: C 1 = ( 0.01 L ) 2 ,   C 2 = ( 0.03 L ) 2 , and L = 255 [41].
Furthermore, Precision and Recall metrics were utilized to measure the alteration detection performance of the proposed method. These metrics are based on the numbers of true positives ( T P ), false positives ( F P ) and false negatives ( F N ) among all pixels:
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N .

5. Analysis of 1-LSB, 2-LSB and 3-LSB Schemes in Embedding Stage

A comparison between the different LSB variants of the designed framework was performed. The watermarked images were compared with the original ones, and the results are reported in Table 1, which shows the average values of the objective quality measures PSNR, SSIM and PSNR-HVS-M for each variant. Furthermore, the comparison was also performed using the bit adjustment for 2-LSB and 3-LSB methods. It can be observed that a minimal enhancement is generally achieved using this adjustment.
During the imperceptibility evaluation of the three variants (1-LSB, 2-LSB or 3-LSB) shown in Table 1, we can observe that the 1-LSB and the 3-LSB variants have the best and the worst performance, respectively, in terms of objective quality evaluation. However, a point worth mentioning is that for the worst case (3-LSB), the embedding of nine copies of the watermark does not produce any recognizable visual modification in the watermarked image.
To measure the authentication and reconstruction process’s performance, experiments changing the tampering rate from 10% to 90% in the image were carried out, adding pseudo-random noise to the image, as can be observed in Figure 4.
The authentication stage evaluation is reported in Table 2, where it can be observed that the 1-LSB method maintains the best results in Precision. In terms of Recall, the proposed variants’ performance increases when this measure is close to one. The variants 2-LSB and 3-LSB have the same high Recall value, even though the 3-LSB variant embeds more copies. Additionally, the probability that the 1-LSB variant extracts copies incorrectly is higher than the other variants, generating more errors during the authentication process. Another point worth mentioning is that the 2-LSB variant achieves Recall’s best performance, avoiding the tampering coincidence problem, embedding only six copies of the recovery watermark.
Finally, the average values of the quality measures PSNR, SSIM and PSNR-HVS-M for the reconstructed images compared with the original ones are shown in Table 3. A significant increase in the values with the 2-LSB method can be recognized. Nevertheless, although the 3-LSB implementation also showed a sharp increment compared with 1-LSB, its results are slightly lower than the 2-LBS method. On the contrary, the PSNR-HSV-M values of the 3-LSB method are slightly higher than the result from 2-LSB. However, the 2-LSB results are still acceptable, and, in general, their PSNR and SSIM values are superior to the other implementations in this study. Consequently, the 2-LSB variation was utilized for the proposed SR-HTR method considering the results given in Table 1, Table 2 and Table 3.

6. Experimental Results and Discussion

Since the 2-LSB method maintains a balance between the quality of the marking, authentication and reconstruction processes, it was selected to insert the watermarks. In this section, the proposed method’s performance with the 2-LSB variation is detailed and compared with other state-of-the-art methods.

6.1. Watermarked Image Quality

Table 4 shows the results after the watermark embedding and the variation with and without the bit adjustment stage. It can be observed that the bit adjustment markedly raised the PSNR and SSIM values. Nonetheless, the best PSNR-HSV-M values fluctuated between both implementations.

6.2. Analysis of Tampering Detection

In order to measure the performance of the proposed method in change detection, different modifications were applied to the test watermarked images. Afterward, the authentication process was executed using the six authentication watermarks and the hierarchical authentication method to discard most of the detection errors, specifically the false negatives.
The tampering detection process evaluation consists of two alteration schemes: the first one was used to estimate the ability to detect alteration rates between 10% and 90% by adding a regular square area of pseudo-random noise into the watermarked image, as displayed in Figure 4. In the second scheme, the alterations were performed by modifying one or multiple irregular areas in the watermarked image using Adobe Photoshop software, maintaining the structure and original nature of the image and avoiding significant falsity in the alteration. Figure 5 illustrates the watermarked images with irregular alterations of Kodak-1 in 46.41%, Kodak-3 in 34.34%, Kodak-11 in 25.06% and Kodak-14 in 40.08% of the entire image.
The detection’s Precision results for the changes using the first scheme are shown in Table 5. It can be noticed that the Precision performance is enhanced as the alteration rate is increased. The Recall metric resulted in values of 1.0 for all the alteration rates between 10% and 80% and 0.9999 for a change of 90% of the image. This is due to the hierarchical method of authentication.
The evaluation of the second scheme, which uses multiple and irregular alterations in the images, is displayed in Figure 6, where images in Figure 6a–d represent the ground truth of the alterations, and images in Figure 6e–h are the results of the change detection obtained by SR-HTR. The pairs of (Precision, Recall) values are: (0.9190, 0.9995) for Kodak-1, (0.9416, 1.0) for Kodak-3, (0.9214, 1.0) for Kodak-11 and (0.9382, 0.9998) for Kodak-14.

6.3. Evaluation of the Reconstruction under Different Tampering Rates

In this section, the results for the evaluation of the reconstruction process are reported. The reconstruction of the test images was evaluated using both alteration schemes. For the first scheme, which is illustrated in Figure 4, the six versions of the reconstruction watermark are utilized for reconstruction. The inpainting method was performed to regenerate the information affected by the tampering coincidence problem.
Figure 7 shows the reconstructed images for Kodak-23. As can be noticed, the reconstruction capability for each alteration rate markedly dropped according to the increment in the alteration rate. Nevertheless, the original content of the image can be clearly distinguished.
Additionally, it can be observed that a granulated effect is obtained according to the alteration rate. The inpainting process gives this effect due to the replenishment of affected areas by the tampering coincidence problem, with neighboring pixels’ intensity values detected as authentic ones.
Finally, this process was executed on all the test images using the first alteration scheme. Figure 8, Figure 9 and Figure 10 report the PSNR, SSIM and PSNR-HVS-M results after the reconstruction process, correspondingly. These graphics show an average decrease of 12.72 dB for PSNR and 14.41 dB for PSNR-HSV-M in reconstructing the modification rates between 10% and 90% of the image and an average reduction of 0.3516 for SSIM values. The results can be considered acceptable given the high alteration rate to which the images are subjected. A point worth mentioning is that other works usually test their methods with an upper boundary of 50% of modification since they do not consider the tampering coincidence problem consequences in their entirety.

6.4. Evaluation of the Image Reconstruction under Multiple and Irregular Attacks

The results of the reconstructed images that were modified using the second alteration scheme, given in Figure 5, are presented in this section. Figure 11 shows a satisfactory visual quality of the reconstructed images, which display the original content substituted with other information.

6.5. Comparison with State-of-the-Art Schemes

The proposed SR-HTR method was compared in terms of performance with other state-of-the-art methods. As previously mentioned, the design of SR-HTR aims to achieve high performance when there is a high rate of modification in a watermarked image. This is accomplished by the insertion of redundancy in the authentication and reconstruction bits. Therefore, the comparison is performed in terms of objective quality in the following areas: watermarked image visualization, change detection rates and reconstruction image visualization.
The first evaluation consists of the comparison between the watermarked image and the original one. Table 6 shows the average values of PSNR, SSIM and PSNR-HVS-M for the set of test images employed.
The novel SR-HTR method and Molina [36] present the best quality results in PSNR and SSIM. This is due to the bit adjustment stage implemented in 2-LSB for both methods. For the PSNR-HVS-M metric, the highest values belong to [23,24], followed by the proposed method, [36], and then by [2,10]. A point worth mentioning is that the bit adjustment of SR-HTR and [36] negatively affects the results of PSNR-HSV-M due to the characteristics of this metric, as described in Table 1 and Table 4. Furthermore, it is essential to notice that the methods [23,24,36] employ the 2-LSB method, and procedures [2,10] use 3-LSB insertion. Additionally, methods [23,24] do not exploit the total insertion capacity to use enough redundant information in the watermarks, leading to lower performance in the reconstruction stage.
The second evaluation is related to change detection. For this test, Precision and Recall metrics were employed to compare the performance of the alteration schemes previously described.
The Precision results of this evaluation using the first alteration scheme are shown in Table 7. The best results belong to [2,10]. Compared with the other methods, these methods possess an authentication scheme that generates multiple authentication bits for each block of n × n pixels. Consequently, there is a higher likelihood to detect an alteration in a block because the number of bits to be compared increases. However, the proposed method was designed to generate a single authentication bit for each block of 4 × 4 pixels, and this significantly reduces the detection capability.
Figure 12 illustrates the images utilized for the Precision and Recall comparison with the second alteration scheme, which considers the modification of multiple and irregular areas. Moreover, Table 8 shows the average Precision values of SR-HTR and the other state-of-the-art methods for the attacks given in Figure 12.
Again, it can be observed that methods [2,10] achieved a higher performance in this test by employing a more significant number of authentication bits. These results are followed by the method [24], then the designed SR-HTR, [36] and finally [23]. Nonetheless, the proposed SR-HTR method accomplished Precision values higher than 0.9. It is further demonstrated that the novel SR-HTR method maintains a better balance between authentication and reconstruction capabilities.
Regarding the Recall measure, all the methods presented a value of 1.0 for both alteration schemes, including the novel SR-HTR method using the hierarchical authentication method.
To perform a visual comparison, the image Kodak-15 was employed. Its modifications can be observed in Figure 13. This comparison is illustrated in Table 9 and Table 10, where Table 9 presents the results for the alteration rates between 20% and 50%, and Table 10 shows the results for alteration rates between 60% and 90%.
It can be noticed in Table 9 that the designed SR-HTR method presents a better visual quality for alteration rates between 20% and 50%, followed by the procedure [36] that inserts three versions of the reconstruction watermark, and then by [10] that inserts only two. The methods [2,23,24] reconstruct the original content but with a significant loss of quality in the reconstructed image due to the insertion of a single instance of the recovery watermark.
The results given in Table 10 show a dramatic fall in quality for the state-of-the-art methods, while SR-HTR maintains an acceptable visual quality proficiency until 90% when the Kodak-15 image is modified. The second-best results belong to [36], which degrades the image quality but still visibly preserves the original content. This decrease in quality for the state-of-the-art methods is due to the low level of redundancy inserted for the reconstruction watermark, which corresponds to the insertion of up to two copies, such as in [10,23]. However, the framework [23] presents a larger number of contiguous pixels without reconstruction compared with [10] because it uses the SPITH compression algorithm with 32 × 32 pixel blocks to generate the reconstruction bits. This method is sensitive to noise applied to the blocks. Therefore, the decompression algorithm will result in wrong values for the whole block with minimal modification of the reconstruction bits.
The values of PSNR, SSIM and PSNR-HVS-M for the visual results shown in Table 9 and Table 10 are reported in Table 11. It can be observed that for alteration rates greater than 30%, the proposed SR-HTR method presents a better reconstruction performance in the metrics PSNR and PSNR-HSV-M. Furthermore, the process [36] provides a higher reconstruction performance for alteration rates lower than 30%. Regarding SSIM, the best results for 20% and 30% rates belong to [23] due to the SPITH transform utilized in the generation of the reconstruction bits, which allows a higher quality in reconstructing the blocks. Nonetheless, the SR-HTR method maintains an acceptable performance when more than 40% of the image is modified. It is important to emphasize that the reconstructed image through SR-HTR is not significantly affected by the growth in the alteration rate.
The results displayed in Table 12, Table 13 and Table 14 represent the average values of PSNR, SSIM and PSNR-HVS-M, correspondingly, for reconstructing the test images using the first alteration scheme. These results reflect the evaluation shown in Table 9, Table 10 and Table 11, where the proposed method demonstrates more balanced results between the change detection and the reconstruction process. Although a lower performance in change detection was obtained (as shown in Table 7 and Table 8), the reconstructed images’ quality drastically increased since the proposed method generates an authentication bit for each set of 15 reconstruction bits. The SR-HTR method aims to obtain a better performance in reconstructing large areas of modified pixels without considerably affecting the change detection process and the quality of the watermarked image, as one can see in Table 6. Subsequently, the method [36] achieves the second-best performance due to its insertion process that uses three copies of the watermarks. Methods [10,23] are placed in the third and fourth positions because of their intolerance to high alteration rates and because of the insertion of only two copies of the reconstruction bits. Finally, methods [2,24], which only insert a single version of the reconstruction bits, have demonstrated a lower performance due to their inability to solve the tampering coincidence problem.
The last evaluation is the comparison of the reconstruction proficiency of multiple and irregular alterations. This test was performed using the modifications illustrated in Figure 12. Their corresponded reconstructed images are shown in Table 15. The images’ alteration rates are Kodak-4 at 85.65%, Kodak-10 at 26.42%, Kodak-17 at 74.36% and Kodak-18 at 30.97%. The PSNR, SSIM and PSNR-HVS-M values of the reconstructed image compared with the original ones are reported in Table 16.
On one hand, it can be observed that the proposed SR-HTR method presents the best objective quality for each reconstructed image. On the other hand, the method given by [23] resulted in better SSIM values for Kodak-10 because the alteration rate is low. However, the visual quality of the reconstructed Kodak-10 given in Table 15 for the proposed method represents an admissible reconstruction.
As can be noticed in Table 9, Table 10, Table 11, Table 12, Table 13, Table 14, Table 15 and Table 16, the novel SR-HTR method demonstrated an excellent performance compared with the other state-of-the-art schemes in terms of regular attacks (as seen in Table 9, Table 10, Table 11, Table 12, Table 13 and Table 14) and multiple and irregular attacks (shown in Table 15 and Table 16).

7. Conclusions

This paper proposes a novel fragile scheme named SR-HTR based on watermarking for color image authentication and self-recovery with high tampering rates. The image protection and extraction method can be implemented in three different variants (1-LSB, 2-LSB or 3-LSB), where it is possible to embed multiple copies (three, six or nine, respectively) of the recovery watermarks, and thus to increase the robustness of the scheme to the tampering coincidence problem. The evaluation of the three embedding–extraction variants of the novel method was carried out. The embedding watermark scheme uses a pseudo-random sequence to embed the recovery watermarks in different blocks to increase the watermarked image’s objective quality.
During the evaluation of the results, various alterations were investigated at different tampering rates (from 10% to 90%) with irregular and multiple alterations. Finally, during the recovery process, the 2-LSB embedding scheme was selected as a balanced solution between the quality of the watermarked image and the recovered images’ quality for different alterations.
The experimental results have shown good quality for the watermarked images obtained by the novel SR-HTR framework compared with state-of-the-art methods. Additionally, the novel scheme has demonstrated a good performance during the detection of regular, irregular and multiple alterations, resulting in Precision and Recall metrics higher than 0.9.
The designed SR-HTR has shown excellent performance in reconstructing the alterations at different tampering rates (from 10% to 90%), which is superior to other state-of-the-art methods. Additionally, in cases of multiple and irregular alterations, the novel color image authentication and self-recovery framework has shown an excellent performance, maintaining high objective criteria values as well as great visual perception via the HVS.
As future work, further investigations should be performed to resist different intentional attacks like cropping, scaling and rotation, and unintentional attacks such as JPEG compression. Additionally, for fast processing in real-time environments, we will consider the possibility of designing parallel fragile watermarking schemes implemented in graphics processing units (GPU) or multicore central processing units (CPU), based on adversarial examples, which have demonstrated remarkable performance in solving similar problems in image and audio domains [42], avoiding weaknesses presented by deep neural networks (DNNs) where if an image is transformed slightly, will be incorrectly classified by a DNN even when the changes are small and unnoticed by the human eye [43].

Author Contributions

Conceptualization, R.R.-R., J.M.-G. and B.P.G.-S.; Formal analysis, C.C.-R.; Investigation, R.R.-R.; Methodology, R.R.-R., V.P. and J.M.-G.; Software, C.C.-R. and B.P.G.-S.; Visualization, C.C.-R. and J.M.-G.; Writing—Original draft, B.P.G.-S. and J.M.-G.; Writing—Review and editing, V.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The authors would like to thank the Instituto Politecnico Nacional (Mexico), Comision de Operacion y Fomento de Actividades Academicas (COFAA) of IPN and the Consejo Nacional de Ciencia y Tecnologia (Mexico) for their support in this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, C.; Wang, Y.; Ma, B.; Zhang, Z. A novel self-recovery fragile watermarking scheme based on dual-redundant-ring structure. Comput. Electr. Eng. 2011, 37, 927–940. [Google Scholar] [CrossRef]
  2. Singh, D.; Singh, S.K. Effective self-embedding watermarking scheme for image tampered detection and localization with recovery capability. J. Vis. Commun. Image Represent. 2016, 38, 775–789. [Google Scholar] [CrossRef]
  3. Wu, W.-C.; Lin, Z.-W. SVD-based self-embedding image authentication scheme using quick response code features. J. Vis. Commun. Image Represent. 2016, 38, 18–28. [Google Scholar] [CrossRef]
  4. Qi, X.; Xin, X. A quantization-based semi-fragile watermarking scheme for image content authentication. J. Vis. Commun. Image Represent. 2011, 22, 187–200. [Google Scholar] [CrossRef]
  5. Chang, C.-C.; Chen, K.-N.; Lee, C.-F.; Liu, L.-J. A secure fragile watermarking scheme based on chaos-and-hamming code. J. Syst. Softw. 2011, 84, 1462–1470. [Google Scholar] [CrossRef]
  6. Liu, S.-H.; Yao, H.-X.; Gao, W.; Liu, Y.-L. An image fragile watermark scheme based on chaotic image pattern and pixel-pairs. Appl. Math. Comput. 2007, 185, 869–882. [Google Scholar] [CrossRef]
  7. Chaluvadi, S.B.; Prasad, M.V.N.K. Efficient image tamper detection and recovery technique using dual watermark. In 2009 World Congress on Nature & Biologically Inspired Computing (NaBIC); Institute of Electrical and Electronics Engineers (IEEE): New York, NY, USA, 2009; pp. 993–998. [Google Scholar]
  8. Zhang, X.; Qian, Z.; Ren, Y.; Feng, G. Watermarking with Flexible Self-Recovery Quality Based on Compressive Sensing and Compositive Reconstruction. IEEE Trans. Inf. Forensics Secur. 2011, 6, 1223–1232. [Google Scholar] [CrossRef]
  9. He, H.; Chen, F.; Tai, H.-M.; Kalker, T.; Zhang, J. Performance Analysis of a Block-Neighborhood-Based Self-Recovery Fragile Watermarking Scheme. IEEE Trans. Inf. Forensics Secur. 2011, 7, 185–196. [Google Scholar] [CrossRef]
  10. Tong, X.; Liu, Y.; Zhang, M.; Chen, Y. A novel chaos-based fragile watermarking for image tampering detection and self-recovery. Signal Process. Image Commun. 2013, 28, 301–308. [Google Scholar] [CrossRef]
  11. Qian, Z.; Feng, G. Inpainting Assisted Self Recovery With Decreased Embedding Data. IEEE Signal Process. Lett. 2010, 17, 929–932. [Google Scholar] [CrossRef]
  12. Qin, C.; Chang, C.-C.; Chen, K.-N. Adaptive self-recovery for tampered images based on VQ indexing and inpainting. Signal Process. 2013, 93, 933–946. [Google Scholar] [CrossRef]
  13. Li, C.; Wang, Y.; Ma, B.; Zhang, Z. Tamper detection and self-recovery of biometric images using salient region-based authentication watermarking scheme. Comput. Stand. Interfaces 2012, 34, 367–379. [Google Scholar] [CrossRef]
  14. He, H.-J.; Zhang, J.-S.; Tai, H.-M. Self-recovery Fragile Watermarking Using Block-Neighborhood Tampering Characterization. In Computer Vision; Springer International Publishing: Berlin/Heidelberg, Germany, 2009; Volume 5806, pp. 132–145. [Google Scholar]
  15. Dadkhah, S.; Manaf, A.A.; Hori, Y.; Hassanien, A.E.; Sadeghi, S. An effective SVD-based image tampering detection and self-recovery using active watermarking. Signal Process. Image Commun. 2014, 29, 1197–1210. [Google Scholar] [CrossRef]
  16. Zhang, X.; Wang, S.; Qian, Z.; Feng, G. Reference Sharing Mechanism for Watermark Self-Embedding. IEEE Trans. Image Process. 2011, 20, 485–495. [Google Scholar] [CrossRef] [PubMed]
  17. Lee, T.-Y.; Lin, S.D. Dual watermark for image tamper detection and recovery. Pattern Recognit. 2008, 41, 3497–3506. [Google Scholar] [CrossRef]
  18. Zhang, X.; Wang, S.; Qian, Z.; Feng, G. Self-embedding watermark with flexible restoration quality. Multimed. Tools Appl. 2011, 54, 385–395. [Google Scholar] [CrossRef]
  19. Bravo-Solorio, S.; Nandi, A.K. Secure fragile watermarking method for image authentication with improved tampering localisation and self-recovery capabilities. Signal Process. 2011, 91, 728–739. [Google Scholar] [CrossRef]
  20. Wang, C.-P.; Wang, X.-Y.; Xia, Z.-Q.; Zhang, C.; Chen, X.-J. Geometrically resilient color image zero-watermarking algorithm based on quaternion Exponent moments. J. Vis. Commun. Image Represent. 2016, 41, 247–259. [Google Scholar] [CrossRef]
  21. Hsu, L.-Y.; Hu, H.-T. Blind image watermarking via exploitation of inter-block prediction and visibility threshold in DCT domain. J. Vis. Commun. Image Represent. 2015, 32, 130–143. [Google Scholar] [CrossRef]
  22. Munoz-Ramirez, D.O.; Reyes-Reyes, R.; Ponomaryov, V.; Cruz-Ramos, C. Invisible digital color watermarking technique in anaglyph 3D images. In Proceedings of the 2015 12th International Conference on Electrical Engineering, Computing Science and Automatic Control (CCE), Mexico City, Mexico, 28–30 October 2015; pp. 1–6. [Google Scholar]
  23. Fan, M.; Wang, H. An enhanced fragile watermarking scheme to digital image protection and self-recovery. Signal Process. Image Commun. 2018, 66, 19–29. [Google Scholar] [CrossRef]
  24. Tai, W.-L.; Liao, Z.-J. Image self-recovery with watermark self-embedding. Signal Process. Image Commun. 2018, 65, 11–25. [Google Scholar] [CrossRef]
  25. Chamlawi, R.; Khan, A.; Usman, I. Authentication and recovery of images using multiple watermarks. Comput. Electr. Eng. 2010, 36, 578–584. [Google Scholar] [CrossRef]
  26. Chamlawi, R.; Khan, A. Digital image authentication and recovery: Employing integer transform based information embedding and extraction. Inf. Sci. 2010, 180, 4909–4928. [Google Scholar] [CrossRef]
  27. Chamlawi, R.; Khan, A.; Idris, A. Wavelet Based Image Authentication and Recovery. J. Comput. Sci. Technol. 2007, 22, 795–804. [Google Scholar] [CrossRef]
  28. Qi, X.; Xin, X. A singular-value-based semi-fragile watermarking scheme for image content authentication with tamper localization. J. Vis. Commun. Image Represent. 2015, 30, 312–327. [Google Scholar] [CrossRef]
  29. Preda, R.O. Semi-fragile watermarking for image authentication with sensitive tamper localization in the wavelet domain. Measurement 2013, 46, 367–373. [Google Scholar] [CrossRef]
  30. Molina-Garcia, J.; Reyes-Reyes, R.; Ponomaryov, V.; Cruz-Ramos, C. Watermarking algorithm for authentication and self-recovery of tampered images using DWT. In Proceedings of the 2016 9th International Kharkiv Symposium on Physics and Engineering of Microwaves, Millimeter and Submillimeter Waves (MSMW), Kharkiv, Ukraine, 20–24 June 2016; pp. 1–4. [Google Scholar] [CrossRef]
  31. Horng, S.-J.; Rosiyadi, D.; Li, T.; Takao, T.; Guo, M.; Khan, M.K. A blind image copyright protection scheme for e-government. J. Vis. Commun. Image Represent. 2013, 24, 1099–1105. [Google Scholar] [CrossRef]
  32. Wang, X.-Y.; Liu, Y.-N.; Han, M.-M.; Yang, H.-Y. Local quaternion PHT based robust color image watermarking algorithm. J. Vis. Commun. Image Represent. 2016, 38, 678–694. [Google Scholar] [CrossRef]
  33. Dutta, T.; Gupta, H.P. A robust watermarking framework for High Efficiency Video Coding (HEVC)–Encoded video with blind extraction process. J. Vis. Commun. Image Represent. 2016, 38, 29–44. [Google Scholar] [CrossRef]
  34. Wang, R.-Z.; Lin, C.-F.; Lin, J.-C. Hiding data in images by optimal moderately-significant-bit replacement. Electron. Lett. 2000, 36, 2069. [Google Scholar] [CrossRef] [Green Version]
  35. Ponomarenko, N.; Silvestri, F.; Egiazarian, K.; Carli, M.; Astola, J.; Lukin, V. On between coefficient contrast masking of DCT basis func-tions, CD-ROM. In Proceedings of the Third International Workshop on Video Processing and Quality Metrics for Consumer Electronics, Scottsdale, AZ, USA, 25–26 January 2007. [Google Scholar]
  36. Molina-Garcia, J.; Garcia-Salgado, B.P.; Ponomaryov, V.; Reyes-Reyes, R.; Sadovnychiy, S.; Cruz-Ramos, C. An effective fragile watermarking scheme for color image tampering detection and self-recovery. Signal Process. Image Commun. 2020, 81, 115725. [Google Scholar] [CrossRef]
  37. Lin, C.-C.; He, S.-L.; Chang, C.-C. Pixel P Air-Wise Fragile Image Watermarking Based on HC-Based Absolute Moment Block Truncation Coding. Electron. 2021, 10, 690. [Google Scholar] [CrossRef]
  38. Kim, C.; Yang, C.-N. Self-Embedding Fragile Watermarking Scheme to Detect Image Tampering Using AMBTC and OPAP Approaches. Appl. Sci. 2021, 11, 1146. [Google Scholar] [CrossRef]
  39. Lee, C.-F.; Shen, J.-J.; Chen, Z.-R.; Agrawal, S. Self-Embedding Authentication Watermarking with Effective Tampered Location Detection and High-Quality Image Recovery. Sensors 2019, 19, 2267. [Google Scholar] [CrossRef] [Green Version]
  40. Kodak Photo CD, Photo Sampler. Available online: http://www.math.purdue.edu/~lucier/PHOTO_CD/ (accessed on 1 March 2021).
  41. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  42. Kwon, H.; Yoon, H.; Park, K.-W. Acoustic-decoy: Detection of adversarial examples through audio modification on speech recognition system. Neurocomputing 2020, 417, 357–370. [Google Scholar] [CrossRef]
  43. Kwon, H.; Kim, Y.; Yoon, H.; Choi, D. Random Untargeted Adversarial Example on Deep Neural Network. Symmetry 2018, 10, 738. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Protection method diagram.
Figure 1. Protection method diagram.
Applsci 11 03187 g001
Figure 2. Reconstruction and authentication method diagram.
Figure 2. Reconstruction and authentication method diagram.
Applsci 11 03187 g002
Figure 3. Images employed from Kodak database [40], (a) Kodak-1, (b) Kodak-3, (c) Kodak-11, (d) Kodak-14, (e) Kodak-4, (f) Kodak-10, (g) Kodak-17, (h) Kodak-18.
Figure 3. Images employed from Kodak database [40], (a) Kodak-1, (b) Kodak-3, (c) Kodak-11, (d) Kodak-14, (e) Kodak-4, (f) Kodak-10, (g) Kodak-17, (h) Kodak-18.
Applsci 11 03187 g003
Figure 4. Tampered images for Kodak-23, (a) 20%, (b) 30%, (c) 40%, (d) 50%, (e) 60%, (f) 70%, (g) 80%, (h) 90%.
Figure 4. Tampered images for Kodak-23, (a) 20%, (b) 30%, (c) 40%, (d) 50%, (e) 60%, (f) 70%, (g) 80%, (h) 90%.
Applsci 11 03187 g004
Figure 5. Multiple and irregular alterations, (a) Kodak-1 original, (b) Kodak-3 original, (c) Kodak-11 original, (d) Kodak-14 original, (e) Kodak-1 altered, (f) Kodak-3 altered, (g) Kodak-11 altered, (h) Kodak-14 altered.
Figure 5. Multiple and irregular alterations, (a) Kodak-1 original, (b) Kodak-3 original, (c) Kodak-11 original, (d) Kodak-14 original, (e) Kodak-1 altered, (f) Kodak-3 altered, (g) Kodak-11 altered, (h) Kodak-14 altered.
Applsci 11 03187 g005
Figure 6. Multiple and irregular alterations, (a) Kodak-1 ground truth, (b) Kodak-3 ground truth, (c) Kodak-11 ground truth, (d) Kodak-14 ground truth, (e) detection for Kodak-1, (f) detection for Kodak-3, (g) detection for Kodak-11, (h) detection for Kodak-14.
Figure 6. Multiple and irregular alterations, (a) Kodak-1 ground truth, (b) Kodak-3 ground truth, (c) Kodak-11 ground truth, (d) Kodak-14 ground truth, (e) detection for Kodak-1, (f) detection for Kodak-3, (g) detection for Kodak-11, (h) detection for Kodak-14.
Applsci 11 03187 g006
Figure 7. Quality measures values (PSNR/SSIM/PSNR-HVS-M) for Kodak-23 reconstruction given different alteration rates: (a) 20%, 37.91/0.9762/34.69; (b) 30%, 31.99/0.9552/28.82; (c) 40%, 31.26/0.9401/27.96; (d) 50%, 30.66/0.9250/27.47; (e) 60%, 29.96/0.9079/26.68; (f) 70%, 28.21/0.8739/24.69; (g) 80%, 27.25/0.8418/23.28; (h) 90%, 25.21/0.7854/20.47.
Figure 7. Quality measures values (PSNR/SSIM/PSNR-HVS-M) for Kodak-23 reconstruction given different alteration rates: (a) 20%, 37.91/0.9762/34.69; (b) 30%, 31.99/0.9552/28.82; (c) 40%, 31.26/0.9401/27.96; (d) 50%, 30.66/0.9250/27.47; (e) 60%, 29.96/0.9079/26.68; (f) 70%, 28.21/0.8739/24.69; (g) 80%, 27.25/0.8418/23.28; (h) 90%, 25.21/0.7854/20.47.
Applsci 11 03187 g007aApplsci 11 03187 g007b
Figure 8. PSNR of the reconstructed images for different alteration rates.
Figure 8. PSNR of the reconstructed images for different alteration rates.
Applsci 11 03187 g008
Figure 9. SSIM of the reconstructed images for different alteration rates.
Figure 9. SSIM of the reconstructed images for different alteration rates.
Applsci 11 03187 g009
Figure 10. PSNR-HVS-M of the reconstructed images for different alteration rates.
Figure 10. PSNR-HVS-M of the reconstructed images for different alteration rates.
Applsci 11 03187 g010
Figure 11. Reconstructed images from multiple and irregular alterations (PSNR/SSIM/PSNR-HVS-M), (a) Kodak-1 (24.53/0.7417/21.94), (b) Kodak-3 (31.82/0.8959/29.27), (c) Kodak-11 (28.12/0.8670/26.08), (d) Kodak-14 (26.38/0.8263/23.41).
Figure 11. Reconstructed images from multiple and irregular alterations (PSNR/SSIM/PSNR-HVS-M), (a) Kodak-1 (24.53/0.7417/21.94), (b) Kodak-3 (31.82/0.8959/29.27), (c) Kodak-11 (28.12/0.8670/26.08), (d) Kodak-14 (26.38/0.8263/23.41).
Applsci 11 03187 g011
Figure 12. Multiple and irregular alterations, (a) Kodak-4 original, (b) Kodak-10 original, (c) Kodak-17 original, (d) Kodak-18 original, (e) Kodak-4 modified, (f) Kodak-10 modified, (g) Kodak-17 modified, (h) Kodak-18 modified, (i) Kodak-4 ground truth, (j) Kodak-10 ground truth, (k) Kodak-17 ground truth, (l) Kodak-18 ground truth.
Figure 12. Multiple and irregular alterations, (a) Kodak-4 original, (b) Kodak-10 original, (c) Kodak-17 original, (d) Kodak-18 original, (e) Kodak-4 modified, (f) Kodak-10 modified, (g) Kodak-17 modified, (h) Kodak-18 modified, (i) Kodak-4 ground truth, (j) Kodak-10 ground truth, (k) Kodak-17 ground truth, (l) Kodak-18 ground truth.
Applsci 11 03187 g012
Figure 13. Altered images for Kodak-15, (a) 20%, (b) 30%, (c) 40%, (d) 50%, (e) 60%, (f) 70%, (g) 80%, (h) 90%.
Figure 13. Altered images for Kodak-15, (a) 20%, (b) 30%, (c) 40%, (d) 50%, (e) 60%, (f) 70%, (g) 80%, (h) 90%.
Applsci 11 03187 g013
Table 1. Least-significant bit (LSB) embedding analysis in terms of the objective quality.
Table 1. Least-significant bit (LSB) embedding analysis in terms of the objective quality.
No Bit AdjustmentWith Bit Adjustment
Embedding1-LSB2-LSB3-LSB2-LSB3-LSB
PSNR51.1743.8937.5144.3337.69
SSIM0.99660.98240.93140.98430.9340
PSNR-HVS59.2149.8941.8049.6941.92
Table 2. Precision and Recall obtained from the authentication process.
Table 2. Precision and Recall obtained from the authentication process.
Embedding10%20%30%40%50%60%70%80%90%
Recall1-LSB0.90330.92110.93690.93450.93560.93860.94040.93890.9384
2-LSB111111111
3-LSB111111110.9999
Precision1-LSB0.94660.97330.98310.9880.99430.99130.99220.99320.9944
2-LSB0.91570.95850.97180.97990.98850.98490.98770.98920.9910
3-LSB0.88930.94480.96370.97440.98430.98190.98550.98740.9900
Table 3. PSNR, SSIM and PSNR-HVS-M from the reconstruction process.
Table 3. PSNR, SSIM and PSNR-HVS-M from the reconstruction process.
Embedding10%20%30%40%50%60%70%80%90%
PSNR1-LSB27.8125.0223.5021.7920.3019.0117.5715.7913.286
2-LSB35.8032.7130.6329.2528.1327.0626.0024.8223.07
3-LSB35.5232.4830.4329.0828.0327.0626.2225.2723.68
SSIM1-LSB0.9530.9030.8510.7870.7130.6290.5330.4100.250
2-LSB0.9670.9350.9010.8670.8320.7930.7500.6960.615
3-LSB0.9620.9250.8880.8510.8130.7720.7300.6800.604
PSNR-HVS-M1-LSB26.1922.8320.9118.9417.0415.2613.5411.498.78
2-LSB33.2930.1928.1426.6925.5724.3723.0221.3218.87
3-LSB33.2530.1828.1526.7325.6924.7123.7622.3519.90
Table 4. Objective quality metrics PSNR, SSIM and PSNR-HVS-M for watermarked images.
Table 4. Objective quality metrics PSNR, SSIM and PSNR-HVS-M for watermarked images.
With Bit AdjustmentWithout Bit Adjustment
PSNRSSIMPSNR-HVSPSNRSSIMPSNR-HVS
Kodak-144.360.992352.4943.930.991453.30
Kodak-243.870.978749.0243.440.975650.17
Kodak-344.260.976248.0243.810.973247.96
Kodak-444.300.981949.7343.900.979850.05
Kodak-544.200.992250.8943.830.991452.12
Kodak-644.410.987449.1043.920.985948.82
Kodak-744.310.981748.9943.920.979749.10
Kodak-844.410.993452.1643.940.992552.50
Kodak-944.470.979049.1044.020.976448.94
Kodak-1044.450.980649.9043.970.978349.57
Kodak-1144.340.985450.9943.840.983450.80
Kodak-1244.340.978849.4743.820.976148.84
Kodak-1344.300.994852.4943.880.994153.46
Kodak-1444.270.989650.7643.890.988451.67
Kodak-1544.340.980048.1043.920.977747.74
Kodak-1644.320.982549.8243.890.980549.87
Kodak-1744.170.981948.4043.840.980048.82
Kodak-1844.140.987950.1143.770.986551.20
Kodak-1944.440.984251.1243.960.982351.03
Kodak-2044.890.984544.9944.270.983244.32
Kodak-2144.320.982849.1643.910.980849.27
Kodak-2244.410.985350.3543.950.983450.38
Kodak-2344.350.975848.4343.940.973148.39
Kodak-2444.270.987649.1643.840.986249.20
Table 5. Precision values for the detection of different alteration rates between 10% and 90%.
Table 5. Precision values for the detection of different alteration rates between 10% and 90%.
10%20%30%40%50%60%70%80%90%
Kodak-10.91700.96290.97480.98400.98980.98500.98900.99030.9930
Kodak-20.91670.96310.97480.98440.98980.98510.98940.99040.9928
Kodak-30.91770.96250.97470.98400.98980.98500.98920.99040.9929
Kodak-40.91070.94540.96300.96730.98460.98450.98350.98570.9854
Kodak-50.91700.96250.97460.98410.98980.98520.98920.99030.9930
Kodak-60.91700.96290.97480.98410.98980.98510.98910.99060.9929
Kodak-70.91700.96270.97470.98400.98980.98500.98900.99060.9929
Kodak-80.91800.96290.97470.98400.98980.98520.98910.99040.9929
Kodak-90.91070.94540.96290.96760.98460.98420.98350.98560.9851
Kodak-100.91070.94500.96300.96700.98460.98440.98360.98590.9853
Kodak-110.91770.96270.97460.98410.98980.98500.98900.99030.9929
Kodak-120.91770.96270.97480.98410.98980.98510.98900.99040.9929
Kodak-130.91700.96290.97480.98410.98980.98520.98910.99030.9929
Kodak-140.91670.96400.97460.98430.98980.98530.98910.99030.9928
Kodak-150.91670.96270.97470.98420.98980.98510.98910.99050.9930
Kodak-160.91800.96290.97470.98430.98980.98510.98900.99040.9928
Kodak-170.91070.94520.96290.96740.98460.98400.98380.98570.9853
Kodak-180.91170.94550.96260.96710.98460.98410.98380.98570.9852
Kodak-190.91210.94460.96260.96780.98460.98410.98360.98570.9851
Kodak-200.91670.96340.97470.98420.98980.98500.98900.99030.9929
Kodak-210.91740.96330.97480.98420.98980.98500.98900.99030.9929
Kodak-220.91740.96250.97480.98400.98980.98500.98940.99050.9929
Kodak-230.91740.96270.97470.98410.98980.98510.98920.99040.9928
Kodak-240.91770.96290.97470.98400.98980.98510.98900.99040.9931
Table 6. Quality comparison using PSNR, SSIM and PSNR-HVS-M between watermarked images and original ones.
Table 6. Quality comparison using PSNR, SSIM and PSNR-HVS-M between watermarked images and original ones.
PSNR (dB)SSIMPSNR-HVS (dB)
SR-HTR44.330.984349.69
Molina [36]44.320.984549.36
Fan [23]44.080.983250.09
Singh [2]37.850.936442.04
Tai [24]44.080.983250.07
Tong [10]37.850.936342.05
Table 7. Comparison of Precision for different alteration rates between 10% and 90%.
Table 7. Comparison of Precision for different alteration rates between 10% and 90%.
10%20%30%40%50%60%70%80%90%
SR-HTR0.91570.95850.97180.97990.98850.98490.98770.98920.9910
Molina [36]0.91520.95800.97160.97970.98840.98480.98760.98910.9909
Fan [23]0.80070.92100.91440.94831.00000.96010.97480.96590.9762
Singh [2]0.98551.00001.00000.99631.00000.99751.00001.00000.9983
Tai [24]0.96700.98550.99030.99391.00000.99430.99580.99630.9972
Tong [10]0.98551.00001.00000.99631.00000.99751.00001.00000.9983
Table 8. Comparison of Precision for multiple and irregular modification detection.
Table 8. Comparison of Precision for multiple and irregular modification detection.
SR-HTRMolina
[36]
Fan
[23]
Singh
[2]
Tai
[24]
Tong
[10]
Kodak-40.97970.97820.945110.99391
Kodak-100.91480.90600.757310.97761
Kodak-170.97840.97700.762810.97791
Kodak-180.92430.91770.895110.98841
Table 9. Visual quality comparison for the reconstruction of Kodak-15 using an alteration rate between 20% and 50%.
Table 9. Visual quality comparison for the reconstruction of Kodak-15 using an alteration rate between 20% and 50%.
20%30%40%50%
SR-HTR Applsci 11 03187 i001 Applsci 11 03187 i002 Applsci 11 03187 i003 Applsci 11 03187 i004
Molina [36] Applsci 11 03187 i005 Applsci 11 03187 i006 Applsci 11 03187 i007 Applsci 11 03187 i008
Fan [23] Applsci 11 03187 i009 Applsci 11 03187 i010 Applsci 11 03187 i011 Applsci 11 03187 i012
Singh [2] Applsci 11 03187 i013 Applsci 11 03187 i014 Applsci 11 03187 i015 Applsci 11 03187 i016
Tai [24] Applsci 11 03187 i017 Applsci 11 03187 i018 Applsci 11 03187 i019 Applsci 11 03187 i020
Tong [10] Applsci 11 03187 i021 Applsci 11 03187 i022 Applsci 11 03187 i023 Applsci 11 03187 i024
Table 10. Visual quality comparison for the reconstruction of Kodak-15 using an alteration rate between 60% and 90%.
Table 10. Visual quality comparison for the reconstruction of Kodak-15 using an alteration rate between 60% and 90%.
60%70%80%90%
SR-HTR Applsci 11 03187 i025 Applsci 11 03187 i026 Applsci 11 03187 i027 Applsci 11 03187 i028
Molina [36] Applsci 11 03187 i029 Applsci 11 03187 i030 Applsci 11 03187 i031 Applsci 11 03187 i032
Fan [23] Applsci 11 03187 i033 Applsci 11 03187 i034 Applsci 11 03187 i035 Applsci 11 03187 i036
Singh [2] Applsci 11 03187 i037 Applsci 11 03187 i038 Applsci 11 03187 i039 Applsci 11 03187 i040
Tai [24] Applsci 11 03187 i041 Applsci 11 03187 i042 Applsci 11 03187 i043 Applsci 11 03187 i044
Tong [10] Applsci 11 03187 i045 Applsci 11 03187 i046 Applsci 11 03187 i047 Applsci 11 03187 i048
Table 11. Comparison of objective quality metrics for the reconstruction of Kodak-15 using alteration rates between 20% and 90%.
Table 11. Comparison of objective quality metrics for the reconstruction of Kodak-15 using alteration rates between 20% and 90%.
20%30%40%50%60%70%80%90%
PSNRSR-HTR33.29531.52429.89728.83628.19427.39826.45924.106
Molina [36]34.79531.37428.28225.49922.92420.70518.54816.474
Fan [23]29.16622.84920.1617.14213.60411.8559.5658.295
Singh [2]20.60317.25114.90313.05911.63210.5149.4528.356
Tai [24]16.56714.55613.31212.01710.9149.9719.0187.998
Tong [10]27.88222.72819.21916.43914.19612.39410.7349.101
SSIMSR-HTR0.97120.93500.89200.85680.82330.78750.75970.7091
Molina [36]0.95910.91030.83970.75370.63970.51770.40120.3187
Fan [23]0.97450.92840.88580.82250.67400.55150.32860.1694
Singh [2]0.82820.72870.62710.52540.41900.31480.21060.1044
Tai [24]0.83330.73410.63330.52770.41560.30920.20360.0997
Tong [10]0.91680.81540.69640.57720.45360.33660.22230.1104
PSNR-HVS-MSR-HTR30.78928.89427.14526.10825.47424.49123.13920.164
Molina [36]32.71728.03224.43521.45718.72316.42314.23612.131
Fan [23]28.12621.79319.23116.05112.38110.6138.4817.185
Singh [2]18.32314.82112.48110.5128.9637.8256.7425.535
Tai [24]18.87915.60513.30511.329.9699.0338.0426.826
Tong [10]25.89520.79417.08214.17711.7989.8958.1576.375
Table 12. Comparison of average PSNR for the reconstruction of the test images with alteration rates between 10% and 90%.
Table 12. Comparison of average PSNR for the reconstruction of the test images with alteration rates between 10% and 90%.
10%20%30%40%50%60%70%80%90%
SR-HTR35.8032.7230.6329.2528.1427.0626.0124.8223.08
Molina [36]37.2833.9731.2028.7126.3823.9921.7719.7918.01
Fan [23]31.6229.3223.7520.9618.4114.9212.9810.999.63
Singh [2]26.7421.7618.6116.3014.5413.0411.7910.709.71
Tai [24]26.3421.0517.9415.7114.0212.5711.3610.339.40
Tong [10]35.2228.6823.9820.5917.9215.6113.6711.9810.46
Table 13. Comparison of average SSIM for the reconstruction of the test images with alteration rates between 10% and 90%.
Table 13. Comparison of average SSIM for the reconstruction of the test images with alteration rates between 10% and 90%.
10%20%30%40%50%60%70%80%90%
SR-HTR0.9670.9350.9010.8680.8330.7930.7500.6960.616
Molina [36]0.9720.9430.9060.8520.7800.6790.5610.4400.336
Fan [23]0.9770.9590.9190.8750.8170.6760.5400.3470.167
Singh [2]0.9340.8410.7430.6390.5370.4310.3260.2200.111
Tai [24]0.9420.8540.7530.6480.5420.4290.3190.2100.103
Tong [10]0.9740.9210.8340.7220.6030.4780.3590.2400.122
Table 14. Comparison of average PSNR-HSV-M for the reconstruction of the test images with alteration rates between 10% and 90%.
Table 14. Comparison of average PSNR-HSV-M for the reconstruction of the test images with alteration rates between 10% and 90%.
10%20%30%40%50%60%70%80%90%
SR-HTR33.2930.2028.1526.7025.5724.3823.0321.3218.88
Molina [36]36.1232.4729.1025.8923.0720.2617.7515.5413.58
Fan [23]31.3729.3323.2120.3317.7314.0912.1310.158.78
Singh [2]23.7918.8515.7113.3611.559.988.677.546.52
Tai [24]25.2919.9816.7714.4112.6711.2710.169.258.43
Tong [10]34.7026.8721.8018.1315.2512.7910.738.947.33
Table 15. Visual comparison of the reconstructed images for multiple and irregular alterations.
Table 15. Visual comparison of the reconstructed images for multiple and irregular alterations.
SR-HTRMolina [36]Fan [23]Singh [2]Tai [24]Tong [10]
Kodak-4 Applsci 11 03187 i049 Applsci 11 03187 i050 Applsci 11 03187 i051 Applsci 11 03187 i052 Applsci 11 03187 i053 Applsci 11 03187 i054
Kodak-10 Applsci 11 03187 i055 Applsci 11 03187 i056 Applsci 11 03187 i057 Applsci 11 03187 i058 Applsci 11 03187 i059 Applsci 11 03187 i060
Kodak-17 Applsci 11 03187 i061 Applsci 11 03187 i062 Applsci 11 03187 i063 Applsci 11 03187 i064 Applsci 11 03187 i065 Applsci 11 03187 i066
Kodak-18 Applsci 11 03187 i067 Applsci 11 03187 i068 Applsci 11 03187 i069 Applsci 11 03187 i070 Applsci 11 03187 i071 Applsci 11 03187 i072
Table 16. PSNR, SSIM and PSNR-HVS-M for reconstructed images presented in Table 15.
Table 16. PSNR, SSIM and PSNR-HVS-M for reconstructed images presented in Table 15.
SR-HTRMolina
[36]
Fan
[23]
Singh
[2]
Tai
[24]
Tong
[10]
PSNRKodak-425.9019.4210.0110.259.9711.22
Kodak-1030.2132.0524.5020.2718.9726.17
Kodak-1726.7620.9511.3311.4611.2013.07
Kodak-1827.9128.9719.1717.4917.2022.85
SSIMKodak-40.68830.34330.19610.14290.13410.1548
Kodak-100.91300.91970.91500.76800.78180.8658
Kodak-170.76610.49940.38580.26500.26020.2933
Kodak-180.88480.88680.85530.71870.72620.8174
PSNR-HVS-MKodak-421.6914.7811.218.2111.279.23
Kodak-1026.5230.2624.0616.6715.7623.65
Kodak-1723.9116.7710.428.3710.2710.13
Kodak-1825.6627.5117.6814.4414.2720.61
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Reyes-Reyes, R.; Cruz-Ramos, C.; Ponomaryov, V.; Garcia-Salgado, B.P.; Molina-Garcia, J. Color Image Self-Recovery and Tampering Detection Scheme Based on Fragile Watermarking with High Recovery Capability. Appl. Sci. 2021, 11, 3187. https://doi.org/10.3390/app11073187

AMA Style

Reyes-Reyes R, Cruz-Ramos C, Ponomaryov V, Garcia-Salgado BP, Molina-Garcia J. Color Image Self-Recovery and Tampering Detection Scheme Based on Fragile Watermarking with High Recovery Capability. Applied Sciences. 2021; 11(7):3187. https://doi.org/10.3390/app11073187

Chicago/Turabian Style

Reyes-Reyes, Rogelio, Clara Cruz-Ramos, Volodymyr Ponomaryov, Beatriz P. Garcia-Salgado, and Javier Molina-Garcia. 2021. "Color Image Self-Recovery and Tampering Detection Scheme Based on Fragile Watermarking with High Recovery Capability" Applied Sciences 11, no. 7: 3187. https://doi.org/10.3390/app11073187

APA Style

Reyes-Reyes, R., Cruz-Ramos, C., Ponomaryov, V., Garcia-Salgado, B. P., & Molina-Garcia, J. (2021). Color Image Self-Recovery and Tampering Detection Scheme Based on Fragile Watermarking with High Recovery Capability. Applied Sciences, 11(7), 3187. https://doi.org/10.3390/app11073187

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop