Next Article in Journal
Corchorus Olitorius-Mediated Green Synthesis and Characterization of Nickel and Manganese Ferrite Nanoparticles
Previous Article in Journal
Gap Solitons in Fiber Bragg Gratings Having Polynomial Law of Nonlinear Refractive Index and Cubic–Quartic Dispersive Reflectivity by Lie Symmetry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Shrinkage and Redundant Feature Elimination Network-Based Robust Image Zero-Watermarking

1
College of Science and Technology, Ningbo University, Ningbo 315000, China
2
Faculty of Information Science and Engineering, Ningbo University, Ningbo 315211, China
*
Author to whom correspondence should be addressed.
Symmetry 2023, 15(5), 964; https://doi.org/10.3390/sym15050964
Submission received: 29 March 2023 / Revised: 19 April 2023 / Accepted: 20 April 2023 / Published: 23 April 2023

Abstract

:
To address the contradiction between watermarking robustness and imperceptibility, a zero-watermarking method based on shrinkage and a redundant feature elimination network (SRFENet) is proposed in this paper. First, in order to have the capability of resisting different image attacks, a dense connection was used to extract shallow and deep features from different convolutional layers. Secondly, to reduce unimportant information for robustness and uniqueness, in SRFENet, a shrinkage module was utilized by automatically learning the threshold of each feature channel. Then, to enhance watermarking uniqueness, a redundant feature elimination module was designed to reduce redundant information for the remaining valid features by learning the weights of inter-feature and intra-feature. In order to increase watermarking robustness further, noised images were generated for training. Finally, an extracted feature map from SRFENet was used to construct a zero-watermark. Furthermore, a zero-watermark from the noised image was generated for copyright verification, which is symmetrical to the process of zero-watermark construction from the original image. The experimental results showed that the proposed zero-watermarking method was robust to different single-image attacks (average BER is 0.0218) and hybrid image attacks (average NC is 0.9551), proving the significant generalization ability to resist different attacks. Compared with existing zero-watermarking methods, the proposed method is more robust since it extracts the main image features via learning a large number of different images for zero-watermark construction.

1. Introduction

With the development of computer and Internet technology, digital media such as images, audio and video can be easily gained and modified [1,2,3]. Some illegal users may tamper, copy and spread images without the permission of owners, which seriously infringes their copyrights [4,5]. Thus, how to effectively protect images is an important challenge in the field of information security, and watermarking technology is one of the main solutions to the problem of image security [6,7,8].
Watermarking is generally divided into fragile and robust watermarking methods, which are used for image authentication and image copyright protection, respectively [9,10]. Huang et al. [11] designed a self-embedding watermark using the least significant substitution (LSB), which could detect tampered regions effectively. Different from the fragile watermarking method, the robust watermarking method can resist different image attacks to realize copyright protection. Thanki et al. [12] decomposed the image by using discrete curvelet transform (DCuT) and redundant discrete wavelet transform (RDWT), and then a watermark was embedded by modifying coefficients of the wavelet coefficients to resist common image attacks. In addition, many other transforms were utilized for robust watermarking, such as discrete cosine transform (DCT), fractional Fourier transform (FFT) and singular value decomposition (SVD), and so on [13,14]. However, the transform-based robust watermarking method was generally weak to geometric attacks. A moment-based watermarking was presented to enhance robustness against geometric attacks [15,16]. Xin et al. [15] used in-variance properties of Zernike moments (ZMs) and pseudo-Zernike moments (PZMs) to achieve robustness on geometric image distortions. Hosny et al. [16] calculated fractional multi-channel orthogonal exponential moments (MFrEMs) for embedding the watermark, which had high robustness on geometric attacks, such as translation, scaling and hybrid attacks. In order to improve the ability to resist image attacks, the above watermarking methods increased the watermarking strength but reduced the image quality. Moreover, in some special fields, such as the military, medicine and law, modifying the image is not allowed. Thus, zero-watermarking technology was designed, which could effectively protect the copyright of the image without damaging any pixel of the image and solve the contradiction between watermarking invisibility and robustness [17,18,19,20].
The zero-watermarking technology extracts correspond with robust features of the image to constructing the relevant information that uniquely identifies the image for protecting copyright. Chang et al. [21] generated a binary zero-watermark depending on the texture properties of each image block. Chang et al. [22] computed an approximation image by using low-pass filtering and downscaling, and then a binary watermark was constructed by performing Sobel edge detection on the approximation image. However, spatial domain-based features are sensitive to most image attacks. Kang et al. [23] decomposed the image using Discrete Wavelet Transform (DWT) and SVD and combined the Frobenius norm of SVD with the majority voting model to extract a robust zero-watermark. However, the transform-based features had a lack of rotation and were fragile to geometric attacks. Kang et al. [24] calculated three polar harmonic transforms (PHT) moments of the image and used the size relationship of adjacent moments to construct a zero-watermark. Yang et al. [25] computed low-order quaternion generic polar complex exponential transform (QGPCET) moments and mixed them with robust features to resist geometric attacks. Xia et al. [26] computed local feature regions (LFR) using the feature scale-adaptive process and then used the quaternion polar harmonic Fourier moments (QPHFMs) of LFRs for robust zero-watermarking. However, the above zero-watermarking methods are not robust to some image attacks for varieties of images since their features are manually extracted by prior knowledge for constructing the zero-watermark. Thus, they do not have the capability of generalization for resisting different image attacks.
In recent years, since deep learning can extract main image features by learning a variety of images for different purposes, different networks have been widely designed in the field of computer vision, such as object detection, saliency detection, image classification, and so on [27,28]. Girshick et al. [29] used convolutional neural networks (CNN) to extract region features for object detection. He et al. [30] presented ResNets for image recognition, in which a skip connection was added directly from the input of each module to overcome the gradient vanishing problem in backpropagation. Sun et al. [31] proposed an automatic CNN architecture based on genetic algorithms for image classification. Ding et al. [32] introduced a feedback recursive convolutional model (FBNet) for image saliency detection. In addition to computer vision, researchers also combined deep learning with watermarking to improve the corresponding performance. Kandi et al. [33] used CNN for watermarking to improve robustness. However, the above method is not blind; that is, the original image is needed in the processes of watermark extraction. Ahmadi et al. [34] proposed a non-blind watermarking method based on CNN, which used the residual structure to improve the quality of the watermarked image. Luo et al. [35] trained the noise adversarial attack and employed the channel coding for robustness, but watermarking redundancy was produced to degrade the image quality due to the channel coding. Rai et al. [36] improved the Chimpanzee Optimization Algorithm (ChOA) and combined a deep fusion convolutional neural network to improve the watermarking performance. However, the above watermarking models still could not solve the contradiction between watermarking invisibility and robustness; that is, when robustness was improved, image quality was greatly reduced. Designing deep learning network models for zero-watermarking is an effective way to obtain high robustness on different attacks without degrading the image quality. To our knowledge, there are only a few reports on zero-watermarking based on network models. Fierro-Radilla et al. [37] designed a zero-watermarking model based on CNN. However, its performance was not high since the designed CNN could not extract robust features.
This paper proposes a shrinkage and redundant feature elimination network (SRFENet)-based zero-watermarking method. First, a dense connection was utilized to extract the robust features for watermarking robustness in the feature extraction module. Then, in order to enhance watermarking robustness and uniqueness, a shrinkage module was designed to reduce fragile features and noises by learning the threshold of each channel feature automatically. Moreover, to increase the watermarking uniqueness, the redundant features were removed further by the learning weights of inter-feature and intra-feature. Third, to obtain the capability of generalization to resist different image attacks, noised images were generated for training. Finally, a zero-watermark was built on the extracted image features from SRFENet. In the verification process, a zero-watermark from the noised image was generated, which was in symmetry to the zero-watermark construction from the original image. The experimental results showed that the proposed zero-watermarking method based on SRFENet was robust to different image attacks and superior to the existing methods. The main contributions to this paper are listed as follows:
  • In order to obtain the robustness of zero-watermarking, the feature extraction module of SRFENet effectively combines with the dense connection and the max pooling to extract robust features.
  • In order to enhance the uniqueness of zero-watermarking, the SRFENet adopts the shrinkage module and the redundant feature elimination module so as to increase the discrimination of different image features. At the same time, they are also beneficial to watermarking robustness.
  • Different from traditional zero-watermarking methods, the proposed method employed noise training on different images for feature extraction instead of prior knowledge. Therefore, the proposed method can resist varieties of image attacks and perform better on different image datasets compared with existing zero-watermarking methods.
The rest of this paper is organized as follows. Section 2 presents the proposed zero-watermarking in detail. Section 3 provides the experimental results and corresponding analysis. Section 4 gives a conclusion.

2. Proposed Zero-Watermarking

In order to obtain watermarking uniqueness and robustness, a zero watermarking model based on shrinkage and a redundant feature elimination network (SRFENet) was proposed. SRFENet was designed to obtain effective image features for constructing a zero watermark, as illustrated in Figure 1. Table 1 shows the variables and related descriptions. For SRFENet training, the original image I and the noised image N were input into SRFENet for obtaining the image features FI and FN, respectively, and FI and FN are hoped to similarly resist image noises. Based on the trained SRFENet, a zero watermark was computed from the original image by using binary image processing for copyright protection. Similarly, the zero watermarks could still be extracted from the noised image for image verification. In the following, the structure of SRFENet, robustness training, zero watermark generation and zero watermark extraction for image verification are depicted in detail.

2.1. Structure of SRFENet

The purpose of SRFENet was used to extract effective image features for watermark uniqueness and robustness. SRFENet included a feature extraction module, a shrinkage module and a feature redundancy elimination module, as illustrated in Figure 2. The feature extraction module was used to obtain robust image features by using the dense connection and convolution pooling. The shrinkage module was utilized to reduce noises and redundant information by learning automatically the threshold of each feature channel for improving robustness and uniqueness. In order to further reduce redundant features for the discrimination of different image features, the feature redundancy elimination module was employed so that the watermarking uniqueness could be enhanced and different images generated different zero-watermarks. The image size was M × N × 3, and three modules of SRFENet were introduced concretely as follows.

2.1.1. Feature Extraction Module

Huang [38] presented a dense convolution network to obtain the main energy of the image by fusing the image features of each network layer. In addition, image features obtained via the maximum pooling layer were robust to geometric attacks such as rotation [39]. Thus, in order to extract robust image features, the dense connection and the max pooling were combined in the feature extraction module.
Specifically, when the original image I was input into the first convolutional layer to extract the initial feature map F0, this was used as the input of the dense connection layer. In the dense connection, the output of each network layer was used as the input of the next layer so that the information flowing between layers was maximized for feature reuse to strengthen the feature propagation of each network layer. For each layer of the dense connection, the output of the layer l could be computed as:
F l = G l ( c o n ( F 0 , F 1 , , F l 1 ) )
where Fl denotes the feature map from the layer l of the dense connection, con(·) connects feature maps, and Gl(·) is the non-linear transformation consisting of the convolution operation of 3 × 3 kernel (Conv), the batch normalization (BN) and the rectified linear unit (ReLU) activation function. For instance, to obtain F2, F0 and F1 are connected and can be processed by using Equation (1). It can be supposed that each network layer generates K features operated by Gl(·), and then the input feature number for layer l is K × (L − 1) + K0, where K0 is the feature number of F0. In order to avoid complex training due to the large number of input features, three layers of the dense connection were employed to generate F3 with the size of M × N × 64, and each layer had 64 convolution kernels.
Then, F3 was convoluted and pooled, in turn, to obtain the feature map F8 with the size H × W × 64, wherein H = M/4 and W = N/4. F8 was computed as:
F 8 = M a x P o o l ( C o n v ( C o n v ( M a x P o o l ( C o n v ( F 3 ) ) ) ) )
where MaxPool(·) is a maximum pooling operation, Conv(·) represents a convolution group including a 3 × 3 convolution operation, and a BN and a ReLU activation function. After the feature extraction module, a robust feature map F8 was obtained, and its size was one-quarter of the original image. Through different network layers, the main energies of the images were extracted, which included many robust features. However, low features were also obtained, some of which were often redundant and fragile. Thus, in order to reduce those redundant and fragile features, a shrinkage module was designed.

2.1.2. Shrinkage Module

As mentioned before, although F8 is robust, it still contains redundant information and some fragile features affecting watermarking uniqueness and robustness. Since soft thresholding is the core step of signal denoising [40], a shrinkage module was designed so that soft thresholding could be inserted as a nonlinear transformation into the proposed network to reduce redundant features, as illustrated in Figure 2. The shrinkage module was designed to decrease redundant information and enhance watermarking uniqueness. Meanwhile, after attacking images, features can be shrunk by removing noises for watermark extraction so that watermarking robustness can be increased as well. Specifically, each feature value xi,j,c of F8 was computed as yi,j,c of F10 by using Equation (3) to remove near-zero features, where (i,j,c) corresponded to the index value for the channel c.
y i , j , c = { x i , j , c t c , x i , j , c > t c 0 , t c x i , j , c t c x i , j , c + t c , x i , j , c < t c
where tc is the value of the channel c of T, and T is computed by learning the channel weight of F8:
T = S ( M L P ( F 9 ) ) × F 9
where MLP(·) is the multilayer perceptron (MLP), S(·) is the Sigmoid function, and F9 can be defined as:
F 9 = A v g P o o l ( | F 8 | )
where AvgPool(·) is the global average pooling in the spatial domain. Different from setting negative features to zero in the ReLU activation function, soft thresholding only sets near-zero features to zero to eliminate noise, fragile features and redundant information so that both useful negative and positive features can be preserved for watermarking robustness and uniqueness [40].

2.1.3. Feature Redundancy Elimination Module

Although the shrinkage module can effectively reduce unimportant features, there is still some redundant information affecting the discrimination of different images. In order to further improve watermarking uniqueness, a feature redundancy elimination module was designed as well. Via learning the inter-feature and intra-feature weights, redundant features were decreased to enhance effective features.
Specifically, the process of inter-feature weight learning includes inter-feature compression, MLP and weighted inter-feature fusion, as illustrated in Figure 2. The purpose of inter-feature compression is to extract the global information of each feature for MLP learning, and F10 can be compressed as:
{ F max c = M a x P o o l ( F 10 ) F a v g c = A v g P o o l ( F 10 )
where MaxPool(·) represents the global maximum pooling, and the sizes of Fcmax and Fcavg are 1 × 1 × 64.
Then, MLP was used to measure the correlations of features by learning the inter-feature weights of F10:
M c ( F d ) = S ( M L P ( F a v g c ) + M L P ( F max c ) )
where Mc(Fd) is the feature weight for inter-feature fusion to compute Fc:
F c = F 10 × M c ( F 10 )
After eliminating the redundant inter-feature, intra-feature weight learning was used to remove the redundant intra-feature, and this process included intra-feature compression and weighted intra-feature fusion. At first, Fc was compressed as:
{ F max c s = M a x P o o l ( F c ) F a v g c s = A v g P o o l ( F c )
The sizes of Fcsmax and Fcsavg were H × W × 1, and then, in the weighted intra-feature fusion, Fcs was computed as:
F c s = F c × M s ( F c )
where the intra-feature weight Ms(Fc) could be calculated as:
M s ( F c ) = S ( C o n v ( c o n ( F max c s , F avg c s ) ) )
Inter-feature weight learning reduces the redundant information between features, and meanwhile, intra-feature weight learning eliminates the redundant content within features. Different from F10, Fcs highlights the distinguishing features that are closely related to the image content. Moreover, Fcs still preserves robust features from F10 and has both high robustness and uniqueness.
Finally, Fcs was operated by global average pooling, and Equation (12) was used to obtain FI with a size of H × W × 1, and FI was used to construct the zero-watermark.
F I = A v g P o o l ( F c s )

2.2. Noise Training

In order to increase zero-watermarking robustness, different noises were added to the image for training, such as JPEG compression, Gaussian filtering, median filtering and Gaussian noise. Zero-watermark extracted from the noised image was supposed to be the same as the zero watermarks constructed from the original image. Since the zero-watermark is constructed based on the feature map extracted from SRFENet, features of the original image and the noised image can be trained for similarity. Specifically, for each iteration, one noise was added to the original image to generate the noised image N, and the corresponding feature map FN was extracted from SRFENet, as illustrated in Figure 1. The loss function was built on:
L M S E = M S E ( F I , F N ) = i = 1 p j = 1 q ( X i , j Y i , j ) 2 H × W
where MSE(·) computes the mean square error and Xi,j, and Yi,j are feature values of FI and FN, respectively.
In addition, in order to resist varieties of image noise, an attack was randomly selected with equal probability from JPEG compression, Gaussian filtering, median filtering and Gaussian noise and operated on a group of images during each iteration of network training. Through continuous training, LMSE continuously decreased so that FI and FN became similar. When LMSE was stable, the parameters of SRFENet were obtained and trained, while SRFENet was used to extract the robust features for constructing a zero-watermark.

2.3. SRFENet-Based Zero-Watermark Construction for Image Copyright Protection

In terms of zero-watermark construction, the feature map FI was divided into blocks with a size of 4 × 4, and each value of the feature block was compared with the mean value of the block to construct a zero-watermark. If the zero-watermark generated from the original image, I is ZW, the original watermark can be W0 with a size of H × W. The main steps for constructing a zero-watermark are listed as follows:
Step 1. The input I into SRFENet and FI is computed.
Step 2. Divide FI into non-overlapping blocks Pk with a size of 4 × 4, wherein k is the index of each block, and each value of Pk can be binarized as:
z i , j k = { 0 , x i , j k < A k 1 , x i , j k A k , i = 1 , 2 , , h ; j = 1 , 2 , w
where Ak is the mean of Pk and xki,j is the value of Pk located at (i,j). Repeat this step until all blocks are computed, and zki,j is used to form a binary map BI.
Step 3. Perform an XOR operation on BI and W0 to generate ZW:
Z W = B I W 0
Step 4. From a time stamp authority, apply a time stamp to be combined with ZW to register in the Intellectual Property Right Database (IPRD). The construction and registration process of the zero-watermark is then finished.

2.4. SRFENet-Based Zero-Watermark Extraction for Image Verification

When receiving an image, the image may be attacked. In order to verify the copyright of the image, the corresponding zero-watermark can be extracted from the noised image. The process of zero-watermark extraction from the noised image N is symmetrical to that of constructing a zero-watermark from the original image I. Detailed steps for this are depicted as follows:
Step 1. Input N into SRFENet and FN is computed.
Step 2. Divide FN into non-overlapping blocks Qk with a size of 4 × 4, wherein k is the index of each block, and each value of Qk is binarized as:
z i , j k = { 0 , x i , j k < A k 1 , x i , j k A k , i = 1 , 2 , , h ; j = 1 , 2 , w
where Ak is the mean of Qk and xk′i,j is the value of Qk located at (i,j). Repeat this step until all blocks are computed, and zk′i,j is used to form a binary map BN.
Step 3. Perform the XOR operation on BN and ZW to compute W1:
W 1 = B N Z W
Step 4. If the extracted watermark W1 is similar to W0, and the extracted timestamp is authenticated, the copyright of the image is verified.

3. Experimental Results and Discussions

To evaluate the effectiveness of the proposed watermarking model, more than 100,000 color images from the COCO dataset [41] were used for training, and stochastic gradient descent (SGD) was used as an optimizer with a learning rate of 0.0001. Furthermore, eight images with a size of 512 × 512 were used for testing in Figure 3.
In order to evaluate the uniqueness of zero watermarks generated from different images, a normalized correlation coefficient (NC) was used for computing the similarity of different zero watermarks, and a bit error rate (BER) was used to detect the similarity between the original watermark and the extracted watermark for evaluating watermarking robustness.

3.1. Uniqueness Evaluation

In order to verify the uniqueness of the zero-watermark from different images, NCs between zero watermarks of different images were computed. Table 2 shows NCs between zero watermarks generated from eight images. It can be seen that the maximum value of NC was 0.6278, which indicates that the similarity between zero-watermarks was low and could prove the copyright of the image.
In addition, we randomly selected 200 images from the COCO dataset for uniqueness testing. The experimental results show that the NCs of zero-watermarks constructed from different images were all less than 0.75. Therefore, the zero-watermark of the proposed model was unique for different images and could protect the copyright of the image effectively. The main reason for this was that the shrinkage module and the feature redundancy elimination module reduced the redundant contents of images and extracted effective features to enhance the discrimination of image features.

3.2. Robustness on Trained and Untrained Attacks

In order to show the resistance of the proposed zero-watermarking model on image attacks, 200 images were randomly selected from COCO [41], VOC [42] and DIV2K [43], respectively, and the average BER of each dataset was computed.
First, the robustness of trained attacks was tested, including JPEG compression (JP), Gaussian filtering (GF), Median filtering (MF), and Gaussian noise (GN) with different strengths, as illustrated in Figure 4. For the same type of attack, when the attack strength increased, the corresponding BERs increased but were all less than 0.03. Thus, this showed the effectiveness of the proposed watermarking model for resisting trained attacks. Furthermore, except for JPEG compression, the BERs of the COCO dataset for most attacks were lower than those of the VOC and DIV2K datasets. This was mainly because the network was trained using the COCO dataset, and since image distributions are different for different datasets, the proposed method performed better on the COCO dataset. Thus, if we wanted to increase the generalization of the proposed method for different images, the trained images could be extended.
Secondly, in order to prove its generalization ability to resist different attacks, watermarking robustness on untrained attacks was tested as well. Average filtering (AF), Salt and Pepper noise (SPN), Scaling (SC), Rotation (RT), and so on were not trained in the proposed watermarking model, but the proposed watermarking method still had strong robustness, and the corresponding BERs were lower than 0.06, as illustrated in Figure 5. The main reason for this is that SRFENet extracts robust features which are not changed much after image attacks because of the feature extraction and shrinkage modules. Furthermore, the trained attacks in the noise network could represent the internal characteristics of other non-trained attacks. These experiments show that the proposed zero-watermarking method had good generalization performance in resisting different image attacks.
Moreover, in order to prove the generalization performance of untrained datasets, VOC and CIV2K datasets were tested as well. Although images from COCO were trained in the proposed network, robust features could still be extracted from the images of VOC and CIV2K datasets, and the corresponding zero watermarks were robust as well, as illustrated in Figure 4 and Figure 5. This is mainly because SRFENet learns different features from varieties of images so that the trained SRFENet can extract robust features from different datasets, which is superior to the transformed-based zero-watermarking methods.

3.3. Robustness Comparisons

At first, in order to verify the high robustness of the proposed zero-watermarking method again, different zero-watermarking methods were used for comparison. Since it is difficult to reproduce their methods completely, the comparison results are from their papers. At first, Kang’s [24] was used for comparison, as illustrated in Figure 6, where the average BER of Lena, Barba, Boats, Airplanes and Fruit was computed. Compared with Kang’s [24], although BERs of the proposed method were a little lower for some of the attacks, such as JPEG Compression (50), the Gaussian Filter (3 × 3) and Average Filter (3 × 3) were higher for most image attacks. Especially when the attack strength increased, the superiority of the proposed method was obvious, and, for instance, BER was 0.04 lower than Kang’s for Salt and Pepper noise (0.05) and Gaussian noise (0.05). The main reason for this is that the trained noises in SRFENet include high-intensity attacks, such as JPEG Compression (10), and this increases watermarking robustness on most image attacks. Moreover, the BERs of the proposed method were lower for untrained attacks, such as Salt and Pepper noise and Scaling, and this proved that the proposed method had the generalization capability of resisting different attacks again.
Second, Vellaisamy’s [17] and Zou’s [19] are used for comparison, as shown in Table 3, where the average BER of the eight images, as illustrated in Figure 3, were computed. From Table 3, we can see that, for most of the attacks, the BERs of the proposed method were lower than those of Vellaisamy’s [17] and Zou’s [19], and especially for rotation, the BER’s decrease was obvious. As the rotation angle increased, the BERs of Vellaisamy’s [17] and Zou’s [19] increased significantly, but the proposed method still performed well. This is mainly because the robust features extracted from SRFENet were stable when the images were rotated. Considering the average BER of all the attacks, the proposed method was superior to Vellaisamy’s [17] and Zou’s [19] and proved that the proposed zero-watermarking method could protect the copyright of the image effectively.
Third, hybrid attacks were used to test the compared methods, and NC was used for the robustness evaluation. As shown in Table 4, although Xiong’s [20] performed best for Median Filter (5 × 5) + JPEG Compression(10), and Kang’s [23] performed best for JPEG Compression (10) + Gaussian noise (0.3) and JPEG Compression (10) + Scaling (2.0), the NCs of the proposed method were higher than 0.95, which denotes that the proposed method can resist these attacks. Moreover, compared with Xiong’s [20] and Kang’s [23], the proposed method was better for other hybrid attacks, and the average NC of the proposed method was higher. In total, the test on hybrid attacks proved that robust features from SRFENet were effective, and the proposed method could resist hybrid attacks well.
From the above comparisons, SRFENet-based watermarking methods were better than transform-based methods. In order to further prove the effectiveness of the proposed method, Fierro-Radilla [37] was used for comparison, which is the CNN-based watermarking method. As shown in Table 5, the proposed method was higher than Fierro-Radilla’s [37] for all image attacks. The main reason for this is that too much pooling and convolution in Fierro-Radilla’s [37] reduced the main energies of the image, which affected the extraction of robust features. Considering the above comparisons, the proposed method is robust in both single attacks and hybrid attacks and performs better than existing zero-watermarking methods, including those that are transform-based and deep learning-based.

3.4. Ablation Analysis

In order to further evaluate the effectiveness of the feature extraction module, the shrinkage module, and the feature redundancy elimination module, feature maps F8 and F10 were usepd to construct a watermark, namely, Method_F8, and Method_F10, respectively. Moreover, in order to test the zero-watermarking performance generated by different features in the processes of the feature extraction module, we constructed a watermark based on the features of maps F3 and F5, namely Method_F3 and Method_F5. Method_F3 to examine the performance of the output features of the dense connection, while Method_F5 could test the performance of the zero-watermarking after the max pooling operation. The training environments and watermark construction of Method_F3, Method_F5, Method_F8 and Method_F10 were the same as those of the proposed method.
Table 6 shows the NC ratios of different ranges for different watermarking methods, wherein 200 images were randomly selected from COCO datasets for testing. From Table 6, we can see that the watermarking uniqueness of the proposed watermarking method was better compared to Method_F3, Method_F5, Method_F8 and Method_F10. The ratios of NC values above 0.75 for Method_F3, Method_F5, Method_F8 and Method_F10 were all higher than 0.3%, which meant that watermarks constructed from some images were similar. The ratios of NC values between 0.65 and 0.75 were still high for Method_F3, Method_F5 and Method_F8, which means that F3, F5 and F8 have much redundant information, so the corresponding watermarking uniqueness decreased. Method_F10 performed better than Method_F8 because the corresponding NC ratios above 0.75 and between 0.65 and 0.75 decreased to 0.0035 and 0.1228, respectively. This suggests that the redundant information of F10 was reduced compared to F8. However, compared with the proposed method, the NC ratios of NC values above 0.75 and between 0.65 and 0.75 were still higher, which meant that the redundant information of F10 still existed and the watermarking uniqueness of some images still could not be ensured after the operation of the shrinkage module. In other words, the redundant feature elimination module was proven to enhance the features effectively by removing redundant features.
In terms of robustness comparisons, although Method_F3 was the worst, it is robust to different image attacks from Table 7. This is mainly because image features that are extracted by a dense connection are robust. According to the experimental results of Method_F3, Method_F5, and Method_F8, it can be found that the average BER decreased as the number of network layers increased. Moreover, the average BER of Method_F8 was higher than that of Method_F10, which denotes that F8 had many more fragile features compared to F10 and still proves that the shrinkage module worked well on reducing fragile features. The robustness of the proposed method was similar to that of Method_F10. However, considering that the watermarking uniqueness of Method_F10 was much less than that of the proposed method, the proposed zero-watermarking method could protect the image effectively. Meanwhile, it also proved that the feature extraction module, the shrinkage module, and the feature redundancy elimination module are vital to the watermarking performance of the proposed method.

4. Conclusions

To solve the contradiction between watermarking invisibility and robustness, a zero-watermarking method based on a shrinkage and redundant feature elimination network (SRFENet) has been presented. Firstly, the feature extraction module consisting of the dense connection and the max pooling operation was designed to extract low and deep features to obtain the ability to resist different image attacks. Second, to improve the uniqueness of zero watermarking, a shrinkage module was deployed to reduce unimportant information by learning the threshold of each feature channel automatically, while the redundant feature elimination module was employed to decrease redundant features by the learning weights of inter-feature and intra-feature. Third, to improve the generalization performance of anti-attacks, different image attacks were used for training. Finally, the features extracted from SRFENet were used to build a zero watermark. The experimental results prove that the proposed zero-watermarking method was superior to existing zero-watermarking methods in terms of resisting single-image attacks and hybrid-image attacks. However, the proposed method was only suitable for the common color image but not for the high dynamic range image since it had to resist special image attacks, such as tone mapping. In future work, we will focus on this, and moreover, the vision transformers (VIT) still can be used to improve the corresponding watermarking performance.

Author Contributions

Conceptualization, Z.H. and T.L.; methodology, Z.H. and L.H.; validation, Z.H., L.H. and Y.S.; formal analysis, Z.H.; investigation, T.L.; data curation, L.H.; writing—original draft preparation, L.H.; writing—review and editing, Z.H. and T.L.; funding acquisition, Z.H. and T.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of China under Grant No. 61971247, the Zhejiang Natural Science Foundation of China under Grants Nos. LY22F020020, LQ23F010011, the Natural Science Foundation of Ningbo under Grant Nos. 2021J134, 2022J136, and the Zhejiang Province Education Department Foundation under Grant Y202248989.

Data Availability Statement

The dataset used in this paper were mainly obtained through open source channels and can be downloaded from the dataset website: http://cocodataset.org (accessed on 14 September 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. He, W.; Cai, Z. Reversible data hiding based on dual pairwise prediction-error expansion. IEEE Trans. Image Process. 2021, 30, 5045–5055. [Google Scholar] [CrossRef]
  2. Li, X.; Li, W.; Ren, D.; Zhang, H.; Wang, M.; Zuo, W. Enhanced Blind Face Restoration with Multi-Exemplar Images and Adaptive Spatial Feature Fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2706–2715. [Google Scholar]
  3. Zhang, G.; Zheng, L.; Su, Z.; Zeng, Y.; Wang, G. M-sequences and sliding window based audio watermarking robust against large-scale cropping attacks. IEEE Trans. Inf. Forensics Secur. 2023, 18, 1182–1195. [Google Scholar] [CrossRef]
  4. Yang, C.; Li, J.; Bhatti, U.A.; Liu, J.; Ma, J.; Huang, M. Robust zero watermarking algorithm for medical images based on zernike-DCT. Secur. Commun. Netw. 2021, 2021, 4944797. [Google Scholar] [CrossRef]
  5. Luo, Y.; Li, L.; Liu, J.; Tang, S.; Cao, L.; Zhang, S.; Qiu, S.; Cao, Y. A multi-scale image watermarking based on integer wavelet transform and singular value decomposition. Expert Syst. Appl. 2021, 168, 114272. [Google Scholar] [CrossRef]
  6. Zhao, Y.; Wang, C.; Zhou, X.; Qin, Z. DARI-Mark: Deep learning and attention network for robust image watermarking. Mathematics 2022, 11, 209. [Google Scholar] [CrossRef]
  7. Zhang, F.; Luo, T.; Jiang, G.; Yu, M.; Xu, H.; Zhou, W. A novel robust color image watermarking method using RGB correlations. Multimed. Tools Appl. 2019, 78, 20133–20155. [Google Scholar] [CrossRef]
  8. Li, Y.M.; Wei, D.; Zhang, L. Double-encrypted watermarking algorithm based on cosine transform and fractional Fourier transform in invariant wavelet domain. Inf. Sci. 2021, 551, 205–227. [Google Scholar] [CrossRef]
  9. Huang, L.; Xiang, Z.; Li, J.; Yao, H.; Qin, C. New Framework of Self-Embedding Fragile Watermarking Based on Reference Sharing Mechanism. Secur. Commun. Netw. 2022, 2022, 2699802. [Google Scholar] [CrossRef]
  10. Zhou, L.; Zuo, M.J.; Shi, H.; Zhang, Y.; Gong, L. Robust watermarking algorithm against the geometric attacks based on non-subsampled shearlet transform and Harris-Laplace detector. Secur. Commun. Netw. 2022, 2022, 7605595. [Google Scholar] [CrossRef]
  11. Huang, L.; Kuang, D.; Li, C.; Zhuang, Y.; Duan, S.; Zhou, X. A self-embedding secure fragile watermarking scheme with high quality recovery. J. Vis. Commun. Image Represent. 2022, 83, 103437. [Google Scholar] [CrossRef]
  12. Thanki, R.; Kothari, A.; Trivedi, D. Hybrid and blind watermarking scheme in DCuT–RDWT domain. J. Inf. Secur. Appl. 2019, 46, 231–249. [Google Scholar] [CrossRef]
  13. Mohammed, A.O.; Hussein, H.I.; Mstafa, R.J.; Abdulazeez, A.M. A blind and robust color image watermarking scheme based on DCT and DWT domains. Multimed. Tools Appl. 2023, 1–27. [Google Scholar] [CrossRef]
  14. Zhang, Y.; Gong, Y.; Wang, J.; Sun, J.; Wan, W. Towards perceptual image watermarking with robust texture measurement. Expert Syst. Appl. 2023, 219, 119649. [Google Scholar] [CrossRef]
  15. Xin, Y.; Liao, S.; Pawlak, M. Circularly orthogonal moments for geometrically robust image watermarking. Pattern Recognit. 2007, 40, 3740–3752. [Google Scholar] [CrossRef]
  16. Hosny, K.M.; Darwish, M.M.; Fouda, M.M. Robust color images watermarking using new fractional-order exponent moments. IEEE Access 2021, 9, 47425–47435. [Google Scholar] [CrossRef]
  17. Vellaisamy, S.; Ramesh, V. Inversion attack resilient zero-watermarking scheme for medical image authentication. IET Image Process. 2014, 8, 718–727. [Google Scholar] [CrossRef]
  18. Hosny, K.M.; Darwish, M.M.; Fouda, M.M. New color image zero-watermarking using orthogonal multi-channel fractional-order legendre-fourier moments. IEEE Access 2021, 9, 91209–91219. [Google Scholar] [CrossRef]
  19. Zou, B.; Du, J.; Liu, X.; Wang, Y. Distinguishable zero-watermarking scheme with similarity-based retrieval for digital rights Management of Fundus Image. Multimed. Tools Appl. 2018, 77, 28685–28708. [Google Scholar] [CrossRef]
  20. Xiong, X.G. A zero watermarking scheme with strong robustness in spatial domain. Acta Automat. Sin. 2018, 44, 160–175. [Google Scholar]
  21. Chang, C.C.; Chuang, J.C. An image intellectual property protection scheme for gray-level images using visual secret sharing strategy. Pattern Recognit. Lett. 2002, 23, 931–941. [Google Scholar] [CrossRef]
  22. Chang, C.C.; Lin, P.Y. Adaptive watermark mechanism for rightful ownership protection. J. Syst. Softw. 2008, 81, 1118–1129. [Google Scholar] [CrossRef]
  23. Kang, X.; Lin, G.; Chen, Y.; Zhao, F.; Zhang, E.; Jing, C. Robust and secure zero-watermarking algorithm for color images based on majority voting pattern and hyper-chaotic encryption. Multimed. Tools Appl. 2020, 79, 1169–1202. [Google Scholar] [CrossRef]
  24. Kang, X.; Zhao, F.; Chen, Y.; Lin, G.; Jing, C. Combining polar harmonic transforms and 2D compound chaotic map for distinguishable and robust color image zero-watermarking algorithm. J. Vis. Commun. Image Represent. 2020, 70, 102804. [Google Scholar] [CrossRef]
  25. Yang, H.; Qi, S.; Niu, P.; Wang, X. Color image zero-watermarking based on fast quaternion generic polar complex exponential transform. Signal Process. Image Commun. 2020, 82, 115747. [Google Scholar] [CrossRef]
  26. Xia, Z.; Wang, X.; Wang, C.; Ma, B.; Wang, M.; Shi, Y. Local quaternion polar harmonic Fourier moments-based multiple zero-watermarking scheme for color medical images. Knowl.-Based Syst. 2021, 216, 106568. [Google Scholar] [CrossRef]
  27. Zhang, H.; Chang, H.; Ma, B.; Wang, N.; Chen, X. Dynamic R-CNN: Towards High Quality Object Detection via Dynamic Training. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 260–275. [Google Scholar]
  28. Nguyen, D.T.; Nguyen, T.N.; Kim, H.; Lee, H. A high-throughput and power-efficient FPGA implementation of YOLO CNN for object detection. IEEE Trans. Very Large Scale Integr. (VLSI) Syst. 2019, 27, 1861–1873. [Google Scholar] [CrossRef]
  29. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Region-based convolutional networks for accurate object detection and segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 142–158. [Google Scholar] [CrossRef]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  31. Sun, Y.; Xue, B.; Zhang, M.; Yen, G.G.; Lv, J. Automatically designing CNN architectures using the genetic algorithm for image classification. IEEE Trans. Cybern. 2020, 50, 3840–3854. [Google Scholar] [CrossRef]
  32. Ding, G.; İmamoğlu, N.; Caglayan, A.; Murakawa, M.; Nakamura, R. FBNet: Feedback-Recursive CNN for Saliency Detection. In Proceedings of the 17th International Conference on Machine Vision and Applications (MVA), Online, 25–27 July 2021; pp. 1–5. [Google Scholar]
  33. Kandi, H.; Mishra, D.; Gorthi, S.R.K.S. Exploring the learning capabilities of convolutional neural networks for robust image watermarking. Comput. Secur. 2017, 65, 247–268. [Google Scholar] [CrossRef]
  34. Ahmadi, M.; Norouzi, A.; Karimi, N.; Samavi, S.; Emami, A. ReDMark: Framework for residual diffusion watermarking based on deep networks. Expert Syst. Appl. 2020, 146, 113157. [Google Scholar] [CrossRef]
  35. Luo, X.; Zhan, R.; Chang, H.; Yang, F.; Milanfar, P. Distortion Agnostic Deep Watermarking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 19 June 2020; pp. 13548–13557. [Google Scholar]
  36. Rai, M.; Goyal, S.; Pawar, M. An optimized deep fusion convolutional neural network-based digital color image watermarking scheme for copyright protection. Circuits Syst. Signal Process. 2023, 1–32. [Google Scholar] [CrossRef]
  37. Fierro-Radilla, A.; Nakano-Miyatake, M.; Cedillo-Hernandez, M.; Cleofas-Sanchez, L.; Perez-Meana, H. A Robust Image Zero-Watermarking Using Convolutional Neural Networks. In Proceedings of the 7th International Workshop on Biometrics and Forensics (IWBF), Mexico City, Mexico, 1 May 2019; pp. 1–5. [Google Scholar]
  38. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  39. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Chen, T. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef]
  40. Zhao, M.; Zhong, S.; Fu, X.; Tang, B.; Pecht, M. Deep residual shrinkage networks for fault diagnosis. IEEE Trans. Ind. Inform. 2019, 16, 4681–4690. [Google Scholar] [CrossRef]
  41. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Zitnick, C.L. Microsoft Coco: Common Objects in Context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  42. Everingham, M.; Winn, J. The pascal visual object classes challenge 2012 (voc2012) development kit. Pattern Anal. Stat. Model. Comput. Learn. 2011, 8, 1–45. [Google Scholar]
  43. Agustsson, E.; Timofte, R. Ntire 2017 Challenge on Single Image Super-Resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 126–135. [Google Scholar]
Figure 1. Zero-watermarking framework based on shrinkage and redundant feature elimination.
Figure 1. Zero-watermarking framework based on shrinkage and redundant feature elimination.
Symmetry 15 00964 g001
Figure 2. Shrinkage and Redundant Feature Elimination Network (SRFENet).
Figure 2. Shrinkage and Redundant Feature Elimination Network (SRFENet).
Symmetry 15 00964 g002
Figure 3. Eight original color images. (a) Lena. (b) Barba. (c) Boats. (d) Clown. (e) Airplane. (f) Fruit. (g) Baboon. (h) Isabe.
Figure 3. Eight original color images. (a) Lena. (b) Barba. (c) Boats. (d) Clown. (e) Airplane. (f) Fruit. (g) Baboon. (h) Isabe.
Symmetry 15 00964 g003
Figure 4. Robustness on trained noises.
Figure 4. Robustness on trained noises.
Symmetry 15 00964 g004
Figure 5. Robustness on untrained noises.
Figure 5. Robustness on untrained noises.
Symmetry 15 00964 g005
Figure 6. Robustness comparison against different image attack.
Figure 6. Robustness comparison against different image attack.
Symmetry 15 00964 g006
Table 1. Symbols and their description.
Table 1. Symbols and their description.
SymbolsDescriptionSymbolsDescription
Ioriginal imageNnoised image
FIimage feature extracted from IFNimage feature extracted from N
BIbinary map based on FIBNbinary map based on FN
W0original watermarkW1extracted watermark
ZWgenerated zero-watermark
Table 2. NCs of constructed watermarks from different images.
Table 2. NCs of constructed watermarks from different images.
LenaBarbaBoatsClownAirplaneFruitBaboonIsabe
Lena1.00000.51290.49610.61230.42260.54730.58470.5661
Barba0.51291.00000.59360.62780.45510.51800.50370.6124
Boats0.49610.59361.00000.58420.59000.56210.45460.5256
Clown0.61230.62780.58421.00000.43770.51860.42240.6218
Airplane0.42260.45510.59000.43771.00000.43800.46020.3990
Fruits0.54730.51800.56210.51860.43801.00000.52760.5746
Baboon0.58470.50370.45460.42240.46020.52761.00000.4559
Isabe0.56610.61240.52560.62180.39900.57460.45591.0000
Table 3. Robustness comparison against single image attacks.
Table 3. Robustness comparison against single image attacks.
Attack TypeBER
Proposed MethodVellaisamy’s [17]Zou’s [19]
JP (30)0.02670.26870.0497
JP (70)0.01690.09330.0253
GF (3 × 3)0.00750.00850.0125
GF (5 × 5)0.01260.01790.0136
AF (3 × 3)0.01770.01530.0134
AF (5 × 5)0.02320.05420.0173
MF (3 × 3)0.01270.01650.0147
MF (5 × 5)0.01700.04670.0183
SPN (0.01)0.01540.01360.0591
SPN (0.02)0.01780.01380.0744
GN (0.005)0.01510.05760.072
GN (0.01)0.02310.08120.0906
SC (0.5)0.01820.00910.0149
SC (2.0)0.01530.0030.0086
RT (3°)0.04730.13110.0875
RT (5°)0.06220.19120.1198
Average0.02180.0639 0.0432
Table 4. Robustness comparison against hybrid attack.
Table 4. Robustness comparison against hybrid attack.
Attack TypeNC
Proposed MethodXiong’s [20]Kang’s [23]
MF (5 × 5) + SPN (0.3)0.96920.8901——
MF (5 × 5) + GN (0.3)0.95980.95010.9366
MF (5 × 5) + JP (10)0.96350.97620.9815
JP (10) + SPN (0.3)0.93220.8929——
JP (10) + GN (0.3)0.95190.95510.9361
RT (2°) + JP (10)0.94280.86280.8920
JP (10) + SC (2.0)0.96920.98560.9694
Average0.95510.93040.9431
Table 5. Robustness comparison with Fierro-Radilla [37].
Table 5. Robustness comparison with Fierro-Radilla [37].
Attack TypeBER
Proposed MethodFierro-Radilla’s [37]
JP (30)0.02670.0288
JP (70)0.01690.0207
MF (3 × 3)0.01270.0214
MF (5 × 5)0.01700.0521
GN (0.002)0.01370.0324
GN (0.02)0.02620.0813
SPN (0.01)0.01540.0217
SPN (0.03)0.01840.0431
RT (1°)0.02750.0424
RT (5°)0.06220.0793
JP (10) + RT (2°)0.07140.0821
MF (7 × 7) + SPN (0.02)0.04250.0804
Average0.02920.0488
Table 6. Uniqueness comparison of zero-watermarking constructed by different methods.
Table 6. Uniqueness comparison of zero-watermarking constructed by different methods.
Range of NCNC Ratio
Proposed Method_F3Method_F5Method_F8Method_F10
Above 0.750.00%0.62%0.58%0.67%0.36%
Between 0.65 and 0.758.26%25.35%23.46%21.27%12.28%
Between 0.5 and 0.6545.27%48.37%54.38%51.23%56.32%
Below 0.546.47%25.66%21.58%26.83%31.04%
Table 7. Robustness comparison of zero-watermarking constructed by different methods.
Table 7. Robustness comparison of zero-watermarking constructed by different methods.
Attack TypeBER
ProposedMethod_F3Method_F5Method_F8Method_F10
JP (50)0.02110.03100.03130.03010.0204
JP (90)0.01080.02780.01960.01920.0097
GF (5 × 5)0.01590.02850.02670.02290.0221
GF (9 × 9)0.02630.03560.02960.02340.0224
AF (5 × 5)0.02460.02870.03210.03300.0281
AF (9 × 9)0.05480.07620.07670.07300.0571
MF (5 × 5)0.01450.03150.03100.03210.0264
MF (9 × 9)0.04090.07250.06430.06010.0443
SPN (0.01)0.01920.02620.02580.02460.0205
SPN (0.05)0.02560.04100.03860.04350.0243
GN (0.01)0.02040.02520.01960.02040.0214
GN (0.05)0.03450.03600.03680.03660.0337
SC (0.5)0.01640.02110.02780.02390.0127
SC (0.8)0.01470.02060.02330.02140.0112
RT (3°)0.04530.06860.06360.06810.0545
RT (5°)0.05080.09350.07100.07330.0613
Average0.02720.04150.03860.03790.0294
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, L.; He, Z.; Luo, T.; Song, Y. Shrinkage and Redundant Feature Elimination Network-Based Robust Image Zero-Watermarking. Symmetry 2023, 15, 964. https://doi.org/10.3390/sym15050964

AMA Style

He L, He Z, Luo T, Song Y. Shrinkage and Redundant Feature Elimination Network-Based Robust Image Zero-Watermarking. Symmetry. 2023; 15(5):964. https://doi.org/10.3390/sym15050964

Chicago/Turabian Style

He, Lingqiang, Zhouyan He, Ting Luo, and Yang Song. 2023. "Shrinkage and Redundant Feature Elimination Network-Based Robust Image Zero-Watermarking" Symmetry 15, no. 5: 964. https://doi.org/10.3390/sym15050964

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop