Abstract

We have proposed an image adaptive watermarking method by using contourlet transform. Firstly, we have selected high-energy image blocks as the watermark embedding space through segmenting the original image into nonoverlapping blocks and designed a watermark embedded strength factor by taking advantage of the human visual saliency model. To achieve dynamic adjustability of the multiplicative watermark embedding parameter, the relationship between watermark embedded strength factor and watermarked image quality is developed through experiments with the peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM), respectively. Secondly, to detect the watermark information, the generalized Gaussian distribution (GGD) has been utilized to model the contourlet coefficients. Furthermore, positions of the blocks selected, watermark embedding factor, and watermark size have been used as side information for watermark decoding. Finally, several experiments have been conducted on eight images, and the results prove the effectiveness of the proposed watermarking approach. Concretely, our watermarking method has good imperceptibility and strong robustness when against Gaussian noise, JPEG compression, scaling, rotation, median filtering, and Gaussian filtering attack.

1. Introduction

Transmitting and sharing digital multimedia have become more convenient with the rapid development of the network. However, such phenomenon results in security issues, such as authentication, copyright protection, and fingerprinting [17]. Digital watermarking can be used as an effective method to address these problems. Generally, in the watermarking process, some useful information (e.g., watermark data) is embedded into an original signal while ensuring its quality. Furthermore, robustness and imperceptibility are the main factors in digital image watermarking. Many image watermarking algorithms have been presented in the literature. On the basis of the embedding method, most algorithms can be divided into three categories, namely, additive, quantization, and multiplication-based watermarking algorithms.

For the additive-embedding watermarking approach, the watermark information is directly added to the host image coefficients or image block of the same size. Generally, the coefficients can be obtained from some common transforms, including discrete wavelet transform (DWT), discrete cosine transform (DCT), and Fourier transform. The additive-embedding watermarking embeds the watermark information in the most important frequency domain of image perception, which is similar to the spread spectrum communication idea in the communication system. Cox et al. [8] first designed a digital watermarking method based on the idea of spread spectrum, which embedded watermark data in the important perception transformation coefficient of the host signal by applying the spread spectrum principle. Cox’s spread spectrum watermarking algorithm has been considered a representative method. The only deficiency is that the digital watermarking algorithm requires participation of the original image when detecting watermark information, indicating that it is not a blind watermarking algorithm. Subsequently, Cheng et al. [9] proposed an additive watermarking approach, which detects the watermark by using the generalized Gaussian distribution (GGD). Experiments show that this distribution can effectively control the detection error probability of the watermark. Liu et al. [10] transformed the test signal into the DCT domain. Moreover, a local optimal detection model that is suitable for any host signal was derived by conducting hypothesis testing analysis in this domain. Although these methods [9, 10] can detect watermark information effectively, their parameter estimation process is complex. To address this problem, Kwitt et al. [11] proposed a lightweight blind optimal detector for additive watermarking; it is expected to be useful in resisting watermark desynchronization. Zhang et al. [12] proposed a high-security additive watermarking algorithm by utilizing gyrator transform and matrix decomposition. A key innovation of this algorithm is to adopt an invariant integer wavelet transform, which transforms the image wavelet coefficients into integers, thereby enhancing the performance of the watermarking.

In the quantization-based watermarking method, the main procedure is to embed the watermark data into the host signal by designing a corresponding quantizer. The watermark data are detected according to the quantization interval of the image transform coefficient to extract watermark. Many watermarking methods with quantization scheme have been proposed in recent years. Chen et al. proposed a digital watermarking method with the quantization index modulation (QIM) scheme; it is the most representative quantization watermarking algorithm based on edge information coding [13]. QIM has the characteristics of high capacity, blind detection, and simple implementation. However, QIM watermarking has two main shortcomings. First, it is sensitive to amplitude scaling attacks; second, it is not robust to gain attacks. Researchers proposed corresponding improvement methods to address these problems. In view of the sensitivity of the QIM watermarking method to scaling attacks, researchers mainly improved it in accordance with the quantization step size. To solve the inconsistency of quantization step between the embedded end and the receiver end, as well as the adaptability problem of quantization step, several watermarking methods have been proposed, such as rational dither modulation [14] and adaptive QIM [15]. Furthermore, to enhance the robustness of the QIM watermarking against gain attack, the quantization watermarking [16], sample projection-based quantization [17], P-norm ratio-based quantization [18], angle quantization [19], complex wavelet domain l1 norm quantization [20], and random projection-based quantization methods [21] have been proposed one after another. These quantization watermarking methods mainly aim to enable the watermark algorithm to obtain invariance to the scaling or gain attacks, and the watermark has strong robustness performance in resisting compression, filtering, and gain attacks. However, the performance of these quantization methods in desynchronization attacks is still inadequate. To further enhance the robustness of quantization watermarking, some researchers have designed corresponding quantization watermarking algorithms by combining the just noticeable distortion (JND) model, image texture complexity, and texture direction features, such as texture direction quantization [22], pair quantization based on extended JND [23, 24], and mixed modulation quantization using singular value decomposition [25]. These quantization methods are combined with image features; they can retain image orientation features and reduce image distortion after watermark embedding. However, these methods are generally vulnerable to noise attack.

The performance of the multiplicative embedding-based watermarking method is similar to that of the quantization watermarking method. The multiplicative watermarking algorithm is usually combined with the human visual perception model, and the embedded strength factor varies with the intensity of the original signal. Moreover, a good trade-off between imperceptibility and robustness can be achieved in the multiplicative watermarking algorithm. Akhaee et al. [26] developed an image watermarking method based on a “scaling” strategy by using the Watson entropy visual masking. The watermark data were embedded into the image block with high entropy to improve the invisibility of the watermark. The algorithm is robust against Gaussian filtering, Gaussian noise, and scaling attacks. However, the entropy value of the image block changes after embedding the watermark; this finding is inconsistent with the entropy of the image block prior to embedding the watermark, thereby reducing the robustness of the watermark against synchronization attacks. Subsequently, Akhaee et al. [27] proposed a scaling-based image watermarking method with contourlet transform in a noisy environment. Experiments demonstrated that the robustness of this watermarking method is good. However, the algorithm has high complexity. Different from the Watson entropy visual masking, Khalilian et al. [28] proposed a multiplicative watermarking algorithm by taking advantage of the visual saliency model. They designed an adaptive embedding factor by combining visual saliency and texture masking. On the one hand, the embedding factor should increase with the distance from the significant region of the image. On the other hand, the watermark embedding strength should be larger in regions with rich texture. This method improves the robustness of the watermarking when against some common image processing attacks. However, the performance of their watermarking still needs to be enhanced in terms of resisting antidesynchronization attack. Moreover, some visual attention-based watermarking methods have been presented in the last few years. For example, Bhowmik et al. [29] embedded high-strength and low-strength watermarks into significant and insignificant regions of vision, respectively, thereby improving the watermarking performance. Hernandez et al. [30] proposed a video watermarking algorithm that took full advantage of the video’s spatiotemporal characteristics and minimized the perceived redundancy of the video. Thus, the trade-off between imperceptibility and robustness has been achieved in their method. Yadav et al. [31] developed an image watermarking algorithm by using an adaptive embedded factor, which only used image variance information to compute watermark embedded factor. However, the performance of this method is weak when resisting rotation attacks.

Inspired by literature [28], an image watermarking algorithm was developed based on the visual saliency model in the contourlet transform domain. The main contributions of our work are summarized as follows:(1)An adaptive watermark embedded strength factor is exploited with a visual saliency model, which can achieve a good trade-off between the robustness and imperceptibility of the watermarking.(2)The watermark information was embedded into the contourlet coefficients with high energy that can enhance the imperceptibility of watermarking.

The remainder of this paper is organized as follows. The belief concept of the contourlet transform is introduced in Section 2. Section 3 introduces the proposed watermark embedding and detection method. Section 4 shows the experimental results of the proposed watermarking and the comparative results with other watermarking approaches. Finally, the conclusions are summarized in Section 5.

2. Brief Introduction of Contourlet Transform

In 2005, Do et al. [32] proposed a “real” 2D representation of images, that is, the contourlet transform. It captures the segmented conic curves of an image by using different subband scales and frequencies, which have directivity and anisotropy, thereby enabling the contourlet transform to obtain a “sparser” representation. Thus, the contourlet transform has the characteristics of sparse representation at both spatial and directional resolutions. In contourlet transform, multiscale and directional analyses are performed separately. First, the image was transformed into one coarse version plus a set of band-pass images by the Laplacian pyramid (LP) method. Second, each LP band-pass image was decomposed into a number of subbands with 2D quincunx filtering and critical subsampling. Therefore, the contourlet transform can decompose images into multidirectional subbands at multiple scales. Figure 1 illustrates a diagram of the contourlet transform. Furthermore, we have utilized the contourlet toolbox to decompose the “Peppers” image [32]. Figure 2 shows the result of applying the contourlet transform on the “Peppers” image. The figure clearly shows that the contourlet transform can decompose the “Peppers” image into multidirectional subbands.

3. Watermark Embedding and Decoding

In this section, Figure 3 shows the proposed watermark embedding and watermark detection procedure. As shown in Figure 3, we embed the watermark data into the contourlet coefficients with high energy in our implementation. In addition, we utilize the visual saliency model to construct the watermark embedded strength factor; thus, a trade-off between the invisibility and robustness of the watermark can be achieved elegantly with the watermark embedded strength factor. In the watermark detection stage, we model the contourlet coefficients with GGD to detect the watermark due to the non-Gaussian property of the contourlet coefficients.

3.1. Proposed Watermark Embedding

The procedure of the proposed watermark embedding in Figure 3(a) can be generalized as follows:Step 1: We segment the host image into blocks and select the first image blocks with high energy. The energy is calculated as the sum of the squares of the absolute values of the pixels of the image block. Consequently, the energy of block [28] can be computed by , where denotes the size of the image block and represents the positions of image block. Generally, a larger value of the energy of image block implies that this image region contains more important coefficients and should be considered a significant image block in comparison with other image blocks. Therefore, to improve the robustness of the watermarking, the watermark is embedded into the image blocks with high energy.Step 2: Then, we decompose each selected image block by using a two-level contourlet transform. Thus, we embed the watermark data into the coefficients of low-frequency subband. The host contourlet coefficient vectors are denoted as , and the watermarked contourlet coefficient vectors are denoted as . Suppose that the watermark is with n components and; the watermark embedding process can be expressed as follows:where denotes the embedded strength factor and its value was calculated in Section 3.2.Step 3: Repeat Step 2 for each image block.Step 4: Two-level inverse contourlet transform on the watermarked image subband is performed, and it is combined with the image subbands, which are not embedded watermark information, to obtain the whole watermarked image.

3.2. Watermark Embedded Strength Factor

The JND threshold has been widely applied in the field of image processing. Its value is often higher in the image texture region [33]. On the basis of [33], the image texture region can hide more information without being perceived by human eyes. Therefore, the embedded strength factor can select a high value. Literature [27] used this fact to develop an image watermarking algorithm. Literature [34] shows that the human visual system tends to focus on the salient areas of an image. As a result, the image salient area hides more distortion, and the embedded strength factor can be enhanced. Therefore, to calculate the embedded strength factor, we take advantage of the texture masking and visual saliency model in this study. The calculation process is summarized as follows.

First, we use a two-level contourlet transform to decompose the host image, which obtains a low-frequency subband, four subbands, and eight subbands from the coarsest scale to the finest scale (Figure 2). Therefore, we compute the energy of directional subband of each block according to the property of the image texture masking. The calculation can be expressed as follows:where is the i-th directional subband’s energy of each image block. Each image block has 12 directional subbands after the two-level decomposition with contourlet transform. Suppose that denotes the average energy of twelve image blocks. When increasing the average energy, the watermark embedded strength factor could increase correspondingly. Hence, according to [27], the watermark embedded strength factor of the high-frequency part can be written as follows:where , , and are set to 1.025, 0.02, and , respectively. These parameters are determined by experimental simulation. In the right part of equation (3), for , the parameter is set to 1.025 for larger image energy , when the exponential function vanishes. This parameter is set to 1.025 in our experiments mainly because it maintains the imperceptibility of the image when used in high image energy when the exponential term disappears. On the contrary, for small image energy, we set parameter to 0.02. Parameter has an important effect in the increasing rate of watermark embedded strength factor; its value is set to mainly because it can achieve a good trade-off between the robustness and imperceptibility of the watermarking. Therefore, the parameter setting of the watermark embedded strength factor is mainly based on the size of image energy. The main reason is to embed the watermark information while maintaining the imperceptibility of the image watermark.

Then, inspired by [35], we modified the embedded strength factor, which is denoted by by applying visual saliency. Suppose that represents the saliency distance of each block and denotes the maximum saliency distance in all image blocks. Therefore, the watermark embedded strength factor can be expressed by . Finally, the modified watermark embedded strength factor can be represented as follows:

3.3. Watermark Decoding

In this section, we model the contourlet coefficients by the GGD. The probability density function of the GGD model is represented as follows:where , and denote the mean value and variance, respectively. is the gamma function when , and denotes the shape parameter. Watermark detection can perform the detection and evaluation of signals. The hypothesis test can be drawn as follows, using the likelihood ratio test (LRT):where and are the null and alternative hypotheses. According to the statistical signal processing method, the maximum likelihood ratio can be represented as follows:

Proofs of (7) are as follows [36]:

Let . Then, the integrand substitutes for ; (8) can be rewritten as follows:

One order Taylor series of () around is expanded as follows:

Therefore, (10) is rewritten as follows:

, is approximately zero in (11). Therefore, equation (11) can be further expressed as follows:

Therefore, .

On the basis of the analysis, we can rewrite the maximum likelihood ratio by combining watermark embedding (1) and the GGD model as follows:

Furthermore, equation (13) can be simply represented as follows:

Thus, we can write the watermark detector as follows:

Next, we can compute the watermark detection threshold. The Gaussian distribution characteristic of the watermark detector under the null hypothesis condition and its mean is zero. As a result, we can calculate the watermark detection threshold as follows:where denotes the watermark detection threshold,  represents the variance, and denotes the right-tail probability of the Gaussian distribution. represents the false alarm probability.

Generally, false alarm is generated due to the existence of the watermark information detected in the unwatermarked image. A missed alarm is the phenomenon in which the watermark detector does not detect the watermark information in the watermarked image. Therefore, the receiver operating characteristic (ROC) curve of the watermarking can be derived as follows.

Suppose denotes the detection probability of watermark. Hence, can represent the missed alarm probability. According to the statistical hypotheses and the central limit theorem, the mean and variance of the distribution of the host image and watermarked image can be estimated; they are denoted as , and , respectively. As a result, can be written as follows:where ,  = 0, and . Let .

Finally, the ROC relationship can be defined as follows:

4. Experimental Results

In this regard, to verify the effectiveness of the proposed watermarking method, several experiments have been performed, including the imperceptibility, robustness, and performance of watermark detection. We have compared the proposed watermarking with other related watermarking approaches. All experiments have been performed on a PC with 4.0 GHz Intel Core i7 CPU and 16 G RAM. The simulation software was MATLAB R2018a that ran in 64-bit Windows 10. In summary, the simulation settings are provided in Table 1.

4.1. Imperceptibility Test

We have tested eight standard images, which include Lena, Barbara, Bridge, Boat, Elaine, Mandrill, Peppers, and Man, to demonstrate the invisibility of the proposed method; the size of each standard image is 512  512. In our implementation, a two-level contourlet transform has been applied to decompose each image block. The filters are set to “Pivka.” Figure 4 only shows the host images and their watermarked version made by applying our method with 16  16 blocks and a 512-bit watermark capacity due to the limited space. Figure 4 shows that the imperceptibility of our method is satisfied. Therefore, finding the difference between the original image and their watermarked version is difficult.

In addition, the embedded strength factor can be adapted and adjusted according to the watermark capacity to further enhance the performance of the proposed method. The relationship between the embedded parameter and watermarked image quality is developed through experiments, and the results are shown in Figures 5 and 6 . The performance is mainly measured by peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) [37]. As shown in Figures 5 and 6, when the watermark embedded strength factor increases, the values of PSNR and SSIM decrease. The range of embedded strength factor can be set within 0.005 to 0.025 to balance the imperceptibility, robustness, and watermark capacity of watermarking.

4.2. Robustness Test

In this section, to assess the robustness of the proposed watermarking, several experiments have been performed in common image processing and some geometric attacks. These attacks include additive whiten Gaussian noise, salt and pepper noise, median filtering, rotation, cropping, flipping, scaling, JPEG compression, and Gaussian filtering attack. Furthermore, to evaluate the effectiveness of our watermarking method, we have compared it with other related watermarking approaches, which include the methods in [27, 31] and [38]. Moreover, the robustness performance is measured through the bit error rate (BER).

We have performed two common experiments under JPEG compression and Gaussian noise attack. The result is shown in Figures 7 and 8. In this work, the watermark capacity is 512 bits. Figure 7 shows that our method has satisfying robustness on JPEG compression attack. Similarly, Figure 8 shows that our watermarking method has good robustness against Gaussian noise attacks. Moreover, Tables 2 and 3 show the comparison of the performance of our method with other methods under common image processing, geometric, and combined attacks. All watermarking methods, for the purpose of comparison, use the same watermark capacity. The watermark capacity of all methods is also 512 bits in Tables 2 and 3.

Table 2 shows the results of the simulation experiments under common image processing attacks, which cover Gaussian noise with noise variance 20, Salt and Pepper noise with noise variance of 0.05, JPEG compression with a quality factor of 20%, Gaussian filtering with the windows of size 3 × 3, and median filtering with the windows of size 5 × 5. Table 3 shows the results of the simulation experiments under geometric attacks, including the rotation attack with 10° angle, amplitude scaling attack with factor 0.75, cropping with factor 50%, combination attack with rotation of 5° angle, and scaling with a factor of 0.50.

As shown in Tables 2 and 3, the proposed watermarking method has slightly better performance than the image watermarking methods. This finding is mainly due to the application of the following factors. First, we embed the watermark information into the image blocks with high energy in the contourlet transform domain. Second, the watermark embedded strength factor was constructed by taking advantage of the visual saliency model and texture masking. Thus, embedding the watermark can be adapted. As such, a good trade-off between the invisibility and robustness of the watermark can be achieved. Finally, the watermark detection performance can be improved by the GGD model.

However, the proposed watermark detector relies on partial original image feature information, such as positions of image blocks; thus, the proposed algorithm becomes semiblind. In the subsequent work, we will design a blind watermarking method.

4.3. Performance of Watermark Detection

The GGD is used to model the contourlet coefficients to further demonstrate the detection performance, and the ROC is utilized to measure the performance of watermark detection according to equation (18) of Section 3.3. Figure 9 shows the results and indicates that the detection performance of our method is satisfied. The main reason is that the contourlet coefficient distribution is highly nonlinear, and the GGD fits the contourlet coefficient effectively.

However, the proposed watermarking method performs weakly when resisting other attacks, including combinational attack amplitude scaling and JPEG compression, Salt and Pepper and Gaussian noise, and global affine transformation and histogram equalization attack. These problems will be addressed by developing some matrix decomposition-based watermarking methods or deep learning-based watermarking algorithms in our future work.

5. Conclusion

We have developed an image watermarking algorithm by using the visual saliency model in the contourlet domain. In watermark embedding, high-energy image blocks are selected for the watermark embedding space, and the watermark embedded strength factor is exploited by taking advantage of texture masking and visual salience. The watermark can be embedded into the contourlet coefficients adaptively by using this strategy. For watermark decoding, the GGD model is used to describe the contourlet coefficients, and the ROC has been derived by applying the statistic signal processing method. Finally, we have performed several experiments to demonstrate the proposed method. Simulation results show that our watermarking method has satisfied imperceptibility and robustness. In the future work, a novel watermark detection approach will be designed using the deep learning or generative adversarial network method.

Data Availability

Eight standard grayscale images Lena, Barbara, Bridge, Boat, Elaine, Mandrill, Peppers, and Man are used as host images in the simulations, which are shown in Figure 4 in this paper. The results in this paper are entirely theoretical and analytical. The main steps of the demonstrations for each result are clearly reported in the text and the paper is fully consistent without the support of any additional data.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the Natural Science Foundation of Jiangxi (no. 20192BAB207013). The authors would like to thank professor M.N. Do for providing the CODE to perform the contourlet transform.