The Scientific World Journal

The Scientific World Journal / 2014 / Article
Special Issue

Recent Advances in Information Security

View this Special Issue

Research Article | Open Access

Volume 2014 |Article ID 454867 | https://doi.org/10.1155/2014/454867

Shang-Lin Hsieh, Chun-Che Chen, Wen-Shan Shen, "Combining Digital Watermarking and Fingerprinting Techniques to Identify Copyrights for Color Images", The Scientific World Journal, vol. 2014, Article ID 454867, 14 pages, 2014. https://doi.org/10.1155/2014/454867

Combining Digital Watermarking and Fingerprinting Techniques to Identify Copyrights for Color Images

Academic Editor: Fei Yu
Received15 Mar 2014
Accepted20 May 2014
Published08 Jul 2014

Abstract

This paper presents a copyright identification scheme for color images that takes advantage of the complementary nature of watermarking and fingerprinting. It utilizes an authentication logo and the extracted features of the host image to generate a fingerprint, which is then stored in a database and also embedded in the host image to produce a watermarked image. When a dispute over the copyright of a suspect image occurs, the image is first processed by watermarking. If the watermark can be retrieved from the suspect image, the copyright can then be confirmed; otherwise, the watermark then serves as the fingerprint and is processed by fingerprinting. If a match in the fingerprint database is found, then the suspect image will be considered a duplicated one. Because the proposed scheme utilizes both watermarking and fingerprinting, it is more robust than those that only adopt watermarking, and it can also obtain the preliminary result more quickly than those that only utilize fingerprinting. The experimental results show that when the watermarked image suffers slight attacks, watermarking alone is enough to identify the copyright. The results also show that when the watermarked image suffers heavy attacks that render watermarking incompetent, fingerprinting can successfully identify the copyright, hence demonstrating the effectiveness of the proposed scheme.

1. Introduction

Many researchers [117] have been engaged in finding the solution to protecting copyrights of digital images, which may be duplicated and distributed over the Internet without the authors’ permission. Generally speaking, there are two approaches to discovering image copyright infringement. One is watermarking [111] and the other is fingerprinting [1217]. The main idea of watermarking is to embed a piece of information (i.e., watermark) in the host image. If a similar watermark can be retrieved from a suspect image, it is then considered a duplicated one. On the other hand, the principle of fingerprinting is to extract unique features (i.e., fingerprints) from both the host image and the suspect one for comparison. If their fingerprints are similar, the ownership of the image can then be confirmed.

There are some general considerations on the two techniques, including the processing time and robustness. In terms of processing time, watermarking is more efficient because fingerprinting needs extra time to compare the image’s fingerprint with those stored in the database. If the database is large, it will be very time consuming. On the other hand, fingerprinting is generally more robust [18, 19] because when a watermarked image suffers some image processing operations that modify the content of the image, the embedded watermark will usually be damaged or even destroyed. On the contrary, since normal image processing does not destroy the features of an image, the fingerprint of the image can therefore be preserved. In summary, fingerprinting is more robust whereas watermarking is more efficient. If the complementary natures of two approaches can be utilized properly, a robust and efficient scheme can then be developed to identify copyrights.

This paper proposes a novel scheme that combines the two techniques to identify copyrights for color images. The proposed scheme generates from the image a fingerprint, which also serves as the watermark. The watermark is then embedded in the host image to produce a watermarked image. When there is a dispute over the copyright of a suspect image, the suspect image will first be processed by watermarking, which tries to retrieve the watermark from the suspect image. If the watermark is identified, the copyright is confirmed at this stage; otherwise, the image will then be processed by fingerprinting, which utilizes the retrieved watermark as the fingerprint and compares it with those stored in the database. If a match is found, then the suspect image will be considered a duplicated one.

The proposed scheme utilizes a special technique called image secret sharing (ISS), whose details can be found in the paper [20] we published in 2008. The following briefly describes the main idea of the technique utilized by the proposed scheme.

The ISS generates a share image from two images. In the proposed scheme, the two images are the logo image and the feature image (as depicted in Figure 1). The logo image can be any identifiable image. The feature image is generated from the input image as follows. First, the input image is split into nonoverlapping 8 × 8 blocks. Then, the 2D DWT is applied to each block to generate four subbands, LL2, LH2, HL2, and HH2. An example of 2D DWT is shown in Figure 2. Finally, the coefficients in the LL2 subband of each DWT block are used to generate the feature image. The ISS then generates a share image from the logo image and feature image. The share image will be used as the fingerprint of the input image by the scheme. The share image also serves as the watermark to be embedded in the host image. The benefit of the ISS scheme is that performing the XOR operation on the feature image and the share image will restore the logo image, which can then be used to identify the copyright.

The proposed scheme contains two phases: the fingerprint and watermarked image generation phase and the authentication logo detection phase. The former phase extracts features from the host image, which, along with a logo image, is used to generate the fingerprint. The fingerprint also serves as the watermark, and the phase embeds it in the host image to produce a watermarked image. On the other hand, the latter phase extracts features and retrieves the watermark from the suspect image. The extracted features and the retrieved watermark are utilized to restore the logo image, which is used to identify the copyright. If it fails, the retrieved watermark then serves as the fingerprint and is compared with those in the database to determine if the suspect image is a duplicated one.

The fingerprint and watermarked image generation phase (shown in Figure 3) works as follows. In the beginning, feature extraction extracts the features of the host image and then logo scrambling disarranges the authentication logo to a scrambled logo image. After that, fingerprint generation takes as input the extracted features and the scrambled logo to generate the fingerprint. Finally, the fingerprint serves as a watermark and is embedded in the host image, which becomes a watermarked image. The fingerprint is also stored in a database for later use in the next phase.

The authentication logo detection phase (shown in Figure 3) checks the watermark first and, if necessary, the fingerprint next. In the beginning, watermark retrieval regains the watermark from the suspect image. Next, the features of the suspect image are extracted by feature extraction. After that, logo restoration takes as input the retrieved watermark (the expected fingerprint of the suspect image) and the extracted features to recover and rearrange the scrambled logo to restore the authentication logo. The phase ends if the accuracy rate of the restored logo determined by logo comparison is high enough; otherwise, the process proceeds to retrieve the next available fingerprint from the database and then returns to logo restoration, which takes as input the retrieved fingerprint instead of the extracted watermark. The phase restores the logo from the retrieved fingerprint as well as the extracted features and proceeds to logo comparison. The looping process continues until the authentication logo is discovered or no fingerprint is available.

3.1. Fingerprint and Watermarked Image Generation Phase

The following paragraphs detail the stages in the fingerprint and watermarked image generation phase, including feature extraction, logo scrambling, fingerprint generation, and watermark embedding.

3.1.1. Feature Extraction

The feature extraction stage takes a color image as input and then extracts its features. The stage has two substages, sampling and feature generation. During sampling, the stage first transforms the RGB image to the YCbCr color space [21, 22]. Then, it partitions each of the three channels into several nonoverlapping blocks of size 8 × 8 (hence, a color image of size will result in nonoverlapping blocks). After partition, for each row of the corresponding blocks in the three channels, the stage takes the first four samples from the Y channel, the next two from the Cb, and the last two from the Cr (as depicted in Figure 4) to generate new packed blocks, each of size 8 × 8.

After sampling, the stage enters its second substage, feature generation. For each of the packed blocks, the stage applies 2D DWT to the block, resulting in four coefficients in the LL2 subband. Next, the stage computes the average (denoted by ) of the four coefficients and then obtains a feature type according to the relationship of the four coefficients and the average as expressed in (1). Consider

According to , , and the mapping table shown in Table 1, a feature share (called FT-share) of size 2 × 2 is determined for each block. The FT-shares represent the features of the input color image. They are assembled to form the feature image.


Feature type Mean value positionWhite logo pixel FT-share
XOR
FP-share 
Black logo pixel FT-share
XOR
FP-share
FT-shareFP-shareFT-shareFP-share

 1  454867.table.001454867.table.002454867.table.003454867.table.004454867.table.005454867.table.006
454867.table.007454867.table.008454867.table.009454867.table.0010454867.table.0011454867.table.0012
454867.table.0013454867.table.0014454867.table.0015454867.table.0016454867.table.0017454867.table.0018
454867.table.0019454867.table.0020454867.table.0015454867.table.0022454867.table.0023454867.table.0024

2 454867.table.0025454867.table.0026454867.table.0027454867.table.0028454867.table.0029454867.table.0030
454867.table.0031454867.table.0032454867.table.0033454867.table.0034454867.table.0035454867.table.0036
454867.table.0037454867.table.0038454867.table.0039454867.table.0040454867.table.0041454867.table.0042
454867.table.0043454867.table.0044454867.table.0045454867.table.0046454867.table.0047454867.table.0048
454867.table.0049454867.table.0050454867.table.0051454867.table.0052454867.table.0053454867.table.0054
454867.table.0055454867.table.0056454867.table.0057454867.table.0058454867.table.0059454867.table.0060

 3 454867.table.0061454867.table.0062454867.table.0063454867.table.0064454867.table.0065454867.table.0066
454867.table.0067454867.table.0068454867.table.0069454867.table.0070454867.table.0071454867.table.0072
454867.table.0073454867.table.0074454867.table.0075454867.table.0076454867.table.0077454867.table.0078
454867.table.0079454867.table.0080454867.table.0081454867.table.0082454867.table.0083454867.table.0084

4 454867.table.0085454867.table.0086454867.table.0087454867.table.0088454867.table.0089454867.table.0090

454867.table.0091 The coefficients of the LL2 subband (a is at the top left position, b the top right, c the bottom left, and d the bottom
right)

FT-share: feature share; FP-share: fingerprint share.

The steps of the feature extraction stage are listed in Algorithm 1.

Input:   A color image H (N × N).
Output:   A feature image FT (N/8 × N/8).
Convert H to the YCbCr color space
Partition each of the Y, Cb, and Cr channels into N/8 × N/8 non-overlapping blocks of size 8 × 8
For each corresponding block of the three channels
    Take the first 4 samples from the Y channel, the next 2 from the Cb, and last 2 from the
    Cr as depicted in Figure 4 to form a packed block of size 8 × 8
End For
Apply 2D-DWT to each packed block to obtain N/8 × N/8 LL2 blocks of size 2 × 2
For each of the LL2 blocks
    Compute the average A of the four coefficients
    Obtain the feature type T according to (1)
    Determine the FT-share according to T, A, and Table 1
End For
Assemble the FT-shares to form the feature image FT

3.1.2. Logo Scrambling

In order to disperse the intensity of attacks, the proposed scheme adopts Torus automorphism [23] to scramble the authentication logo. The stage uses a predetermined key, , and the following equation to scramble the logo. Consider where (, ) are the coordinates in state t and is the coordinate size of the given image. Figure 5 shows an example of the authentication logo scrambled four times with , .

3.1.3. Fingerprint Generation

The proposed scheme uses ISS (mentioned in Section 2) to generate the fingerprint. It determines a FP-share for each FT-share (generated in feature extraction) according to the color of the corresponding pixel in the scrambled logo by looking Table 1 up. For example, if the feature type is 2 and the FT-share is the same as the one in Table 2, then the FP-share will be either one of the two FP-shares in Table 2 according to the color of the corresponding pixel in the scrambled logo. After every FP-share is determined, the stage gathers all the FP-shares to form the fingerprint image. The fingerprint also serves as the watermark in the next stage.


Feature type Mean value positionWhite logo pixel FT-share
XOR
FP-share 
Black logo pixel FT-share
XOR
FP-share
FT-shareFP-shareFT-shareFP-share

2 454867.table.0038454867.table.0037454867.table.003454867.table.0038454867.table.0038454867.table.006

3.1.4. Watermark Embedding

The watermark embedding stage uses the resulting fingerprint image as a watermark and then embeds it in the host image. Generally speaking, for the RGB color space, the human visual system is more sensitive to the G channel than to the other two [2228]. Therefore, the proposed scheme embeds the watermark into the less sensitive R and B channels of the host image. To be more specific, there are four areas: two in the R channel ( and ) and two in the B channel ( and ), used for watermark embedding (see Figure 6). The stage applies the 2D DWT to the R and B channels of the host image and next applies the 1D DWT to the resulting HL2 and LH2 to obtain the four blocks, , , , and .

The proposed scheme embeds the watermark into the image by adjusting the coefficients in the to blocks according to a predefined . The value of affects the robustness and the quality. The larger the is, the more robust the embedded watermark is, but the worse the quality of the watermarked image is. Figure 7 illustrates the adjustment of the coefficients. If the watermark bit is 1 and the current coefficient is between and , then the coefficient will be adjusted to be or , whichever is closer to the current coefficient. On the other hand, the adjusted coefficient will be or if the watermark bit is 0.

The embedding process is described as follows. First, the stage uses the current coefficient and the predefined to calculate S, sign, and according to (3), (4), and (5), respectively. In the equations, sign indicates that the is positive or negative; is the remainder for watermark bit value 0 and, similarly, for watermark bit value 1. Then, the stage determines the C_Low and C_High according to the value of the watermark bit. If the value is 0, (6) will be used, otherwise, (7). Finally, the stage adjusts the coefficient to according to (8). Consider

The steps of the watermark embedding stage are listed in Algorithm 2.

Input:   A color image H ( ) and a fingerprint image FP (N/4 × N/4).
Output:   A watermarked image ( ).
Apply DWT to the R and B channels of H to obtain the four embedding blocks
For each coefficient c of the blocks and the corresponding pixel of FP
    Compute the sign, , and ( , ) by (3), (4), and (5), respectively
    Compute the C_Low and C_High by (6) if watermark bit is 0, or (7) if otherwise
    Adjust the coefficient to by (8)
End For
Perform inverse DWT to produce the watermarked image

3.2. Authentication Logo Detection Phase

The phase is activated when a dispute over the copyright of a suspect image occurs. The following details the stages in the phase, including watermark retrieval and logo restoration.

3.2.1. Watermark Retrieval

The stage regains the watermark from the suspect image. First, the stage applies the DWT to the R and B planes of the suspect image in the same way as that in watermark embedding to obtain the embedding blocks, to (refer to Figure 6). Then, for each coefficient of the blocks, it obtains the watermark bit according to where is the same as that in (3).

Finally, the watermark (which is supposed to be the fingerprint of the image) is restored by assembling every for each coefficient in the blocks.

Algorithm 3 lists the steps of the watermark retrieval stage.

Input:   A suspect color image S ( ).
Output:   A fingerprint (watermarked) image FP′ ( ).
Apply DWT to the R and B channels of and obtain the embedding blocks
For each coefficient of the blocks
    Obtain w according to (9)
End For
Assemble every watermark bit w to restore the fingerprint image FP′

3.2.2. Logo Restoration

The stage restores the authentication logo. As shown in Figure 8, it has four substages: scrambled logo recovery, logo unscrambling, logo enhancement, and logo resizing.

Scrambled Logo Recovery. After retrieving the fingerprint image (i.e., watermark) by watermark retrieval and extracting the feature image by feature extraction, the substage performs XOR operation on each pixel of the fingerprint image and the corresponding pixel of the feature image to retrieve the scrambled logo. Because both of the two images are black-and-white, each pixel is either 0 (black) or 1 (white). Therefore, the substage simply performs bitwise XOR operation on the two images and obtains a scrambled logo.

Logo Unscrambling. As mentioned in Section 3.1.2, the logo was scrambled before it is used to generate fingerprint in the fingerprint and watermarked image generation phase. The substage adopts (10), which is the inverse equation of (2), to rearrange the scrambled logo and restore the logo. Consider

Logo Enhancement. The substage enhances the restored logo by erosion and dilation. Erosion removes pixels on object boundaries in an image and therefore can be used to remove smaller islands in the image; dilation, on the other hand, adds pixels to the boundaries of objects in an image and hence can be used to remove bright areas from the image. The substage performs erosion followed by dilation once, which is illustrated in Figure 9.

Erosion is performed by resetting each of the pixels according to (11). That is, if one of the neighbors is black, then the current pixel will be set to black; otherwise, it will be set to white. Let L(x, y) be the current pixel value to be determined and , , and its neighbors; then

Dilation, on the contrary, resets the neighbors of each pixel rather than the pixel itself. If L(x, y) is black; the substage sets all of its neighbors, , , and , to black; otherwise, the neighbors remain unchanged.

Logo Resizing. The proposed scheme adopts ISS to retrieve the authentication logo, which causes pixel expansion because one logo pixel is mapped to a share of four pixels (mentioned in Section 3.1.3). As a result, the retrieved logo from the previous stage will be larger than the original one. To resize the logo to its original size, the substage partitions the enhanced logo into several blocks of size 2 × 2, each of which is then reduced into one pixel with value according to the following rule:

4. Experimental Results

Two kinds of experiments were conducted to prove the effectiveness of the proposed scheme. The first experiment shows the robustness of our scheme and the other demonstrates the capability of unique identification. In the experiments, the authentication logo used to generate the watermark (fingerprint) is shown in Figure 10.

Two common measurements used to estimate the robustness of our scheme are described as below.

(1) Peak Signal to Noise Ratio (PSNR). The measurement to estimate the color image quality after image processing is a variant version of normal PSNR [29]. The variant PSNR listed below does not consider the influence of the green channel because the channel is not modified by our scheme. Consider where MSE is the mean square error between the original image and the modified image, which is defined as follows: where represents the original pixel value and denotes the modified pixel value.

According to the definition of PSNR, the higher the value is, the better the quality of the modified image is. Generally, if the PSNR is greater than 30 dB, the quality of the modified image is acceptable.

(2) Accuracy Rate (AR). The measurement shown below is used to evaluate the correctness of the logo after it has been restored. Consider where NP is the number of pixels in the original logo and CP is the number of correct pixels obtained by comparing the pixels of the original logo with the corresponding ones of the restored logo. Figure 11 shows the restored logos with different AR values. As can been seen, restored logos with AR higher than 81% (the ones in the upper row) still can be visually recognized whereas the restored logo with AR equal to 76% is hard to identify. However, according to the description in Section 4.2, if AR is higher than 75%, the scheme still can identify the copyrights of the image.

4.1. Robustness Experiments

The experiments proved our scheme is robust to different kinds of attacks. The test images used in the experiment, including “Lena,” “Mandrill,” “Sailboat,” and “Peppers,” are shown in Figure 12. The commercial image processing software “Adobe Image Photoshop CS” was used to simulate several kinds of image attacks, some of which for “Lena” are shown in Figure 13.

The experimental results of the test images are shown in Table 3. As mentioned above, if AR is less than 75%, the scheme cannot identify the copyrights of the image. Table 3 shows that our scheme failed to verify the copyrights for the images suffering the heavier attacks of JPEG, contrast, Gauss blurring, and scaling in watermark verification. Nevertheless, the duplications of those images can all be determined in fingerprint verification. In summary, our scheme can identify the copyrights of the suspect images under moderate attacks in watermark verification and determine the duplications of those suffering the heavy attacks in fingerprint verification.


AttacksAR (%) (Watermarking/fingerprinting)
LenaMandrillSailboatPeppersAverage

None99.93/99.9399.76/99.7699.98/99.9899.63/99.6399.83/99.83

JPEG(90)88.31/99.7186.84/99.4488.04/99.8888.16/99.4987.84/99.63
(80)79.76/99.7376.17/99.0277.29/99.5877.91/99.1277.78/99.36
(70)74.58/99.4672.12/98.7372.61/99.5473.05/99.1773.09/99.23
(60)71.58/99.6370.14/98.170.19/99.4469.95/98.8370.47/99.00
(50)71.48/99.4170.09/97.6870.43/99.2469.8/98.8870.45/98.80

Contrast(10)92.6/99.8580.71/99.7888.62/99.9590.31/99.2788.06/99.71
(20)82.86/99.7673.12/99.5679.27/99.7881.81/98.6879.27/99.45
(30)75.85/99.4670.61/99.0775.22/99.6374.63/98.0774.08/99.06
(40)72.61/99.2770.83/97.9272.31/99.1972.51/98.1472.07/98.63
(50)71.7/98.9370.41/96.1971.19/98.7571.09/96.5871.10/97.61
(60)70.39/97.6870.65/92.770.95/98.170.53/95.870.63/96.07
(70)70.8/94.6570.65/90.3171.12/97.7171.36/9570.98/94.42

Gauss blurring(0.5)83.64/99.4173.44/97.2978.17/99.1780.54/99.0578.95/98.73
(1.0)76.1/98.8870.92/95.7874.17/98.4473.88/98.3473.77/97.86
(1.5)73.63/98.1770.95/94.3472.44/97.2972.34/97.7372.34/96.88

Scalingw: 256, h: 25671.66/97.9270.75/92.5371.22/96.8370.73/95.871.09/95.77
w: 128, h: 12872.83/96.0273.41/86.0473.49/94.0473.41/92.9973.29/92.27

Brightness(10)96.66/96.7894.41/94.9297.85/97.8594.46/94.4695.85/96.00
(20)90.65/91.4889.06/90.795.9/95.988.33/88.3390.99/91.60
(30)85.72/87.6283.98/86.4393.55/93.5883.69/83.6486.74/87.82

Croppingw: 480, h: 48089.62/92.9288.99/92.4691.48/95.0488.77/91.8789.72/93.07
w: 448, h: 44882.71/86.8281.71/85.6783.79/88.8982.03/85.8282.56/86.80
w: 432, h: 43279.88/83.0879.25/83.0181.13/86.2379/83.379.82/83.91
w: 424, h: 43278.88/81.7678.49/81.9680.52/85.8678.59/83.0679.12/83.16
w: 424, h: 42478.22/81.278/81.4980.25/85.3378.1/82.4278.64/82.61

Moreover, there were 100 attacks in total and 53 of them resulted in an AR value higher than 75% in watermarking verification. That is to say, the copyrights of the 53% of the attacked images can be successfully identified in watermark verification, and hence only 47% of them need fingerprint verification.

4.2. Uniqueness Experiments

The experiment showed that our scheme has the capability of unique identification and is able to distinguish a copyrighted image from different ones. The four copyrighted images (with embedded watermarks) in Figure 12 along with seven unwatermarked images (Figures 14(a)14(g)) were processed by our scheme to identify the copyright. The stored fingerprints of the watermarked images were used in fingerprint verification to restore the logos for all of the images.

The results shown in Table 4 demonstrated the extraordinary unique identification capability of our scheme. It can be clearly seen that all the restored logos except the ones of the copyrighted images (those on the rightmost side) are unrecognizable, which proves that our scheme is actually able to distinguish a copyrighted image from different ones. It will not mistakenly identify the copyrights of an unwatermarked image.

(a) Lena

MandrillSailboatPeppersF-16OperaWatchNew YorkTiffanyTerrauxFourviereLena

Watermark454867.table.0092454867.table.0093454867.table.0094454867.table.0095454867.table.0096454867.table.0097454867.table.0098454867.table.0099454867.table.00100454867.table.00101454867.table.00102

AR (%)73.0773.6973.0769.8070.3970.0971.3470.2669.7870.2699.93

Fingerprint454867.table.00103454867.table.00104454867.table.00105454867.table.00106454867.table.00107454867.table.00108454867.table.00109454867.table.00110454867.table.00111454867.table.00112454867.table.00113

AR (%)59.0656.0349.2462.7259.6244.3462.3560.4045.6563.2199.93

(b) Mandrill

LenaSailboatPeppersF-16OperaWatchNew YorkTiffanyTerrauxFourviereMandrill

Watermark454867.table.00114454867.table.00115454867.table.00116454867.table.00117454867.table.00118454867.table.00119454867.table.00120454867.table.00121454867.table.00122454867.table.00123454867.table.00124

AR (%)73.7473.6973.0772.7373.5273.7972.6474.3573.8272.299.76

Fingerprint454867.table.00125454867.table.00126454867.table.00127454867.table.00128454867.table.00129454867.table.00130454867.table.00131454867.table.00132454867.table.00133454867.table.00134454867.table.00135

AR (%)58.257.8162.0857.8459.4560.0359.6454.3559.0160.3899.76

(c) Sailboat

LenaMandrillPeppersF-16OperaWatchNew YorkTiffanyTerrauxFourviereSailboat

Watermark454867.table.00136454867.table.00137454867.table.00138454867.table.00139454867.table.00140454867.table.00141454867.table.00142454867.table.00143454867.table.00144454867.table.00145454867.table.00146

AR (%)73.7473.0773.0772.7373.5273.7972.6474.3573.8272.299.98

Fingerprint454867.table.00147454867.table.00148454867.table.00149454867.table.00150454867.table.00151454867.table.00152454867.table.00153454867.table.00154454867.table.00155454867.table.00156454867.table.00157

AR (%)54.3259.0653.4457.9155.152.3756.4755.3753.7859.499.98

(d) Peppers

LenaMandrillSailboatF-16OperaWatchNew YorkTiffanyTerrauxFourvierePeppers

Watermark454867.table.00158454867.table.00159454867.table.00160454867.table.00161454867.table.00162454867.table.00163454867.table.00164454867.table.00165454867.table.00166454867.table.00167454867.table.00168

AR (%)73.7473.0773.6972.7373.5273.7972.6474.3573.8272.299.63

Fingerprint454867.table.00169454867.table.00170454867.table.00171454867.table.00172454867.table.00173454867.table.00174454867.table.00175454867.table.00176454867.table.00177454867.table.00178454867.table.00179

AR (%)47.0260.1346.0855.5958.6949.2953.2652.854.6655.7799.63

Moreover, the resulting AR values of all the unwatermarked images are all lower than 75%. Hence, it is reasonable for our scheme to confirm the copyright when AR is higher than 75%.

4.3. Discussion and Comparison

The fingerprint extracted from an image is more robust than the watermark embedded in the image. This can be seen from Figure 15, which shows the average AR values of the images suffering different attacks for watermarking and fingerprinting. The AR values for fingerprinting are all higher than those for watermarking. When the image undergoes image processing operations that heavily damage the embedded watermark, the extracted fingerprint still can survive the attacks. However, because linear comparison is a computationally intensive process, fingerprinting is costly in time if there are many fingerprints in the database.

Table 5 shows the processing time for watermark extraction and logo restoration and comparison in our scheme. The experiments were carried out on a computer equipped with the following hardware and software:CPU: 3.16 GHz Intel(R) Xeon(R) CPU E3120,RAM: 4 GB,OS: Windows 7,Computing language and environment: MATLAB.


Stage LenaMandrillSailboatPeppersAverage

Watermark extraction218209200213210

Logo restoration and comparison2029392829

Watermark extraction + logo restoration and comparison238238239241239

As Table 5 shows, it takes about 0.21 seconds to extract the watermark from the input image and 0.029 seconds to restore and compare the logo. Therefore, the total processing time (in seconds) of our scheme can be expressed by where represents the number of the fingerprints retrieved from the database. If the copyright can be identified in watermark verification, is 0. That is, our scheme only needs 0.24 seconds in the best case. Otherwise, our scheme needs additional 0.029 seconds for each retrieved fingerprint in the database.

Because our scheme combines watermarking and fingerprinting techniques, it can be as efficient as pure watermarking schemes and also as robust as pure watermarking schemes. A pure watermarking scheme is very efficient because it only needs to make one watermark comparison to verify the copyright. However, it may not be as robust as a pure fingerprinting scheme when dealing with images that have suffered heavy attacks. Table 6 shows the comparison of our combined scheme and the other two pure schemes. If the copyright of the input image can be identified by watermarking verification, our scheme can be as efficient as pure watermarking schemes. If watermarking verification fails to identify the copyright, our scheme is still able to determine duplication in fingerprinting verification, which makes our scheme as robust as pure fingerprinting schemes.


Pure
watermarking scheme
Pure
fingerprinting scheme
Our
combined scheme

Robustness
Efficiency

: good; △: fair.

5. Conclusion

This paper presented a copyright identification scheme that takes advantage of the complementary nature of digital watermarking and fingerprinting. The experimental results showed that when the watermarked image suffers moderate attacks, watermarking verification alone is enough to identify the copyright, and there is no need for fingerprinting verification. In other words, the proposed scheme can identify the copyright efficiently in this situation. On the other hand, the experimental results also showed that when the watermarked image suffers heavy attacks that render watermarking verification incompetent, fingerprinting verification, although more time consuming, can successfully determine the duplication, hence demonstrating the robustness of the proposed scheme.

One distinguishing characteristic of the proposed scheme is that it does not need a separate watermark for watermarking verification and a separate fingerprint for fingerprinting verification. The proposed scheme extracts features from the input image to generate the fingerprint, which also serves as the watermark. Hence, only one piece of information is needed for both watermarking and fingerprinting verifications.

To further improve the scheme, retrieving the stored fingerprint image from the database to restore the correct authentication logo more quickly is worth studying. When there are more than one fingerprint image in the database, the original fingerprint of the host image must be correctly retrieved; otherwise, the correct authentication logo cannot be restored. Retrieving every stored fingerprint image to restore an authentication logo for comparison is very time consuming. The scheme should provide a more efficient way that is able to find the proper one in less time, which is the future work of the research.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgment

Financial support of this study by Tatung University, Taipei, Taiwan, under Grant B100-I07-036 is gratefully acknowledged.

References

  1. G. Coatrieux, H. Huang, H. Shu, and L. Luo, “A watermarking-based medical image integrity control system and an image moment signature for tampering characterization,” IEEE Journal of Biomedical and Health Informatics, vol. 17, no. 6, pp. 1057–1067, 2013. View at: Google Scholar
  2. K.-C. Liu, “Colour image watermarking for tamper proofing and pattern-based recovery,” IET Image Processing, vol. 6, no. 5, pp. 445–454, 2012. View at: Publisher Site | Google Scholar
  3. Y.-L. Chen and C.-T. Hsu, “Detecting recompression of JPEG images via periodicity analysis of compression artifacts for tampering detection,” IEEE Transactions on Information Forensics and Security, vol. 6, no. 2, pp. 396–406, 2011. View at: Publisher Site | Google Scholar
  4. P. -C. Su, Y. -C. Chang, and C. -Y. Wu, “Geometrically resilient digital image watermarking by using interest point extraction and extended pilot signals,” IEEE Transactions on Information Forensics and Security, vol. 8, no. 12, pp. 1897–1908, 2013. View at: Google Scholar
  5. A. Piper and R. Safavi-Naini, “Scalable fragile watermarking for image authentication,” IET Information Security, vol. 7, no. 4, pp. 300–311, 2013. View at: Google Scholar
  6. M. Pawlak and Y. Xin, “Robust image watermarking: an invariant domain approach,” in Proceedings of the IEEE Canadian Conference on Electrical and Computer Engineering, vol. 2, pp. 885–888, May 2002. View at: Google Scholar
  7. W.-H. Lin, S.-J. Horng, T.-W. Kao, P. Fan, C.-L. Lee, and Y. Pan, “An efficient watermarking method based on significant difference of wavelet coefficient quantization,” IEEE Transactions on Multimedia, vol. 10, no. 5, pp. 746–757, 2008. View at: Publisher Site | Google Scholar
  8. I.-K. Kong and C.-M. Pun, “Digital image watermarking with blind detection for copyright verification,” in Proceedings of the 1st International Congress on Image and Signal Processing (CISP '08), vol. 1, pp. 504–508, May 2008. View at: Publisher Site | Google Scholar
  9. L. A. Elrefaey, M. E. Allam, H. A. Kader, and M. Selim, “Robust blind image-adaptive watermarking,” in Proceedings of the 25th National Radio Science Conference (NRSC '08), March 2008. View at: Publisher Site | Google Scholar
  10. M.-H. Lee and O.-J. Kwon, “Color image watermarking based on DS-CDMA using Hadamard kernel,” in Proceedings of the 10th International Conference on Advanced Communication Technology, pp. 1592–1597, February 2008. View at: Publisher Site | Google Scholar
  11. N.-Y. Lee and C.-C. Wang, “Yet another wavelet watermarking scheme for copyright protection,” in Proceedings of the 9th IEEE International Conference on E-Commerce Technology, pp. 421–424, July 2007. View at: Publisher Site | Google Scholar
  12. H.-M. Sun, C.-J. Hong, and C.-H. Chen, “A new approach to feature-based copyright protection of images,” in Proceedings of the 3rd International Conference on Information Technology: Research and Education (ITRE '05), pp. 233–237, June 2005. View at: Publisher Site | Google Scholar
  13. C.-C. Chang and J.-C. Chuang, “An image intellectual property protection scheme for gray-level images using visual secret sharing strategy,” Pattern Recognition Letters, vol. 23, no. 8, pp. 931–941, 2002. View at: Publisher Site | Google Scholar
  14. C.-S. Lu, C.-Y. Hsut, S.-W. Sun, and P.-C. Chang, “Robust mesh-based hashing for copy detection and tracing of images,” in Proceedings of the IEEE International Conference on Multimedia and Expo (ICME '04), vol. 1, pp. 731–734, June 2004. View at: Google Scholar
  15. J.-S. Lee and K.-S. Yoon, “The system integration of DRM and fingerprinting,” in Proceedings of the 8th International Conference Advanced Communication Technology (ICACT '06), vol. 3, pp. 2180–2183, February 2006. View at: Google Scholar
  16. S.-H. Yang and C.-F. Chen, “Robust image hashing based on SPIHT,” in Proceedings of the 3rd International Conference on Information Technology: Research and Education (ITRE '05), pp. 110–114, June 2005. View at: Publisher Site | Google Scholar
  17. F. Ahmed and M. Y. Siyal, “A secure and robust DCT-based hashing scheme for image authentication,” in Proceedings of the 10th IEEE Singapore International Conference on Communication Systems (ICCS '06), pp. 1–6, October 2006. View at: Google Scholar
  18. J. S. Seo, J. Haitsma, T. Kalker, and C. D. Yoo, “A robust image fingerprinting system using the Radon transform,” Signal Processing: Image Communication, vol. 19, no. 4, pp. 325–339, 2004. View at: Publisher Site | Google Scholar
  19. P. C. Vila, Content-based audio search: from fingerprinting to semantic audio retrieval [Ph.D. thesis], University Pompeu Fabra, Barcelona, Spain, 2007.
  20. S.-L. Hsieh, I.-J. Tsai, B.-Y. Huang, and J.-J. Jian, “Protecting copyrights of color images using a watermarking scheme based on secret sharing and wavelet transform,” Journal of Multimedia, vol. 3, no. 4, pp. 42–49, 2008. View at: Google Scholar
  21. V. Nabiyev and A. Günay, “Towards a biometric purpose image filter according to skin detection,” in Proceedings of the Second International Conference on Problems of Cybernetics and Informatics, 2008. View at: Google Scholar
  22. C. Poynton, “Frequently asked questions about color,” 2014, http://www.poynton.com/ColorFAQ.html. View at: Google Scholar
  23. G. Voyatzis and I. Pitas, “Applications of torus automorphisms in image watermarking,” in Proceedings of International Conference on Image Processing, vol. 3, pp. 237–240, 1996. View at: Google Scholar
  24. G. Voyatzis and I. Pitas, “Chaotic mixing of digital images and applications to watermarking,” in Proceeding of European Conference on Multimedia Applications, Services and Techniques (ECMAST '96), vol. 2, May 1996. View at: Google Scholar
  25. R. C. Gonzalez and R. E. Woods, Digital Image Processing, Prentice Hall, New York, NY, USA, 2nd edition, 2002.
  26. E. J. Stollnitz, T. D. DeRose, and D. H. Salestin, “Wavelets for computer graphics: a primer, part 1,” IEEE Computer Graphics and Applications, vol. 15, no. 3, pp. 76–84, 1995. View at: Publisher Site | Google Scholar
  27. KeyLawk, “Colors: the human eye is most sensitive to GREEN,” 2014, http://keylawk.blogspot.com/2007/10/colors-human-eye-is-most-sensitive-to.html. View at: Google Scholar
  28. S. Zhao, “Wavelength of maximum human visual sensitivity,” 2014, http://hypertextbook.com/facts/2007/SusanZhao.shtml. View at: Google Scholar
  29. I. Nasir, W. Ying, and J. Jianmin, “Novel multiple spatial watermarking technique in color images,” in International Conference on Information Technology: New Generations (ITNG '08), pp. 777–782, April 2008. View at: Publisher Site | Google Scholar

Copyright © 2014 Shang-Lin Hsieh et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views2314
Downloads751
Citations

Related articles

We are committed to sharing findings related to COVID-19 as quickly as possible. We will be providing unlimited waivers of publication charges for accepted research articles as well as case reports and case series related to COVID-19. Review articles are excluded from this waiver policy. Sign up here as a reviewer to help fast-track new submissions.