Abstract

Due to the powerful image editing tools images are open to several manipulations; therefore, their authenticity is becoming questionable especially when images have influential power, for example, in a court of law, news reports, and insurance claims. Image forensic techniques determine the integrity of images by applying various high-tech mechanisms developed in the literature. In this paper, the images are analyzed for a particular type of forgery where a region of an image is copied and pasted onto the same image to create a duplication or to conceal some existing objects. To detect the copy-move forgery attack, images are first divided into overlapping square blocks and DCT components are adopted as the block representations. Due to the high dimensional nature of the feature space, Gaussian RBF kernel PCA is applied to achieve the reduced dimensional feature vector representation that also improved the efficiency during the feature matching. Extensive experiments are performed to evaluate the proposed method in comparison to state of the art. The experimental results reveal that the proposed technique precisely determines the copy-move forgery even when the images are contaminated with blurring, noise, and compression and can effectively detect multiple copy-move forgeries. Hence, the proposed technique provides a computationally efficient and reliable way of copy-move forgery detection that increases the credibility of images in evidence centered applications.

1. Introduction

With the advancements in imaging technologies, the digital images are becoming a concrete information source. Meanwhile, a large variety of image editing tools have placed the authenticity of images at risk. The ambition behind the image content forgery is to perform the manipulations in a way, making them hard to reveal through the naked eye, and use these creations for malicious purposes. For instance, in 2001, after the 9/11 incident, several videos of Osama bin Laden over the social media were found counterfeited through the forensic analysis [1]. In the same way, in 2007, an image of tiger in forest forced the people to believe in the existence of tigers in the Shanxi province of China. The forensic analysis, however, proved the tiger to be a “paper tiger” [2]. Similarly, in 2008, an official image of four Iranian ballistic missiles was found to be doctored, as one missile was revealed to be duplicated [3]. Hence, the famous saying “seeing is believing” [4, 5] is no longer effective. Therefore, ways that can ensure the integrity of the images especially in the evidence centered applications are required.

In recent years, an exciting field, digital image forensics, has emerged which finds the evidence of forgeries in digital images [6]. The primary focus of the digital image forensics is to investigate the images for the presence of forgery by applying either the active or the passive (blind) techniques [2]. The active techniques such as watermarking [7] and digital signatures [6] depend on the information embedded a priori in the images. However, the unavailability of the information may limit the application of active techniques in practice [8]. Thus, passive techniques are used to authenticate the images that do not require any prior information about them [810].

Images are usually manipulated in two ways such as image splicing and region duplication through copy-move forgery. In image splicing, regions from multiple images are used to create a forged image. However, in copy-move forgery, image regions are copied and pasted onto the same image to conceal or increase some important content in the pictured image. As copied regions are apparently identical with compatible components (i.e., color and noise), it becomes a challenging task to differentiate the tempered regions from authentic regions. Furthermore, a counterfeiter applies various postprocessing operations such as blurring, edge smoothing, and noise to remove the visual traces of image forgeries. An example of copy-move forgery is shown in Figure 1.

In the present work copy-move forgery detection is addressed through the discrete cosine transform (DCT) and Gaussian RBF kernel PCA that are used to investigate the similarity between duplicated regions. The benefits of our algorithm compared against several existing CMFD methods are(i)utilization of the lower length of feature vectors;(ii)lower computational cost;(iii)robustness against various postprocessing operations over the forged regions;(iv)ability to detect multiple copy-move forgeries.

The rest of the paper is organized as follows: Section 2 presents the related work regarding copy-move forgery detection (CMFD). Section 3 presents the details of proposed method. Experimental results are presented in Section 4. Finally, the conclusions are drawn in Section 5.

Various CMFD techniques have been proposed so far to effectively address the region duplication problem. In this regard, the research is intended towards the representation of image regions in a more powerful way to accurately detect the duplicated regions. In [11], Fridrich et al. for the first time presented the copy-move forgery detection technique using DCT on small overlapping blocks. The feature vectors are formed using DCT coefficients. The similarity between blocks is analyzed after sorting the feature vectors lexicographically. In [13], image blocks are represented through principal component analysis (PCA). Exploiting one of the features of PCA, the authors used about half of the number of features utilized by [11]. It makes this technique effective but failed to detect copy-move forgery with rotation. In [15], a sorted neighborhood technique based on Discreet Wavelet Transform (DWT) is proposed. The image is decomposed into four subbands and applied the Singular Value Decomposition (SVD) on low frequency components for getting the feature vector. The technique is robust to JPEG compression up to the quality level 70 only. In [16], a technique based on blur moment invariants up to seventh order for extracting the block features and kd-tree matching is introduced. In [12], the application of scaling and rotation invariant Fourier-Mellin Transform (FMT) is suggested in combination with bloom filters on the image blocks for detecting the image forgery. In [14], an improved DCT-based technique is proposed by introducing a truncating process to reduce the dimension of feature vector for forgery detection. In [17], a solution through DCT and SVD is proposed for detecting image forgeries. The algorithm is shown to be robust against compression, noise, and blurring but fails when images are even slightly rotated. In [18], an efficient expanding block technique based on direct block comparison is proposed. In [19], circle block extraction is performed and the features are obtained through rotation invariant uniform local binary patterns (LBP). The technique is robust to blurring, additive noise, compression, flipping, and rotation. However, this technique failed to detect forged regions rotated with arbitrary angles. In [20], the authors employed a new powerful set of keypoint-based features called MIFT for finding similar regions in the images. In [21], the authors extracted feature vectors from circular blocks using polar harmonic transform (PHT) for detecting image forgeries. In [22], an adaptive similarity threshold based scheme is presented in the block matching stage. The detection of forged regions is determined using thresholds proportional to blocks standard deviations. In [23], a method using the Histogram of Oriented Gradients (HOG) is suggested to detect the copy-move forged regions. In [24], the multiscale Weber’s law descriptor (multi-WLD) and multiscale LBP features are extracted for image splicing and copy-move forgery detection from chrominance components. The authors employed SVM for classifying an image as authentic or forged.

3. Proposed Methodology

In this paper, copy-move forgery detection is performed through the DCT and Gaussian RBF kernel PCA using the squared blocks. The reason to use the DCT for block representation is the robustness against several postprocessing operations, for example, compression, blurring, scaling, and noise [25], as it is a common practice in image forgery that the counterfeited images always undergo various postprocessing operations. Hence, it makes the forgery detection very difficult. Although the DCT is effective against mentioned transformations, still there are situations where the block representations through DCT will be nominal; for example, if rotation operation is applied over the forged regions, the DCT representations results are affected as well. To overcome this limitation we apply Gaussian RBF kernel PCA over the DCT frequency coefficients due to their rotation invariant nature compared against PCA [25]. Another motivation to use kernel PCA with DCT is the nonlinear nature of RBF kernel PCA and linear nature of DCT. Hence, it makes the feature representation more diverse and also appears as a better choice compared to PCA that is also linear in nature like DCT. Gaussian RBF kernels have some other advantages such as having fewer hyperparameters; hence, they are numerically less difficult as kernel values are bounded between 0 and 1.

3.1. Framework of the Proposed Algorithm

The discussion above draws forth the framework of CMFD that is described in Figure 2. The steps of the proposed CMFD technique are given as follows:(1)Dividing the grayscale image into fixed sized overlapping blocks.(2)Applying DCT to each extracted block.(3)Extracting Gaussian RBF kernel PCA-based features from each DCT square block.(4)Matching similar block pairs.(5)Removing the isolated block and output the duplicated regions.

3.2. Preprocessing and Blocks Extraction

For the implementation of proposed method, the algorithm is applied over the grayscale images. Thus, as a first step, a color input image of size is converted to a grayscale image using where , , and are the red, green, and blue components of image , respectively.

Once image is converted into a grayscale image, a window of size is slided one pixel along from the top left corner to the bottom lower right corner for obtaining the overlapping blocks. Each block is represented as , where and are the starting points of the block’s row and column, respectively, as shown inwhere , , and .

Thus, can be divided into overlapping blocks as shown in

3.3. Feature Extraction

For an image block of size , where , are , we decompose the block in terms of basis function. The result occurs in the form of a coefficients matrix of size that contains the DCT coefficients: where

The coefficients matrix can be ordered to a zig-zag pattern to reflect the amount of information stored for block representation. For a tempered block that is a duplication of block and does not undergo a postprocessing operation, the zig-zag coefficients are similar, and duplicated regions perfectly match but if geometric transformation such as rotation is applied on duplicated regions as where and represent the block coordinates after rotation with a rotation angle , the DCT coefficients may not match; that is, where is the distance between regions in metric norm. For region matching the value of must be less than a threshold . Therefore, to overcome this limitation of DCT coefficients, we compute the eigenvalues and eigenvectors satisfying with , where is a nonlinear function. Thus, we can substitute in the eigenvector equation. Therefore, all solutions must lie in the span of . Equivalently, So there exist coefficients such thatSubstituting and (9) in (8) and defining matrix by give where represents the column vector and is the symmetric Gram matrix [26] as given by

If represents the eigenvalues of and and denotes the eigenvectors, whereas denotes the normalized eigenvectors provided for , with an assumption that only the first eigenvalues are nonzero and positive. Therefore, for any test point , the th nonlinear principal components are given by [26]

In our implementation, a Gaussian RBF kernel function is chosen, which is defined by the mapping function such thatwhere and represents the Euclidean distance. For any input space , a kernel is a positive definite function for any integer , satisfying , for any [26] and any . Hence, the Gaussian RBF kernel is given as where is the Gaussian kernel parameter. By applying (14) on (4), we get a transformed representation as

The dimensionality of matrix can be reduced to by using

3.4. Block Representation Details

In our implementation, the size of each block is , where the value of and is . When the DCT is applied to block , we get a coefficients matrix of the same size that is further transformed through the Gaussian RBF kernel PCA. By applying (16), we obtain first most informative principal components that serve as the feature vector for the block representation. The reason to select these principal components [27] is that we want to reduce the feature vector size due to large number of blocks and the curse of dimensionality due to which the time requirements for computation increase. Another reason is that the principal component analysis is orthogonal linear transformation such that the greatest variance by some projection appears as the first principal component, the second greatest variance appears as second principal component, and so on. In our implementation, the 10 most informative principal components are selected as the elements of feature vector for a block. Hence, a matrix having the feature vectors of all the blocks is produced as given in

The matrix is sorted lexicographically which makes identical features closer to each other. In the meantime record the left corner’s coordinate of each block that is represented by a square block. Lexicographical sorting before the feature vector matching procedure helps in decreasing the computational cost because a vector is compared against neighboring feature vectors to judge the similarity. In our implementation, window size is to effectively handle various postprocessing operations such as blurring, noise, and compression.

Table 1 gives the comparison between the proposed method and other methods in terms of feature vector dimensions. In comparison with other methods our technique uses the lower dimension of feature vector and hence is computationally more efficient.

3.5. Forgery Detection

The target of the CMFD is to find duplicated regions where the similarity index between regions (feature vectors) is less than a certain threshold and the duplicated regions are nonoverlapping. The reason for the threshold based region matching is due to the nature of counterfeited images that undergo postprocessing operations; and the probability of being similar in terms of features is almost negligible. Therefore, for CMFD, two conditions are imposed over the duplicated block detection procedure: (1) the blocks are nonintersecting and nonoverlapping, and (2) the similarity index does not exceed a threshold.

To meet the first requirement of the CMFD, that is, the matching between nonoverlapping blocks, the shift distance criterion is used. For this, let us consider that and are the top left corner coordinates of the two blocks that are represented by the features vectors and , and then

If the two feature vectors satisfy (18) then we consider these feature vectors for similarity index calculation to meet the second requirement of the CMFD. For this Euclidean distance is adopted as

To show the result of forgery detection, the algorithm produces a black map image and the regions that are considered to be duplicated are highlighted as the desired output .

3.6. Morphological Operations

To show the final detection result of the algorithm, morphological opening and closing operations at a scale defined by the size of the structural element are applied to without losing any details of interest. Thus, the opening operation with a structural element of size is used to remove the small and unwanted blocks of , while closing operation with a structural element of size is used to fill the holes in the highlighted regions of .

3.7. Computational Analysis

For the clear description of process, we are adopting some standard notations as found in the computation text [28]. Suppose that the size of each block is which was originally , where , and there are total blocks. As a first step, we compute the 2D discrete cosine transform for a single block that takes time. Once the DCT is computed for a single block, we apply Gaussian RBF kernel PCA that takes time [27]. Therefore, the machine instructions required for a single block is and as in time complexity we are interested only in the largest factor; therefore, the time required to compute features for a single block is . As the process of block representation is applied to , therefore the feature matrix is obtained in time. After this step, the lexicographical sorting is applied on that takes time. As is larger then or even , therefore the time required against the machine instructions of is . For comparison of blocks the algorithm takes the linear time that does not affect the time complexity of the algorithm. Hence, the time complexity of the algorithm remains , whereas the total machine instructions required are .

4. Experimental Results and Discussions

The experimental results of proposed technique are presented in this section. Adobe Photoshop is used to forge the images and all the experiments are performed on a plate form with Intel 1.70 GHz Core i5 processor and MATLAB 2011. The performance of the proposed technique is evaluated on two datasets. We used grayscale images with the size 128 × 128 pixels from the DVMM Columbia University dataset [29]. The second dataset is collected from the Internet, containing the images of sizes 256 × 256 and 512 × 512 pixels. In the experimentation, we set the parameter values for (current block) of size , (number of rows to compare) = 20, (block distance threshold) = 40, (number of principal components) and (similarity distance between vectors) = , respectively. The experimentation details are presented in the following sections.

4.1. Performance Evaluation

Practically, the most significant property of a detection technique is its capability to discriminate forged and authentic images. In addition to this, the power of locating the forged area correctly is also very important which gives a strong evidence to expose digital forgeries. Thus, the performance of our algorithm is evaluated at two levels: at image level, where we are concerned about the fact that the detected image is truly a forged image, and at pixel level, where we evaluate how accurately the forged areas can be located. To show the accuracy of the proposed technique at image level, the computation of precision “” indicates the probability that an identified forgery is indeed a forgery; and recall “” denotes the probability that actually a forged image is detected [25]:where represents the total number of correctly detected forged images, represents the total number of authentic images mistakenly detected as forged, and represents the total number of forged images incorrectly missed.

To show the accuracy at pixel level the true positive rate (TPR) and the false positive rate (FPR) are calculated as follows: where represents the pixels of original area, the pixels as the forged area, the pixels as the detected original area, and the pixels as the detected forged area. Hence, the TPR shows the performance of technique by correctly identifying the pixels of the copy-moved areas in the forged image, while FPR reflects the pixels which are not contained in forged region but mistakenly included by the implemented technique. Therefore, both the above parameters point out how accurately the proposed technique can locate duplicated areas. The more the TPR is close to 1 and FPR is close to 0, the more precise the technique would be.

4.2. Effectiveness Testing

In order to test the effectiveness of the proposed algorithm, the first experiment for detecting copy-move forgery is performed on the images where the forged region is translated to another location of the image. All the images used in this experiment are without any postprocessing operation. We selected grayscale images from dataset-I with the size of 128 pixels × 128 pixels. The detection results of the experiment can be seen from Figure 3, where left to right is the original, the forged, and the resultant image. As can be seen from Figure 3, there are large similar areas which are difficult to differentiate. However, the algorithm detected the forged regions efficiently. In the second experiment, we selected some color images from dataset-II with the dimensions 512 pixels × 512 pixels. The detection results are given in Figure 4, where left to right is the original, the forged, and the resultant image. We can see from Figure 4 that all the forged objects are irregular; however, the algorithm detected the forged objects precisely.

4.3. Robustness and Accuracy Test of the Algorithm

With the help of image editing software, a counterfeiter usually makes his best efforts to create a forged image. In real life, in order to achieve some purpose, a counterfeiter intentionally performs some postprocessing operations such as blurring, noise, and compression to create imperceptible forged images. Additionally, multiple copy-move forgeries are also a means of image tempering, where there are multiple duplicated areas. In this section, we take these into account and presented some experiments to show the robustness and accuracy of the algorithm. However, in [11, 13, 24, 30], such experiments are not given. The forgery detection results are shown in Figures 57. However, the test for detecting multiple copy-move forgeries is given in Figure 8.

Moreover, to evaluate the robustness and accuracy of the proposed technique quantitatively, 100 authentic images are selected from the two datasets for generation of forged images. To obtain four relatively different forged images from a selected authentic image, a square area of size pixels is copied from a random location and pasted onto a nonoverlapping area. Adopting the same approach, the squared area of size pixels is used to generate four more tempered images for the selected image. In this way the forged image dataset comprising 800 images is generated for the selected authentic images. These forged and original images are then contaminated with postprocessing operations such as Gaussian blurring, AWGN, and JPEG compression. The results are given in Tables 24, which evaluate the robustness and accuracy of the algorithm at image and pixel level.

The results given in Tables 24 show that the detection performance would be better when the duplicated region is larger. Table 2 indicates that the detection performance of the algorithm is high when the images are blurred by Gaussian blurring; even when the images have poor quality (, ) and small forged area ( pixels), our technique fails to detect only 14 out of 800 forged images (). Table 3 shows that the algorithm also performed well in the case of AWGN distorted images. From Table 3, we can draw a conclusion that the proposed algorithm is also capable of detecting forged regions precisely in the case of slightly compressed images with quality factor ( and ).

In the last experiment, the proposed technique is compared with other approaches: DCT-based [11], PCA-based [13], FMT-based [12], and improved DCT-based [14]. For this purpose, we selected 100 authentic images of size 512 × 512 pixels and generated 400 forged images. Here, to obtain two relatively different forged images, a square area of size 48 × 48 pixels is copied from a random location and pasted onto a nonoverlapping area. The overall average performance comparison of over 400 forged images blurred with Gaussian blurring, distorted with AWGN, and with JPEG compression level is shown in Figures 9, 10, and 11, respectively. In the case of Gaussian blurring, Figure 9 indicates the results, where the forged images are blurred by Gaussian filter ( = 5 and = 0.5, 1, 1.5, 2, 2.5, and 3). It is observed that decreasing the radius of Gaussian filter results in higher TPR but lower FPR for all methods. Moreover, TPR curve of the proposed technique achieves higher performance than other techniques, with TPR ≥ 85%, even when the radius of blurring is increased. The FPR curve also gives satisfactory performance that the proposed technique has lower FPR, even with larger blurring radius (). A similar behavior can be observed in the case of noise; the results are shown in Figure 10, where the forged images are distorted with AWGN (SNR = 20, 25, 30, 35, and 40). It is observed for all the techniques that increasing the SNR levels increases the TPR and decreases the FPR. It is also observed that the overall performance of PCA-based method is lower when SNR drops to about 20 dB. However, the proposed technique exhibited better performance by achieving higher TPR and lower FPR than other related techniques. Figure 11 is showing the results with different JPEG compression levels ( = 70, 75, 80, 85, and 90), which indicates that the proposed technique performed well when the forged images were slightly compressed.

5. Conclusion

In this paper, we focused on finding the ways through which we can assure the detection of copy-move forgery in digital images. The main consideration of this paper was to reduce the dimension of the feature length and find the forged objects in the suspected image. Therefore, we have applied DCT and kernel PCA for feature extraction which considers the identical objects found in the forged image. Furthermore, this technique does not require any prior information embedded into the image and works in the absence of digital signature or digital watermark. From the results, a conclusion can be drawn which is that the proposed technique not only effectively detects multiple copy-move forgeries and precisely locates the forged areas but also has nice robustness to postprocessing operations such as Gaussian blurring, AWGN, and compression. Moreover, comparing the detection performance of the proposed technique with existing standard copy-move forgery systems [1114], the results of our technique are reasonably good in terms of average TPR and FPR.

Competing Interests

The authors declare that there are no competing interests regarding the publication of this paper.

Acknowledgments

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) grant funded by the Ministry of Science, ICT & Future Planning (MSIP) (2013R1A1A2008180).