Research Article | Open Access
Bo Liu, Chi-Man Pun, Xiao-Chen Yuan, "Digital Image Forgery Detection Using JPEG Features and Local Noise Discrepancies", The Scientific World Journal, vol. 2014, Article ID 230425, 12 pages, 2014. https://doi.org/10.1155/2014/230425
Digital Image Forgery Detection Using JPEG Features and Local Noise Discrepancies
Wide availability of image processing software makes counterfeiting become an easy and low-cost way to distort or conceal facts. Driven by great needs for valid forensic technique, many methods have been proposed to expose such forgeries. In this paper, we proposed an integrated algorithm which was able to detect two commonly used fraud practices: copy-move and splicing forgery in digital picture. To achieve this target, a special descriptor for each block was created combining the feature from JPEG block artificial grid with that from noise estimation. And forehand image quality assessment procedure reconciled these different features by setting proper weights. Experimental results showed that, compared to existing algorithms, our proposed method is effective on detecting both copy-move and splicing forgery regardless of JPEG compression ratio of the input image.
Nowadays altering digital images via intuitive software is an operation of simpleness with very low cost; thus every individual can synthesize a fake picture. For the widely accessible Internet, the false information disseminates extremely fast. As a consequence, the facts may be distorted and the public opinion may be affected, yielding negative social influence. It can be even worse in the justice when pictures are presented as evidence. Therefore, there is a strong demand for valid and robust authentication method to discern whether a picture is original or not.
Two means are commonly utilized to make a forgery: copy-move and splicing. In the former case, a part of a picture is duplicated and then pasted onto other regions to cover any unwanted portion within the same picture . In the latter case, tampered image consists of two sources and retains the majority of one image for detail . Researchers and scientists have proposed many methods  to expose such intended manipulations. Passive forensic methods fulfill the task without additional information except for the image itself, thus showing advantages over active algorithms like watermarking and other signature schemas. Hence most research work is absorbed in developing blind authentication methods.
The forged picture leaves some clues which can be used to locate the manipulated regions. In the practice of copy-move operation, because the pasted area, though it may probably be altered geometrically, shares some similar features with the original region which is duplicated, searching for analogous features abstracted from local area is a possible solution. SIFT feature can be used to locate clone areas [3, 4]. For splicing tampered image detection, considering that there may be some discrepancies between the host image and the spliced region attempts to find the difference to expose that forgeries make sense. For instance, Kakar et al.  took advantage of motion blur discrepancy to detect fake pictures. For so many pictures stored or disseminated in JPEG format, some traces left by JPEG compression algorithm can be used. Estimating the quantization matrix used in JPEG compression, regions that possess inconsistent DCT coefficients are regarded as splicing area since an intact JPEG format picture should be coded by only one quantization table. Hamdy et al. developed this idea in . However, this approach fails to deal with double compression. Indeed, this method is effective only on detecting BMP format image which is composed of two JPEG pictures with different compression table. Although complex situation in double compression was discussed , multicompressions more than twice are still very hard to analyze.
Our goal is to automatically detect copy-move as well as splicing forgery within a single process without any prior knowledge about the forgery type of the doubtful picture. The reason is obvious. Rather than putting the same picture into different algorithm, which may be effective on only one certain type of forgery, one single method is time saving and avoids evaluating every detection result, which will be very hard to discern the true output from various results. Lin and Wu in  proposed an integrated method to detect both copy-move and splicing forgery. But this method just connects two separate processes together. The errors in the step of forgery type judgment will greatly affect the detection result. Actually it is unnecessary to classify the pictures into certain forgery type if there is a tool or feature sensitive to two attack practices. Li et al.  proposed a method based on JPEG block artificial grids (BAG) detection to expose both splicing and copy-move forgeries. But there are two major problems in that paper; the first is that the algorithm must be adjusted before applying to different forgery; the author revised the splicing detection algorithm to deal with copy-move practice. However in practice we do not know the forgery type for a doubtful image. The other defect is that the algorithm is sensitive to highly compressed image and ineffective on high quality picture with little compression. To overcome this shortcoming, we inducted noise feature to compensate for BAG algorithm. When the picture is less compressed, clear BAG is very difficult to extract. We consider that local noise level and category can be used as a feature to identify different source of regions in a picture. Inconsistency and discrepancy from regions to regions provide the other clues except for BAG to locate forged area. Therefore, we created an integrated feature combining BAG and noise feature to verify the authenticity of a doubtful picture.
The paper is organized as follows. Section 2 introduces the block artificial grids and noise patterns for detecting forgeries, and the following Section 3 details our proposed integrated method. In Section 4, experimental results will be presented to show the effectiveness of our method and we also compared it with existing methods. Finally we conclude this paper in Section 5.
2. Block Artificial Grid and Noise Estimation
2.1. Block Artificial Grid Extraction
It is universally known that the lossy JPEG compression will introduce some visually vertical or horizontal breaks in the image. These breaks called block artificial grid (BAG) appear at the border of each pixel block. This property can be used to determine whether a picture is altered or not. If the picture is intact, block artificial grids should only present on block borders, while there is a great possibility that copied and pasted or spliced regions will bring their original BAGs which may appear within the block rather than at borders. Some papers [9, 10] noticed this and Figure 1 illustrates the phenomenon. Theoretically speaking, if we extract all the BAGs from a given image, areas with BAGs within the block border are regarded as forged regions. Li et al.  introduced steps to extract BAGs. As it is mentioned before, artificial grids are visually vertical and horizontal lines, and they are very weak when comparing to the border lines of objects in the picture. And the main purpose of extraction procedures is to enhance these weak lines and to make them visible. However lines are also strengthened which may be the edges of objects or just objects themselves. This will interfere with the detection result because we only need BAGs. To allay the side effect, we preprocessed the doubtful image by excluding the edges of objects. But it should be noticed that BAG can also be regarded as vertical or horizontal edges. For preserving BAGs we only excluded edges within certain range.
Suppose that was the grayscale version of image and then the edge was obtained by , where represents Sobel operator and “” denotes convolution. Then we defined whether a pixel is excluded using where denotes gradient of the pixel and means excluded pixels. Then we begin to extract BAGs.
Firstly weak horizontal edges were extracted by calculating second-order difference of an image. For the test image, , absolute second-order difference was obtained by
Then all differentials larger than 0.1 or are discarded. In subsequence, enlarged horizontal lines are accumulated from every 33 columns, as shown in (3). Then a median filter is used to refine the result in
Weak horizontal edge is further periodical median filtered as
Similarly, the vertical BAGs can also be attracted. As a result, final BAG is obtained by adding two components together in
2.2. Noise Estimation
Highly compressed by JPEG, the picture shows visual block artificial grids across the whole frame which can be extracted by algorithm described in Section 2.1. However, under some circumstances, when the picture is not highly compressed and stored in high quality, the way by using BAG only becomes harder to detect forgery. To increase the versatility of the algorithm, we use noise feature. The noise comes from imaging sensor and internal circuits within a camera. And the number of noise changes in accordance with camera settings especially ISO sensitivity and exposure time. As an example, Figure 2 shows that the visual noise of images is captured from a Nikon D7000 camera. We can see that more noise appears in the image as the ISO speed rises. In Figure 3 we can see that different camera model from different manufacture also shows unequal noise amount and forms although the pictures were taken in the same scenery with equal ISO speed. So the noise can be used to help distinguish difference sources of a picture. When two pictures are spliced together the noise level or patterns are inconsistent between regions. By estimating the pattern or level of noise in different regions the forgery can be exposed via noise discrepancies.
In most cases, the alien region has a specific shape, such as a tree, a bird, or a person. The forged object may possess different noise level compared to that of its surroundings. To estimate every region’s noise level, the image should be firstly divided into small segments. Most previous methods divide the picture into small overlapping blocks with equal size. But, in our application, this will lead to bad performance in next steps which need accurate noise estimation of each region to compare noise discrepancy. This is because the forged area is not rectangle in most cases, and the small block will contain original and alien pixels. Therefore we segment picture into sets of pixels, not regular shaped, also known as superpixels. Employing this approach makes segments more meaningful and easier to be processed in the following steps because the segmentation algorithms locate the objects and boundaries other than the same-size blocks. The result of image segmentation is a set of segments that collectively cover the entire image or a set of contours extracted from the image. Each of the pixels in a region is similar with respect to some characteristic or computed property, such as color, intensity, or texture. Adjacent regions are significantly different with respect to the same characteristic(s) .
In our application, SLIC (simple linear iterative clustering) superpixels algorithm  was used to segment picture. This algorithm is easy but better than other segmentation methods. Given an image where denotes different color channel. The meaning of subscript will be explained later and
In essence, SLIC is a clustering algorithm. Similar to other clustering methods, two steps are evolved with SLIC segmentation. In the initialization step, cluster centers are assigned by sampling pixels at regular grid. Note that the picture is segmented in LAB color space. Then cluster centers are moved to the lowest gradient position in a neighborhood. In the assignment step, each pixel is associated with the nearest cluster center and an update step adjusts the cluster centers to be the mean vector of all the pixels belonging to the cluster. A residual error between the new and previous cluster center locations is computed. Once the algorithm stops. We assigned subscript which denoted segment number to every pixel.
Before construction of noise feature for every segment, we excluded sharp transitional area since noise estimation was adversely affected by heterogeneous image content . We estimated sharp area using its gray-scale image which was calculated by The sharpness edge of image was then obtained by , where represents Sobel operator and “” denotes convolution. We then define whether a pixel is in the sharp area using where means that the pixel is located in sharp transitional area. To guarantee that these areas will not affect noise estimation in the next step, we expand boundaries via dilation by where is a structure element of ones. is the expanded sharp area.
To extract noise feature of each segment produced by previous SLIC algorithm, we firstly employed denoising algorithm across the whole picture. The estimated noise at location of image was calculated by where and filter , represents five different filters used to trace different aspects of the noise . They are median filter, Gaussian filter, averaging filter, and adaptive Wiener denoising with two neighborhood sizes and , respectively. For instance, high frequency noise can be detected by using Gaussian filter and median filter addresses “salt and pepper” noise.
For each combination of color channel and denoising filter we calculated the mean and standard deviation values of each segment as the noise feature , where As a result we computed dimensional feature vector of a segment.
3. Integrated Method for Forgery Detection
We proposed an integrated method effective to both copy-move and splicing forgery. Based on combination of block artificial grid extraction with analysis of local noise discrepancies, the algorithm showed valid performance to high compression JPEG pictures, as well as high quality images lack of BAGs. To implement the authentication process, we built an indicator for every nonoverlapped blocks of the doubtfulful picture. The indicator mathematically described the possibility of the block being a forged area, with higher value denoting higher probability. As described in Figures 4, 5, and 6, for each block, BAG feature and assigned label based on noise discrepancies were integrated by estimated compression indicator. (see Algorithm 1 for a detailed description).
3.1. Block BAG Feature Construction
In Section 2.1 we introduced how to extract BAGs. For an intact picture, the BAGs appear at the border of each block, while for a picture with intentional copy-move or splicing operation some BAGs will be presented at some abnormal positions, such as the center of the block. For a fixed block , these abnormal BAGs can be calculated  by
3.2. Noise Discrepancy Label Assignment
The noise feature of each segment had been calculated in Section 2.2 and then our goal was to segment the image into two regions with noise discrepancies. To achieve the target the energy-based graph cuts can be used.
Energy minimization via graph cuts is proposed by Boykov et al.  to solve labeling problems with low computation cost. In a common label assignment problem, the labels should vary smoothly almost everywhere while preserving sharp discontinuities existing at object boundaries. These two constraints can be formulated as , where is a labeling that assigns each pixel a label and measures the extent to which is not piecewise smooth while measures the disagreement between and observed data. The goal is to minimize the function. Specifically the energy function can be rewritten as where is neighboring pixels, is the penalty of pairs in the first term, and is nonnegative and measures how well label fits pixel. Local minimum value can be obtained with the help of graph cuts. The simplified problem is illustrated in Figure 7. Since many algorithms have been proposed to solve min-cut problem, if proper weight value is assigned to each edge, the problem of minimizing energy function changes to min-cut problem. The weight is seen in Table 1. The calculation result is a cut which separates two labels. Figure 7 shows two possible cuts and the label is assigned to the pixel when cut contains the edge connecting that label to the pixel. For example, in left case of Figure 7, label is assigned to pixel while is assigned to because cut contains edge and .
Our forgery detection task can also be regarded as a labeling problem. In our application, there are two labels that need to be assigned to each segment produced by previously introduced SLIC algorithm: forged area as they show inconsistency to rest segments in terms of noise level or pattern and the original area. And each segment is processed as a pixel. The reason why we avoid employing widely used outlier detection algorithms  and Otsu’s automatic thresholding method  is the property of noise. From Figure 2 we observe that even the picture is taken by one camera and the amount of noise differs in different illumination. The color of object may also affect the noise level. Accordingly the ideal algorithm should tolerate these local deviations and inconsistencies. In other words, it should keep “smooth” across the image while preserving “sharp” discontinuity in inconsistent boundaries. This requirement is identical to label assignment problems described previously while normal outlier detection algorithms are not capable of this.
“Smooth” constraint is realized by proper assignment of and “sharp” discontinuity requirement is supported by . We firstly discuss the weight of edge and . We computed average value of feature vector of all segments in 30 dimensions and named it the mean vector . Then we found the vector whose Euclidean distance was the largest from by searching for all segments and called it . For a feature vector the weight was obtained by where was “original” label while was “forged” and denoted Euclidean distance between two vectors.
From (15) we can find that, if the noise level of a segment is close to the average value across the whole picture, the weight assigned is small while is large and vice versa. This meets the requirement of discontinuity preserving. Then it is the turn to discuss smooth constraint. Proper value of interaction penalty tolerates local deviations of noise which is affected by illumination or color. There are many forms proposed. For an instance, or an important function given by the Potts model where is 1 if its argument is true and otherwise 0. This penalty function possesses good feature of piecewise smooth, so we used it in the experiment.
Graph cut based on noise discrepancy assigned every segment a label indicating whether the area was classified as forged () or not (). And we assigned every pixel belonging to the segment the same label . At last, block indicator described the possibility of forgery and was calculated by
3.3. Feature Generation for Forgery Detection
In this step, we combined together two features described already with proper coefficient. Since the method based on BAG extraction is only sensitive and feasible to highly compressed images, the form of combined feature is described as and denotes the coefficient assigned and is a function of evaluated image compression rate or the JPEG image quality; namely, .
We firstly evaluated the quality of the picture and then found the function . Proposed by Wang et al. , the quality assessment algorithm is nonreferenced and sensitive to JPEG compression rather than noise which was tested and verified by our experiment. We took 20 pictures in raw file (no compression) and then saved them as JPEG format pictures with different compression ratio. In our experiment, 100% means saving with the highest quality and the lowest compression. We assessed the picture quality of compression rate as 100%, 80%, 60%, 40%, 20%, and 5%, respectively, and averaged the scores. See Figure 8 for result: less compressed pictures show higher quality scores.
However, the algorithm is less sensitive to noise effect. In the experiment, for each image set with certain compression rate, we added 10%, 20%, and 40% monologue Gaussian noise to the image, respectively, and then obtained the average quality scores. See Figure 9 for result: noise does not largely affect quality scores. Therefore, we consider that the dominated factor affecting quality score in algorithm  is JPEG compression rate.
Then we discuss how to generate the function . In the experiment, we made 60 fake pictures and every 10 pictures were compressed in a certain rate. And then we used BAG feature only to detect the forgeries. Table 2 shows detection accuracy in different compression rate. The experimental result confirmed that the BAG method is good at dealing with the pictures with low quality score. Therefore, the value of should approach 1 when declines near to 2 for its detection accuracy is 100%. Meanwhile, should be set to 0 when rises to 9 or so because of its low accuracy.
The function we recommended based on experimental result is
In order to filter out some isolated false marked areas and improve the integrity of suspect forged region, morphological operations including closing and opening are used. The final result comes from , where and are circular structure with radius of 5 and 3 pixels, respectively.
4. Experimental Results and Discussion
This part firstly exhibits the experimental results and compares our results with existing algorithm. Then we consider the situation when the input image is slightly compressed. In this circumstance, there are few conspicuous block artificial grids; noise feature becomes predominated since . In order to verify the effectiveness of the proposed method, we tested under two situations noise discrepancies from artificial added noise and from digital cameras.
4.1. Detection Results and Comparison
As it is mentioned at the beginning, our proposed method can deal with both copy-move and splicing forgery with one authentication process. Two detection results are shown in Figure 10: the marked white area is detected forged region. Our algorithm shows good performance in these two types of forgery.
Then we compared our proposed method with existing algorithm in . In the experiment, we prepared six sets of test images. In each set, there were 25 pictures including intact and fake pictures with copy-move or splicing forgery. The difference between sets was the image quality-JPEG compression rate. Table 3 shows the comparison of detection accuracy between two methods; when the image is greatly degraded by high JPEG compression, two methods present valid performance. However, if the forged image is saved with slightly compression, the detection accuracy of Li’s method drops significantly, while our method still maintains high accuracy. Figure 11 shows an instance of detection result comparison between two methods.
4.2. Simulation Results
In this part we present a simulated forgery case that the noise is added to implanted region. This simulation also reflects a real splicing attack that in order to make the alien area visually resemble the rest part of picture noise may be applied. Since Photoshop is a popular image editing tool, we add noise to picture with provided filters by software. There are two noise distribution options: Gaussian and uniform and two noise patterns: monochrome and colored. Therefore four combinations are available and the user can alter the noise amount in percentage. The experiment is designed to demonstrate the sensibility of algorithm which is the lower limit amount of added noise that can be detected by our method. Figure 12 shows the detection accuracy of four groups, each of which contains five forged pictures. We conclude that the effective lower limit for detection is 1.4% for Gaussian noise and 2.2% for uniform noise regardless of monochrome or colored noise pattern.
4.3. ISO and Detection Results
Two image datasets are prepared to verify the effectiveness of our proposed method. In the first set, all source pictures were taken by a Nikon D7000 DSLR camera and used to make splicing forgeries in combination of different ISO speed seen in Table 4. There are 10 forged pictures in the test set. The data in this table is the detection accuracy or true positive rate.
The ISO speed setting in camera is discrete without the same interval and we find that the higher TP rate appears at combination of two ISO speeds with big gap. In order to see this phenomenon clearly, we can see Figure 13. The horizontal axis is marked by interval stop(s) which denotes the interval ISO speed. For instance, the interval stop of ISO 100 and 200 is 1; this is the same with ISO 1600 and 3200, while that of ISO 200 and 1600 is 3. The average TR rate is calculated from Tables 4 and 5. We conclude that our method shows good performance in two or more interval stops.
|not verified in experiment.|
The second experiment is to verify the effectiveness of detecting forgery in pictures combined from two different cameras. And in the paper we just show an extremely hard situation when the source pictures are taken in the same ISO speed. Two cameras are Nikon D7000 and Canon 550D, respectively. And 10 forged images in the set are used to the test. The TP rate is shown in Figure 14. And the accuracy increases as the ISO speed rises. The reason is that the image processing ability of two camera models is not the same. In lower ISO speed, less noise appears in the picture and this processing difference is small; therefore the TP rate is very low at 10%, while in high ISO settings, the method shows effectiveness again. Note that, in real situation, the ISO of two source pictures may not be the same; only one interval stop will highly enhance the accuracy as it is shown in the first experiment.
In this paper, we concentrated on exposing the two main types of image manipulation, copy-move and splicing forgery. We proposed an integrated algorithm to locate forged regions by a single authentication process. In our method, JPEG block artificial grids and local noise discrepancies were used to generate features which were combined with image quality score as coefficient. Experimental result shows that our approach is valid to both highly compressed and high quality pictures. Comparing to existing algorithms, our method has competitive advantages and a larger range of application.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
The authors would like to thank the referees for their valuable comments. This research was supported in part by the Research Committee of the University of Macau and the Science and Technology Development Fund of Macau SAR (Project nos. 034/2010/A2 and 008/2013/A1).
- J. Granty Regina Elwin, T. S. Aditya, and S. Madhu Shankar, “Survey on passive methods of image tampering detection,” in Proceedings of the International Conference on Communication and Computational Intelligence (INCOCCI '10), pp. 431–436, December 2010.
- J. A. Redi, W. Taktak, and J.-L. Dugelay, “Digital image forensics: a booklet for beginners,” Multimedia Tools and Applications, vol. 51, no. 1, pp. 133–162, 2011.
- I. Amerini, L. Ballan, R. Caldelli, A. Del Bimbo, and G. Serra, “A SIFT-based forensic method for copy-move attack detection and transformation recovery,” IEEE Transactions on Information Forensics and Security, vol. 6, no. 3, pp. 1099–1110, 2011.
- X. Pan and S. Lyu, “Detecting image region duplication using sift features,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '10), pp. 1706–1709, March 2010.
- P. Kakar, N. Sudha, and W. Ser, “Exposing digital image forgeries by detecting discrepancies in motion blur,” IEEE Transactions on Multimedia, vol. 13, no. 3, pp. 443–452, 2011.
- S. Hamdy, H. El-Messiry, M. Roushdy, and E. Kahlifa, “Quantization table estimation in jpeg images,” International Journal of Advanced Computer Science and Applications, vol. 1, no. 6, pp. 17–23, 2010.
- F. Huang, J. Huang, and Y. Q. Shi, “Detecting double JPEG compression with the same quantization matrix,” IEEE Transactions on Information Forensics and Security, vol. 5, no. 4, pp. 848–856, 2010.
- S. D. Lin and T. Wu, “An integrated technique for splicing and copy-move forgery image detection,” in Proceedings of the 4th International Congress on Image and Signal Processing (CISP '11), vol. 2, pp. 1086–1090, October 2011.
- W. Li, Y. Yuan, and N. Yu, “Passive detection of doctored JPEG image via block artifact grid extraction,” Signal Processing, vol. 89, no. 9, pp. 1821–1829, 2009.
- W. Li, N. Yu, and Y. Yuan, “Doctored JPEG image detection,” in Proceedings of the IEEE International Conference on Multimedia and Expo (ICME '08), pp. 253–256, June 2008.
- L. G. Shapiro and G. C. Stockman, Computer Vision, Prentice-Hall, New Jersey, NY, USA, 2011.
- R. Achanta, A. Shaji, K. Smith, A. Lucchi, P. Fua, and S. Süsstrunk, “SLIC superpixels compared to state-of-the-art superpixel methods,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no. 11, pp. 2274–2282, 2012.
- J. Fan, H. Cao, and A. C. Kot, “Estimating EXIF parameters based on noise features for image manipulation detection,” IEEE Transactions on Information Forensics and Security, vol. 8, no. 4, pp. 608–618, 2013.
- H. Gou, A. Swaminathan, and M. Wu, “Intrinsic sensor noise features for forensic analysis on scanners and scanned images,” IEEE Transactions on Information Forensics and Security, vol. 4, no. 3, pp. 476–491, 2009.
- Y. Boykov, O. Veksler, and R. Zabih, “Fast approximate energy minimization via graph cuts,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 11, pp. 1222–1239, 2001.
- V. J. Hodge and J. Austin, “A survey of outlier detection methodologies,” Artificial Intelligence Review, vol. 22, no. 2, pp. 85–126, 2004.
- N. Otsu, “A threshold selection method from gray-level histograms,” Automatica, vol. 11, pp. 285–296, 1975.
- Z. Wang, H. R. Sheikh, and A. C. Bovik, “No reference perceptual quality assessment of JPEG compressed images,” in Proceedings of the International Conference on Image Processing (ICIP '02), vol. 1, pp. I/477–I/480, September 2002.
Copyright © 2014 Bo Liu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.