The Scientific World Journal

The Scientific World Journal / 2014 / Article
Special Issue

Security of Information and Networks

View this Special Issue

Research Article | Open Access

Volume 2014 |Article ID 606570 | 10 pages | https://doi.org/10.1155/2014/606570

Improving RLRN Image Splicing Detection with the Use of PCA and Kernel PCA

Academic Editor: Iftikhar Ahmad
Received02 Jun 2014
Revised26 Jul 2014
Accepted13 Aug 2014
Published14 Sep 2014

Abstract

Digital image forgery is becoming easier to perform because of the rapid development of various manipulation tools. Image splicing is one of the most prevalent techniques. Digital images had lost their trustability, and researches have exerted considerable effort to regain such trustability by focusing mostly on algorithms. However, most of the proposed algorithms are incapable of handling high dimensionality and redundancy in the extracted features. Moreover, existing algorithms are limited by high computational time. This study focuses on improving one of the image splicing detection algorithms, that is, the run length run number algorithm (RLRN), by applying two dimension reduction methods, namely, principal component analysis (PCA) and kernel PCA. Support vector machine is used to distinguish between authentic and spliced images. Results show that kernel PCA is a nonlinear dimension reduction method that has the best effect on R, G, B, and Y channels and gray-scale images.

1. Introduction

For many decades, photography has served a vital function in the lives of people and has been considered as one of the most important revolutions in recording moments. However, photographs have recently lost their trust because of the increasing use of manipulation tools, which have made photographs less trustable. The history of image tampering (i.e., creating photographs that never really happened in real life) is as old as the art of photography itself. Images have been applied for malicious purposes in many instances. In 2003, the Los Angeles Times printed on its front cover an image from photojournalist, Brian Walski, which showed a British soldier in Iraq trying to control a crowd of civilians in a passionate manner. However, the depicted moment never happened. The photograph was a composite of two different photographs merged to create an appealing image. The image tampering was discovered, and Walski was fired [1]. In 2009, the Brazilian newspaper Folha de Sao Paulo published an article that contained a spliced image created from different gray-scale images to show how the Brazilian Chief of Staff actively participated in the resistance against the military regime through such actions as the planning of and preparation for robberies and kidnappings [1].

An innovative research area called digital image forensics emerged from the necessity to regain the trustability of photographs. Digital image forensics is generally categorized into two basic groups: active methods and passive methods [2]. Active methods borrow additional information to insert into the original image, whereas passive methods require no prior information. Digital signatures and watermarking are the most common active detection methods. Passive methods focus on statistical modifications made on image content. Such modifications include splicing, retouching, healing, copy-move, and blurring. Image splicing is best defined as a combination of two or more parts of different photographs to create a new image, which is known as a spliced image. Image splicing deceives the human visual system to achieve its malicious purposes.

Studies were conducted on this topic, and several image splicing detection methods were proposed and developed. Farid [3] considered speech signal splicing as a highly nonlinear process and applied higher order spectral analysis to solve this problem. Ng et al. [4] extended the previously mentioned scheme into an image splicing detection method with the assumption that the image splicing procedure is nonlinear and that the image involved is nonstationary. Their expanded method achieved an unsatisfactory detection performance of 72%.

Fu et al. [5] proposed an image splicing detection approach using the Hilbert-Huang transform (HHT). They considered the high nonlinearity and nonstationary nature of the image splicing operation and merged this technique with a statistical natural image model on the basis of moments of characteristic functions with wavelet decomposition. Their method obtained an accuracy rate of 80.15%. Li et al. [6] combined two methods, namely, moment features based on first-order histogram of the image discrete wavelet transform (DWT) coefficients and HHT, to extract features for image splicing detection. Their method obtained an average accuracy rate of 85.87%.

Shi et al. [7] proposed a blind, passive, and natural image splicing detection model based on statistical feature extraction methods. Their model includes the Markov transition probabilities and moments of characteristic functions of wavelet subbands applied to different 2D arrays and to the 2D arrays of multisize block discrete cosine transform. Their model achieved a detection accuracy rate of 91.8%, which is a promising improvement in the image splicing detection area.

Zhang et al. [8] applied an idea in image steganalysis [9] by merging the Markov features and discrete cosine transform (DCT) features. They achieved a detection rate of 91.5%. He et al. [10] expanded the natural Markov-based model applied in [7] by capturing the interblock correlation between the block DCT coefficients proposed in [11] and merging it with the features generated from the DWT domain. They also reduced their features by applying a feature selection method called support vector machine recursive feature elimination (SVM-RFE), which achieved a detection rate of 93.55% on a digital video multimedia (DVMM) image dataset [12].

Dong et al. [13] proposed an approach based on statistical features obtained from the run length method and on image edge statistics from the blind image splicing detection method. He et al. [14] improved the method proposed in [13] and developed a detection algorithm based on approximate run length. Their results showed a moderate detection accuracy rate (75% versus 69%) but with a lesser amount of time than in the original algorithm (6D versus 12D). Zhao et al. [15] proposed applying the run length run number (RLRN) vectors in four directions of the chroma spaces because detecting spliced images in one color space is difficult. Their results showed that the features extracted from the chroma channels were more accurate than those extracted from the R, G, B, and Y channels.

The existing methods concentrate only on the actual image splicing detection techniques, and handling the extracted high dimensional and redundant features can be a difficult and time consuming process. Therefore, dimensionality reduction methods are applied to remove redundant information from the extracted features and to obtain the most discriminative information with less dimensionality.

In this paper, we compared the two dimension reduction methods (PCA versus kernel PCA) that were developed. This paper aims to evaluate the detection accuracy and computational time of PCA and kernel PCA on RLRN [15] in dimensionalities of 10 and 50. The rest of this paper is organized as follows. The proposed approach is described in the second section, the experimental results are discussed in the third section, and the conclusion and future works are presented in Section 4.

2. The Proposed Approach

Figure 1 demonstrates the approach proposed in this paper. The approach includes four phases: preprocessing, feature extraction, dimension reduction, and classification. The given image from image dataset 1 [12] or image dataset 2 [16] is initially preprocessed for extraction. The features are then extracted from the preprocessed image using RLRN [15]. The dimensionality is then reduced to 10 and 50 dimensions by applying PCA and kernel PCA. Finally, the dimensional (D) feature vector is used as input for the support vector machine (SVM) classifier to calculate the detection accuracy and computational time.

2.1. Preprocessing

Preprocessing improves image content by reducing undesired distortions and/or enhancing image features relevant to further processing. A difference between 2D arrays is used to reduce the correlation between image pixels/coefficients and the effect caused by the diversity in image content [15, 17]. The difference between 2D arrays is computed as follows [15, 17]: where represents the th and th element of the gray value of the images in the image matrix and , , , and represent the elements along the horizontal, vertical, diagonal, and minor-diagonal directions, respectively.

2.2. Feature Extraction Methods

Image features include the global and local properties of an image such as average gray levels, intensity histogram shapes, circles, lines, texels, and contour shapes. Different methods have been developed to extract these features, and these methods have been applied in various image processing fields. Given that image splicing detection is a two-class problem (i.e., one class for authentic images and another class for spliced images), features extracted from images serve an important function in the detection and classification process, which aims to distinguish authentic images from spliced ones. In this paper, a recently used feature extraction method is presented to examine the effect of dimension reduction techniques on detection performance.

2.2.1. RLRN

RLRN is not extensively applied as a feature extraction method. However, the results which have been obtained in [15] show that this method can be used as an image splicing detection approach. The method was initially used in [18] in texture analysis to classify a set of terrain samples. A new run length algorithm was developed in [19] to extract texture features based on a multilevel dominant eigenvector estimation method, which improves classification accuracy.

This paper applied RLRN and its definition and mathematical equations are presented in this section. Zhao et al. [15] suggested using this method. A run in an image is defined as the number of pixels with the same gray level value in a specific direction. For a given image, a run length matrix is defined as the number of runs with gray level and run length along a specific direction. Hence, the run length vector is defined as follows [15]: where represents the number of gray levels and is the maximum value of run lengths. Vector demonstrates the sum distribution of runs with length in a given image. In the equation, run length represents the spread of the image structure and texture. The image with a long run length is smoother than that with a short run length because the latter has different regions with different structures. The gray level run length pixel number matrix is used to equally emphasize all run lengths in this paper. The matrix is defined as follows [15, 19]: where where is the feature vector applied and referred to as RLRN. Four RLRN vectors were captured in four directions (0°, 45°, 90°, and 135°) from every channel (R, G, B, Y, Cb, and Cr) in image dataset 2 and from gray-scale images in image dataset 1 to distinguish the spliced images from the authentic images. The RLRN feature vector is shown as a 100D feature vector by obtaining the first 25 features of every vector in each orientation .

The run length of an image represents its structure and texture, and the splicing procedure modifies the pixel correlations and image structure. Therefore, the RLRN feature extraction method can represent discontinuities and nonconformity and can be efficiently used as an image splicing detection method [13, 15].

2.3. Dimension Reduction Methods

Dimension reduction methods are applied to reduce feature dimensionality by eliminating redundant features and keeping important dimensions in the feature vector. Humans and machine learning methods find it difficult to interpret high dimensional data. Given that a feature matrix has rows that each represents a specific instance of an object and a large number of features exponentially increase the computational time. Thus, decreasing information into smaller sizes enhances method analysis and improves the training and testing phases during classification [20]. Several experiments were conducted to test and analyze this idea.

Different approaches such as finding the linear or nonlinear manifold that lies within the high dimensional data space can simplify interpretation. In this section, the PCA and kernel PCA are presented to improve the features extracted by the RLRN feature extraction technique discussed in the previous section. Moreover, linearity (PCA) and nonlinearity (kernel PCA) were considered in selecting the dimension reduction techniques to investigate their effects on the extracted features.

2.3.1. PCA

PCA is the most common and popular linear dimension reduction approach [2123]. It has been used for years because of its conceptual simplicity and computation efficiency. The approach is applied in many areas such as noise reduction, pattern recognition, regression estimation, and image indexing [24]. It maps a dataset of dimensions to a linear subspace with dimensions, where , and attempts to maintain most of the variability in the mapped dimensions. PCA is considered a second-order approach depending on the covariance matrix of the variables. The approach has different names in different fields such as singular value decomposition, Karhunen-Love transform, Hotelling transform, and empirical orthogonal function method [25].

PCA is based on finding the orthogonal linear vectors, known as principal components, of dimensions with maximum variance. Therefore, the number of reduced dimensions is not more than . The approach works well if the most significant modes of variability are almost linear. Hence, high dimensional samples are best remade from their low dimensional linear projections. Otherwise, PCA becomes ineffective if the most vital significant modes of variability are nonlinear [21]. In mathematical terms, PCA finds as the new feature vector set with dimension , in which is the original feature vector set with dimension [26]: To find linear mapping M, PCA attempts to maximize the following function: where is the covariance of the original feature vector set . However, consists of principal eigenvectors of the sample covariance matrix of the zero-mean data [26]. Therefore, the following eigenproblem must be solved for the principal components :

2.3.2. Kernel PCA

PCA is a linear dimension reduction method. Some datasets have a nonlinear nature, and PCA cannot reduce the dimensions of these datasets efficiently. Thus, kernel PCA was designed to address this problem. Kernel PCA was applied in some pattern recognition experiments [24] and exhibited better recognition rates than linear PCA. Kernel PCA is a nonlinear form of PCA that attempts to identify complicated correlations between given features. It computes principal components in the original dataset through nonlinear mapping. It also discovers major components that are nonlinear in relation to the input space by running, which results from nonlinear mapping in which the low dimensional hidden structures are likely to be simple [21]. Kernel PCA locates the principal eigenvectors of the kernel matrix instead of the covariance matrix [26]. Thus, the computational complexity of kernel PCA is independent of the dimensionality of the feature set, which allows it to work on feature sets with different possible dimensionalities. Kernel PCA does not require any nonlinear optimization; it only needs to solve an eigenvalue problem as in the case of standard PCA. Thus, kernel PCA is free of local minima trap during training. The original feature set must be mapped to a higher dimensional feature set to calculate kernel PCA [24]: Then, the covariance matrix of data is calculated to obtain the principal components by solving the eigenvalue problem using the following equations: Subsequently, the eigenvector can be expressed as a linear combination of features: Therefore, the kernel matrix is defined as follows: where represents the elements of kernel matrix , is the feature set, and is the kernel function with conditions that result in a positive semidefinite kernel [26]. may not be zero-mean such that the features must be centered. The corresponding kernel is obtained using the following equation: Consequently, the following equation represents the low dimensional feature set : where represents the th value in the vector .

High correlations are generally found among the extracted features using RLRN. PCA and kernel PCA are applied to reduce the correlations by eliminating the information redundancies from the features. Figure 2 shows the standard deviation distribution of the features extracted from gray-scale images (R and Cb channels of the colored images) before and after applying PCA and kernel PCA, respectively. The standard deviation measures and shows how data are spread out from the mean. In this case, a high standard deviation implies a high correlation between the features. Figure 2 shows that the original features are highly correlated and their standard deviations are spread over a wide range in 100D. After applying PCA, the standard deviations mostly concentrate on the first few features and decrease as dimensionality increases. However, the standard deviations are still high and considerable. In contrast, the standard deviations greatly reduced after applying kernel PCA on the original features. The features after applying kernel PCA were obviously highly uncorrelated.

2.4. Classifier

SVM is one of the most popular supervised machine learning algorithms applied in pattern recognition. The Matlab codes for this classifier are available in [27]. In this work, LIBSVM was specifically used as a classifier, while radial basis function was used as a kernel function. A grid search method was applied to obtain the best value for the and parameters. All authentic images were labeled as −1 and all spliced images were labeled as +1 during the classification process.

3. Experimental Results

A set of experiments that demonstrate the effectiveness of the proposed approach are described. Our classification system was implemented using MATLAB R2012a on a 2.40 GHz Intel (R) Core i5 processor with 4 GB RAM on a Windows 7 platform.

3.1. Image Dataset

Two image datasets (gray and colored) were applied to evaluate the proposed method. The first image dataset was the Columbia Image Splicing Detection Evaluation Dataset provided by the Digital Video MultiMedia (DVMM) Laboratory, Columbia University (2007) [12]. This dataset contained 1845 gray-scale images (933 authentic images and 912 spliced images) in BMP format. DVMM was the only gray-scale image dataset designed for image splicing detection evaluation. Almost all papers applied DVMM, so it was also used in this work for better comparison with other methods. Another image dataset designed by the Chinese Academy of Sciences, Institute of Automation (CASIA), with high resolution images was also applied.

The CASIA tampered image detection evaluation database [16] is another image dataset designed to evaluate image splicing detection methods. Version 1.0 of this dataset included 1721 color images (800 authentic images and 921 spliced images) with 384 pixels × 256 pixels in JPEG format and was used in our experiments to evaluate the proposed approach. Examples of both image datasets are presented in Figure 3 (DVMM image dataset) and Figure 4 (CASIA image dataset). The first row consists of authentic images, while the second row consists of spliced images [12, 16]. Each image dataset is divided into two groups during the classification process: training (5/6 of the images) and testing (1/6 of the images). These groups were randomly selected to reduce the nondeterministic properties of the classifier.

3.2. Classification

For satisfactory results, PCA and kernel PCA were applied on RLRN feature extraction method with different dimensionalities (10D and 50D) to evaluate the detection accuracy and computational time. The results were presented in three tables. The true positive (TP) and true negative (TN) represent the detection rate of the authentic and spliced images, respectively. Accuracy (Acc.) represents the average detection rate. Computational time is represented in seconds (s) in each table.

Table 1 illustrates the results from the original dimension of the RLRN method with 100D obtained from different channels (RGB, gray, and YCbCr). The results show that the detection accuracies in the R, G, and B channels varied in a close range (64.66% to 69.30%), which verifies the strong correlation among the three channels because of the color filter array interpolation process. The luma channel (Y) that is correlated with the RGB channel (i.e., Y is a linear combination of R, G, and B channels) [15] also shows similar results (63.09%) with RGB. The results obtained from the Cb and Cr channels exhibited the best detection accuracies (89.59% and 93.80%) among the channels, since RLRN is more sensitive to the chroma channels (Cb and Cr) than the luma one [15]. However, the computational times for the Cb and Cr channels were the least among the channels.


100D
TP (%)TN (%)Acc. (%)Time (s)

R63.9174.6869.30388.43
G57.8971.4364.66402.83
B55.6475.3265.48426.27
Gray70.5175.0072.75489.72
Y58.6567.5363.09408.01
Cb90.2388.9689.59258.73
Cr94.7492.8693.80252.32

Table 2 presents the detection accuracy and computational time results from applying PCA on RLRN with 10 and 50 dimensions. The results indicate the same behavior among different channels with the results obtained from 100D. A reduction in detection accuracy was observed for 10D compared with 100D. All detection accuracies slightly increased for 50D, except for the gray and Cr channels. Computational time in 50D also reduced compared with 100D. The results presented in Table 2 indicate a high correlation between the features after PCA application, which was proven in Figure 3.


PCA10D50D
TP (%)TN (%)Acc. (%)Time (s)TP (%)TN (%)Acc. (%)Time (s)

R60.9073.3867.14447.8666.1774.6870.43302.57
G55.6372.0863.86457.4661.6577.9269.79270.28
B67.6769.4868.57504.2164.6672.7368.69293.34
Gray67.9475.0071.43462.4965.3873.6869.53297.29
Y54.8972.7363.81449.5559.4070.7865.09275.09
Cb81.2087.0184.11203.2587.9793.5190.74182.66
Cr83.4692.2187.83187.7190.9887.6689.32187.93

Table 3 demonstrates the detection accuracy and computational time that resulted from combining the RLRN method with kernel PCA. These results show that gray-scale images with 50D exhibited a considerable increase in detection accuracy compared with those obtained from the original and the combined methods (88.28% versus 72.73% and 69.53%), respectively. The R, G, B, and Y channels also exhibited a substantial growth of 25% in detection accuracy. These results were anticipated from Figure 3, which indicated a low correlation among features after kernel PCA application.


Kernel PCA10D50D
TP (%)TN (%)Acc. (%)Time (s)TP (%)TN (%)Acc. (%)Time (s)

R84.9683.1284.04122.0785.7190.9188.31183.91
G84.9683.1284.04114.487.2290.2688.74193.37
B84.9683.1284.04125.6785.7190.9188.31223.77
Gray100050.0063.8790.3886.1888.28236.49
Y82.7186.3684.54112.7887.2290.2688.74188.15
Cb84.9683.1284.04111.6787.8190.9189.36186.7
Cr84.9683.1284.04111.5885.7190.9188.31186.62

However, the Cb and Cr channels did not follow the same trend and exhibited a slight decrease in 50D. The other dimensionalities were also tested for the Cb and Cr channels because of this decrease, and the results indicated that the optimal dimensionality for these two channels was in 95D with detection accuracy of 92.68%. Thus, the Cb and Cr channels contain important information focused on the first 95 dimensions. All computational time generally reduced for 10D and 50D in comparison with 100D.

Combining the RLRN method with kernel PCA in 50D generally exhibited the best results among the methods, which verifies the nonlinear nature of RLRN. Figures 5, 6, and 7 demonstrate the receiver operating characteristic (ROC) curves for gray-scale images and R and Cr channels, respectively. A comparison of the features extracted from the original feature extraction method in 100D was made in each figure with the extracted features from the merged ones with PCA and kernel PCA dimension reduction methods in 50D. The results obtained from 10D were ignored because of the high amount of results presented here. Figure 5 indicates the best effect of kernel PCA on RLRN, while PCA has almost the same results with the original features.

The R, G, B, and Y channels have similar behaviors; the R channel was selected to represent the ROC in Figure 6. According to this ROC, the R, G, B, and Y channels performed better when the dimensionality of the features was reduced to 50D by applying kernel PCA.

Figure 7 shows an inverse effect of kernel PCA on the obtained features from the Cb and Cr channels. The results show that the original features performed the best, and the dimension reduction methods did not improve accuracy rate.

3.3. Comparison with Other Methods

A comparison of some of state-of-the-art image splicing detection methods was conducted for a comprehensive evaluation of the entire system. Table 4 indicates the comparison between different methods and the proposed framework for gray-scale images in image dataset 1.


Feature extraction methodsDimensionalityTP (%)TN (%)Acc. (%)

Expanded DCT Markov [10]10089.9290.2190.07
DWT Markov [10]10087.5885.3986.50
Expanded DCT Markov + DWT Markov [10]10093.2893.8393.55
HHT + moments of characteristic functionswith wavelet decomposition [5]11080.2580.0380.15
Run length and edge statistics based model [13]16383.2385.5384.36
RLRN + kernel PCA (proposed)5090.3886.1888.28

Table 4 shows that the accuracy rates exhibited different trends. Computational time could not be compared because it was not provided in the other methods. The best results were observed for the expanded DCT Markov + DWT Markov and expanded DCT Markov [10], which reduced to 100D by applying the SVM-RFE method (93.55% and 90.07%). The next best accuracy rate was observed for our proposed method (88.28%) with 50D. The results show that our proposed method performed the best compared with similar methods [13] with less dimensions (163D versus 50D).

Table 5 shows another comparison between the original RLRN [15] in 60D and the proposed method (RLRN + kernel PCA) in 50D on image dataset 2 (CASIA). The Gray indicated in the table was obtained by converting the colored images in image dataset 2 to gray-scale images. As previously discussed, a huge increase in detection accuracies in the R, G, B, and Y channels, as well as the gray one, was observed.


RLRN [15]RLRN + kernel PCA
TP (%)TN (%)Acc. (%)TP (%)TN (%)Acc. (%)

R56.3083.7070.985.7190.9188.31
G51.8083.2068.5087.2290.2688.74
B57.2083.7071.3085.7190.9188.31
Gray60.4081.7071.8086.4790.9188.69
Y53.3083.1069.2087.2290.2688.74
Cb91.7096.5094.3087.8190.9189.36
Cr91.8097.1094.7085.7190.9188.31

4. Conclusion

The literature review presented in this work showed that several image splicing detection methods were proposed. Unfortunately, many of them are unable to handle the extracted high dimensional and redundant features well. In addition, processing these features is time consuming. Therefore, this paper focused on evaluating the effectiveness of dimension reduction methods on image splicing detection methods to reduce dimensionality and computational time and remove redundant information from the features. Two dimension reduction methods (PCA: linear and kernel PCA: nonlinear) were selected and applied on an instance image splicing method (RLRN). A set of experiments were designed and tested to determine the effectiveness of these methods on RLRN. Each dimension reduction method was applied on RLRN to reduce dimensionality to 10D and 50D. The results showed that the R, G, B, and Y channels and gray-scale images performed best when merged with kernel PCA and in 50D, which verifies the nonlinear nature of the RLRN features. However, the results also demonstrate that this area of study requires further research. Other dimension reduction methods and their effects on other image splicing detection methods must be investigated, and the optimal dimension number for every image splicing detection approach must be identified.

Abbreviations

D: dimensional
RLRN:Run length run number
PCA:Principle component analysis
SVM:Support vector machine
HHT:Hilbert-Huang transform
DWT:Discrete wavelet transform
DCT:Discrete cosine transform
SVM-RFE:Support vector machine recursive feature elimination
DVMM:Digital video multimedia
CASIA:Chinese Academy of Sciences, Institute of Automation
TP:True positive
TN:True negative
Acc.:Accuracy
ROC:Receiver operating characteristics.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

The authors would like to thank both reviewers for their valuable comments which helped to improve the paper. This research is supported by University of Malaya Research Grant (UMRG) RG312-14AFR.

References

  1. A. Rocha, W. Scheirer, T. Boult, and S. Goldenstein, “Vision of the unseen: current trends and challenges in digital image and video forensics,” ACM Computing Surveys, vol. 43, no. 4, article no. 26, 2011. View at: Publisher Site | Google Scholar
  2. W. Wang, J. Dong, and T. Tan, “A survey of passive image tampering detection,” in Digital Watermarking, pp. 308–322, Springer, 2009. View at: Google Scholar
  3. H. Farid, Detecting Digital Forgeries Using Bispectral Analysis, 1999.
  4. T.-T. Ng, S.-F. Chang, and Q. Sun, “Blind detection of photomontage using higher order statistics,” in Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS '04), pp. V-688–V-691, IEEE, May 2004. View at: Google Scholar
  5. D. Fu, Y. Q. Shi, and W. Su, “Detection of image splicing based on hilbert-huang transform and moments of characteristic functions with wavelet decomposition,” in Digital Watermarking, pp. 177–187, Springer, New York, NY, USA, 2006. View at: Google Scholar
  6. X. Li, T. Jing, and X. Li, “Image splicing detection based on moment features and Hilbert-Huang Transform,” in Proceedings of the IEEE International Conference on Information Theory and Information Security (ICITIS '10), pp. 1127–1130, December 2010. View at: Publisher Site | Google Scholar
  7. Y. Q. Shi, C. Chen, and W. Chen, “A natural image model approach to splicing detection,” in Proceedings of the 9th workshop on Multimedia & security, pp. 51–62, ACM, September 2007. View at: Publisher Site | Google Scholar
  8. J. Zhang, Y. Zhao, and Y. Su, “A new approach merging Markov and DCT features for image splicing detection,” in Proceedings of the IEEE International Conference on Intelligent Computing and Intelligent Systems (ICIS '09), pp. 390–394, November 2009. View at: Publisher Site | Google Scholar
  9. Y. Q. Shi, C. Chen, and W. Chen, “A Markov process based approach to effective attacking JPEG steganography,” in Information Hiding, Springer, 2007. View at: Google Scholar
  10. Z. He, W. Lu, W. Sun, and J. Huang, “Digital image splicing detection based on Markov features in DCT and DWT domain,” Pattern Recognition, vol. 45, no. 12, pp. 4292–4299, 2012. View at: Publisher Site | Google Scholar
  11. C. Chen and Y. Q. Shi, “JPEG image steganalysis utilizing both intrablock and interblock correlations,” in Proceedings of the IEEE International Symposium on Circuits and Systems (ISCAS '08), pp. 3029–3032, Seattle, Wash, USA, May 2008. View at: Publisher Site | Google Scholar
  12. T.-T. Ng, S.-F. Chang, and Q. Sun, “A data set of authentic and spliced image blocks,” ADVENT Technical Report 203-2004, Columbia University, New York City, NY, USA, 2004, http://www.ee.columbia.edu/ln/dvmm/. View at: Google Scholar
  13. J. Dong, W. Wang, T. Tan, and Y. Q. Shi, “Run-length and edge statistics based approach for image splicing detection,” in Digital Watermarking, pp. 76–87, Springer, 2009. View at: Google Scholar
  14. Z. He, W. Lu, and W. Sun, “Improved run length based detection of digital image splicing,” in Digital Forensics and Watermarking, pp. 349–360, Springer, 2012. View at: Google Scholar
  15. X. Zhao, J. Li, S. Li, and S. Wang, “Detecting digital image splicing in chroma spaces,” in Digital Watermarking, vol. 6526 of Lecture Notes in Computer Science, pp. 12–22, Springer, Berlin, Germany, 2011. View at: Publisher Site | Google Scholar
  16. “CASIA tampered image detection evaluation database,” http://forensics.idealtest.org/. View at: Google Scholar
  17. D. Zou, Y. Q. Shi, W. Su, and G. Xuan, “Steganalysis based on Markov model of thresholded prediction-error image,” in Proceedings of the IEEE International Conference on Multimedia and Expo (ICME '06), pp. 1365–1368, IEEE, July 2006. View at: Publisher Site | Google Scholar
  18. M. M. Galloway, “Texture analysis using gray level run lengths,” Computer Graphics and Image Processing, vol. 4, no. 2, pp. 172–179, 1975. View at: Google Scholar
  19. X. Tang, “Texture information in run-length matrices,” IEEE Transactions on Image Processing, vol. 7, no. 11, pp. 1602–1609, 1998. View at: Publisher Site | Google Scholar
  20. K. Anusudha, S. A. Koshie, S. S. Ganesh, and K. Mohanaprasad, “Image splicing detection involving moment-based feature extraction and classification using artificial neural networks,” ACEEE International Journal on Signal & Image Processing, vol. 1, no. 3, p. 9, 2010. View at: Google Scholar
  21. A. Ghodsi, Dimensionality Reduction a Short Tutorial, Department of Statistics and Actuarial Science, University of Waterloo, Ontario, Canada, 2006.
  22. J. E. Jackson, A User's Guide to Principal Components, John Wiley & Sons, New York, NY, USA, 1991.
  23. I. T. Jolliffe, Principal Component Analysis, vol. 487, Springer, New York, NY, USA, 1986.
  24. B. Schölkopf, A. Smola, and K.-R. Müller, “Kernel principal component analysis,” in Artificial Neural Networks—ICANN '97, vol. 1327 of Lecture Notes in Computer Science, pp. 583–588, Springer, Berlin, Germany, 1997. View at: Google Scholar
  25. I. K. Fodor, “A survey of dimension reduction techniques,” Tech. Rep. UCRL-ID-148494, Lawrence Livermore National Laboratory, 2002. View at: Google Scholar
  26. L. J. E. O. Postma, H. J. van den Herik, and L. J. van der Maaten, “Dimensionality reduction: a comparative review,” Journal of Machine Learning Research, vol. 10, pp. 66–71, 2009. View at: Google Scholar
  27. C.-C. Chang and C.-J. Lin, “LIBSVM: a Library for support vector machines,” ACM Transactions on Intelligent Systems and Technology, vol. 2, no. 3, article 27, 2011. View at: Publisher Site | Google Scholar

Copyright © 2014 Zahra Moghaddasi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

1515 Views | 543 Downloads | 24 Citations
 PDF  Download Citation  Citation
 Download other formatsMore
 Order printed copiesOrder

We are committed to sharing findings related to COVID-19 as quickly and safely as possible. Any author submitting a COVID-19 paper should notify us at help@hindawi.com to ensure their research is fast-tracked and made available on a preprint server as soon as possible. We will be providing unlimited waivers of publication charges for accepted articles related to COVID-19.