Abstract

Currently, digital images are widely communicated by media using social media applications. The general public captures the digital images for preserving the family and personal memories and to share with their friends and family. Digital images have been used extensively in forensic science to present the digital images as proof in the court and law enforcement agencies, which present a loophole for the culprits to forge the digital image and change the proofs and evidence. Copy-move forgery (CMF) is among the most widely employed image manipulation methods. In this method, the area of the image is duplicated to some other part to modify its content by applying different postprocessing operations on images like blurring, color reduction, and scaling which is a challenging research problem in copy-move forgery detection (CMFD). In this paper, an efficient and effective CMFD method is presented to identify the single and multiple altered areas in an image in the presence of postprocessing operations. The proposed CMFD method divides the image into circular blocks. It computes a rotation-invariant feature vector from each circular block of the image by applying local intensity order pattern (LIOP) features. The computed feature vectors are then compared using Euclidean distance to locate the suspected image’s forged areas. The experimental results of the proposed CMFD method are reported on three standard datasets of the CMF, namely, CoMoFoD, KLTCI, and MICC-F220. The experimental analysis of the proposed CMFD method on these datasets indicates that it produces robust performance (detection accuracies of 97.29% on the CoMoFoD dataset, 98.53% on the KLTCI dataset, and 97.57% on the MICC-F220 dataset) as compared with state-of-the-art CMFD methods in terms of the standard performance evaluation parameters of the CMF.

1. Introduction

In the digital revolution age, digital samples are a common means of information. The availability of powerful imaging tools like Photoshop and Corel Draw has enabled modifying the image content much easier without compromising quality. On a negative note, it paved the way for forgeries. The abusive use of image tampering has brought major security challenges. Thus, it is of utmost importance to verify the authenticity of the images. Therefore, it is a challenging task to verify an image’s originality accurately. Image validation is required in information sensitive departments like the military, media, court, medical appliances, banking, and news agencies. In CMF, a portion from the same image is taken and pasted to another location; significant characteristics, i.e., chrominance information, noise, light, and brightness variations, remain the same which makes the identification procedure a difficult task [1]. Different approaches like pixel, camera, format physical, and geometry-based are applied to identify image alteration. Because of the multidimensional behavior of image forgery problems, some methods can show better detection results. In contrast, others can no longer be useful, and it depends upon the forgery attack to which an image is subjected [1]. In [2], Zernike moments of circular blocks were utilized to analyze tempered images for locating copy-move regions. The Zernike moment-based features do not produce robust performance in the case of applying scaling operation on the blocks of the forged image during copy-move operation. There are also high computational costs associated with moment-based techniques. To reduce the dimensions of the feature vectors, principal component analysis (PCA) was applied to small-sized image blocks. However, it is not able to identify the transformations of rotation or scaling [3].

Keypoint-based approaches work to calculate features only for associated keypoints and improve computational efficiency as well as the effectiveness of CMF like keypoint-based methodologies, namely, scale-invariant feature transform (SIFT) and speeded-up robust features (SURF) [4]. These frameworks may not be successful in detecting flat duplicate regions. Figure 1 presents an example of CMF and splicing forgery attacks. A 9-dimensional vector was previously introduced to find the fixed angle rotation to detect image forgery [5]. However, this technique could not detect small copied regions in a forged image. Furthermore, a technique was proposed to detect the image features using a local binary pattern (LBP) [6], which exhibits robustness against blurring, flipping, and rotation. Still, its performance degraded while detecting the areas having general angles. A CMFD technique was proposed to identify duplicated regions using a special ordering of the wavelet coefficients from lower to higher frequency subbands. However, this technique was inefficient to defect forgery in compressed and noisy areas of an image [7]. Another technique was proposed to detect corresponding blocks’ similarity in a forged image using kernel PCA. However, this technique does not consider some geometric operations like scaling and shearing. Moreover, a technique was proposed to detect signal resampling using the expectation/maximization (EM) algorithm, which was helpful in image forgery detection. However, this technique limits evidence with the uncompressed and high-resolution format [8].

As a consequence of incredible image handling instruments, the forgeries in digital images effectively become a genuine social issue. This is done with the intent of highlighting a certain object or hiding or removing an object from the image and by applying different postprocessing attacks like blurring, color reduction, scaling, and rotation on images. By and large, a counterfeiter uses some relative changes to roll out the improvements outwardly flawless. Most existing copy-move recognition strategies are not compelling when duplicated locales are under mathematical twists. Unlike block-based approaches, which extract features from every block, the keypoint-based approach takes descriptors from selected places on the image. As a result of the reduced number of descriptors, a faster CMF detection approach based on keypoints can be designed. SIFT, SURF, binary robust invariant scalable keypoints (BRISK), and other feature descriptor methods are among the options. When a tampering operation is performed on an image, the actual pattern of the image is destroyed because it loses/changes the main meaning of the information carried within an image. Two detection methods have been defined; the active method and the passive method for CMF. In the active method, previous knowledge, idea, or concept of images is required to insert the information/image into the original image. For example, digital signs and the image watermark are used to identify duplicated areas of the image. The passive method is the simple method, it does not require any previous or basic information. The tampered or copied areas do not leave any real-time signs but it becomes very essential to know about the basic statistics of authentic images and resolve the inconsistent pattern in the image.

The following are the main contributions of the proposed CMFD method.

It computes LIOP features over circular blocks (instead of standard square blocks used by traditional CMFD methods) to generate the rotation-invariant feature vector. LIOP features present the more detailed image information by employing the direction-based relation of the central pixel to its neighborhoods. Therefore, it helps identify those features that are robust to postprocessing attacks of CMF (like scaling and rotation changes) and assist in effectively detecting the CMF.

It is proficient in identifying single and multiple CMF from digital images.

It also minimized the feature vector’s size for block representation which reduces the computational complexity for CMFD without utilizing any clustering and machine learning methods.

The remaining sections of this paper are structured as follows: state-of-the-art CMFD methods are reviewed in Section 2. In Section 3, the methodology of the proposed CMFD method is presented in detail. The experimental results are presented in Section 4. Finally, the conclusion along with future work directions is presented in Section 5.

The section presents details of the active and passive image region duplication detection techniques. Alkawaz et al. [9] applied discrete cosine transform with overlaying blocks to detect forgery regions in digital images. They also applied lexicographic sort for feature sorting. Bilal et al. [10] employed a fusion of speed-up robust features and binary robust invariant scalable keypoint features for image forgery areas detection. Huang et al. [11] extracted SIFT descriptors of an image. They achieved good detection results for many types of postprocessing, i.e., JPEG compression, size, and angle variations, and under the presence of noise attacks. But the method is not efficient to the low signal-to-noise ratio (SNR) and small-sized tampered areas. Bayram et al. [12] used counting bloom filters (CBF) and fluorescence molecular tomography (FMT) for CMFD. They used counting bloom filters (CBF) as an alternate to lexicographic sorting. They proposed FMT features for rotated, scaled, and highly compressed images. They achieved time efficiency by using CBF. However, robustness is reduced in that effort. Pan and Lyu [13] used a technique based on SIFT features of the image invariant to traditional sample transform like geometrical and illumination distortions. They used efficient approximation to achieve the correct representative set of features matching and transforming among copied areas. As their performance relies on identifying efficient SIFT features, it causes a limitation for the regions with few visual structures. Another limitation is for the images with intrinsically identical areas, which leads to an increase in a high false-positive rate (FPR). Bo et al. [14] introduced an effective approach for locating CMF based on SURF. It is used for detecting SURF interest points to find the possible copied areas in the suspected samples. This framework is efficient for image transformations and postprocessing attacks like noise, blurring, size, and angle variations; however, it is inefficient for automatically locating forged regions and their boundary. Guo et al. [15] presented a hybrid LBP scheme that matches globally rotation invariant with locally invariant LBP keypoints using local binary pattern variance (LBPV). LBPV is used to measure the directions of the texture image and the distinction between images. This method could decrease the feature dimensions while maintaining the classification accuracy better and better categorization certainty than existing rotation invariant LBP.

Hu moments were used by Liu et al. [16] to detect image forgery. By Gaussian pyramid decomposition, they acquire a low-frequency subimage. They used the overlapping circular subimage blocks and extracted from those circular blocks the Hu moments’ characteristics. As they computed Hu moments on the square’s inscribed circle, their method is robust to the rotation. So, the false alarm rate increases due to discarding the pixels outside of the inscribed circle. For image forgery detection, Bravo-Solorio and Nandi [17] made attempts based on log-polar maps. To obtain 1D rotation/reflection-invariant descriptors, they used log-polar maps. Then, to those descriptors, they mapped independent overlapping pixel blocks. Then, they mapped independent overlapping blocks of pixels to those descriptors. Their method is efficient in memory usage because of dimension reduction of blocks but is not much robust to high postprocessing like reflection, rotation, and scaling. By integrating discrete cosine transform (DCT) coefficients analysis, Lin and Wu [18] presented a technique for splicing and CMF localization. It blends them with double JPEG compression and SURF. To locate the copy-and-paste operations of sample portions, the former approach is used. To find duplicates of the same entity, SURF is used. This method can identify the forged regions effectively, identify the nonoriginal regions, and detect multiple artifacts. Shivakumar and Baboo [19] suggested detecting SURF and -dimensional tree-(KD-Tree-) based copy-move forgery. To find copied regions of various sizes, SURF is used. For multidimensional data matching, the latter procedure is used. In image distortion and scaling, this method works well but cannot detect tampered regions of a small scale. A tempered image detection approach based on improved DCT was introduced by Cao et al. [20]. They applied DCT and extracted four features from each block on the fixed-sized image blocks to minimize keypoint vectors’ size. For multiple CMF, Gaussian blurring, and noise pollution, this method is successful. It uses the function vector of a reduced dimension.

A passive forgery identification approach employing a dyadic wavelet transform (DWT) was suggested by Muhammad et al. [21]. This scheme’s input image is split into approximation (LL1) and subbands of detail (HH1). Matched pairs are obtained based on the resemblance between LL1 and HH1. For booming copy-move detection, this method offers the best approach to finding matched and unmatched portions among blocks of a sample. Kakar et al. [22] suggested a booming content-based image recovery technique to find image forgery based on the MPEG-7 image signature tools. Using a keypoint corresponding procedure that uses the critical constrictions in similar feature pairs distinguishes the copied parts. These techniques help us get high true-positive rates and low false positives for real and synthesized forgeries. Nevertheless, it is accurate only for separate samples and wide duplicated regions. Zhao and Guo [23] suggested a DCT- and SVD-based CMF identification methodology. To obtain quantized blocks further distributed into subblocks, the former technique is employed for fixed-size overlapping image blocks. SVD is implemented in each block, and its highest singular value is used to decrease the dimension as a feature. For a sample influenced by Gaussian blurring or JPEG compression, this technique may identify the forged regions. Lynch et al. [24] used comparison blocks as a dominant feature to minimize the processing cost of image forgery detection. As the dominant function, they used the average grey value of blocks. If their dominant feature varies greatly, a deep level comparison (made by a statistical hypothesis test) of the two blocks will not be made. For images influenced by JPEG compression, Gaussian blurring, etc., or in the case of lighter or darker duplicated areas, their technique worked well. Li et al. [25] used circular block characteristic vectors for CMFD. Feature vectors using rotation-invariant LBP were mined. Their system works well for typical postprocessing activities such as blurring or scaling and geometric transformations such as rotation or flipping. Their methodology, however, lacks the required range of feature dimensions and has less robustness for lower-angle rotations such as 200 or 300 degrees.

Hashmi et al. [26] introduced an effective method for CMF localization employing discrete wavelet transform (DWT) and SIFT features. The former method distributes the suspected sample into four parts. SIFT is used to find the key features. These key features are used to detect the resemblances between descriptor vectors, which helps find the tempered image. This framework identifies image tempering even when the image is scaled/rotated and then pasted. Li et al. [27] worked on the CMFD for the image subjected to affine transforms. They extracted rotation and scaling-invariant features from the image’s overlapping circular blocks using polar harmonic transform (PHT). Then, circular blocks were lexicographically sorted, and their Euclidean distances were compared to detect tampered regions. Their method performs well for lower rotation angles like 100 or 120 degree and general angles like 300 or 350 degree. In [28], a digital image forensic technique is proposed that focuses on the detection of image forgery based on the assessment of the light source for an image. This is an object-based method that estimates the lighting properties and hence detects the forgeries for various objects in the image. The model for digital forensics identifies the lighting discrepancies in the objects of an image and provides results indicating a difference between real and fake images. Lighting directions are estimated using azimuth and orientation parameters. The errors that appeared in the results are overcome by implementing the automatic selection of probes. This approach generates the 3D shape which is further used to detect the 3D lighting in the image for various objects. Finally, least-square optimization is used to improve the accuracy of this technique. Zandi et al. [29] presented an accommodative similarity threshold for a CMF identification approach based on locality-sensitive hashing (LSH). LSH is employed for finding the adjacent neighbors, which helps in detecting the forged image. The proposed technique could greatly decrease the number of false matches, improving computational cost and performance. Tian et al. [30] presented a method to locate the forensic changes from the digital images. Initially, the input sample was divided into overlapping blocks. The oriented FAST and rotated BRIEF (ORB) algorithm was employed over each block to compute the features. After that, the cosine and Jaccard distance metrics were used to measure the similarity between the computed keypoints. The approach in [30] exhibits better forgery detection accuracy which utilize ORB features and novel similarity measure technique; however, it may not perform well for the images showing huge scale variations. Kumar et al. [31] proposed a technique to correctly identify the forgery and imitations in the digital images. This technique works for images having any type of object present in the scene. By assessing the lighting parameters, this technique identifies the manipulated object and returns the angle of incidence concerning the light source direction. Two patches are taken from the original part while one patch is chosen from the fake part of the image. The differences between their angle values prove that the image is fake concerning lighting assessment in the scene. In this approach, if the angle difference is more or less than 10 degrees between the fake and original patch, then the patch is considered a forgery. If the angle is within this defined angle difference range, then image patches are considered consistent patches; otherwise, the patches are inconsistent and therefore belong to a forged part. The proposed method is tried on JPEG images browsed from online forged images or taken from well-known research datasets. The demonstrated results produce a robust forgery recognition rate on an image dataset comprising various types of manipulated images.

Abdel-Basset et al. [32] introduced a method to identify the image manipulations. Initially, the SIFT approach was applied to an input image to calculate the keypoints. Then, density-based spatial clustering technique was used to compute the clusters from detected features. Thirdly, RANSAC was utilized to remove the false clusters. Finally, the Structural Similarity Index (SSIM) metric was used to detect the matching areas. The method in [32] is robust to copy-move forgery detection; however, it may not detect the forgeries made within flat image areas. The method in [33] shows better manipulation detection accuracy, however, at the expense of huge computational complexity. Niyishaka and Bhagvati [34] introduced a framework to identify the alterations made within digital images. After preprocessing, the Laplacian of Gaussian (LoG) was applied to compute the blobs of the input sample. Then, BRISK features were computed from each blob, and Euclidian distance was computed among them to locate the matching areas. The method in [34] shows better performance than CMFD; however, for input samples with a large background area, its detection accuracy degrades badly. Soni et al. [35] presented a method to detect image forgeries. Initially, the input sample was distributed into nonoverlapping blocks on which the SURF algorithm was applied to detect the keypoints. Then, similar regions were localized. After this, the maximally stable extremal regions (MSER) technique was used to form the large blocks from each detected matched region. The SURF features were extracted from the identified MSER blocks. In the next step, the affine transform was used to remove outliers. Finally, the similarity was computed to locate forensic changes. The method in [35] performs well for CMFD; however, it may not show good performance under the presence of multiple CMF attacks in an image. Rani et al. [36] introduced the image manipulation detection approach to find image forgeries. To achieve this, a pixel-based forgery detection framework for copy-move and splicing-based forgeries is suggested in this paper. The preprocessing is performed on data to enhance textural information by applying enhanced SURF. Various features are estimated, and template matching is done for the identification of fake image regions. The relevant key parameters are estimated and compared with the calculated threshold value. The demonstrated results show that the proposed framework attains robust detection accuracy compared to the state-of-art image forgery detection techniques. In [37], a technique is proposed for multiple light source-based forgery detection for heterogeneous image surfaces, nearby surface geometry, and texture information to assess the lighting environment. The Phong reflection model is implemented to estimate the structural profiles. The method works in two phases. In the first phase, preprocessing over selected patches is performed followed by angle and error estimations in the second phase. To identify such forgeries, elevation angles concerning mounted light sources are estimated. The value of the elevation angle is computed for various patches of the image. This method is also validated for synthetic image datasets and tested for generalized forged images. An experimental result demonstrates better forgery detection accuracy compared with state-of-the-art methods in this domain. [38] proposed a technique to detect image manipulation. By using a single light source, the incident light source angle is returned which is achieved by estimating illumination from the image. This technique is validated by computing the angle for all objects in the scene. The incident and reflection angle for all the image objects should be consistent; however, if these conditions are violated then, objects are not consistent concerning light source direction. Errors are approximated by applying a least-square estimation. The technique shows that from subpatch estimations, the calculation of light source direction can be approximated which itself is unique to detecting image forgery. The obtained differences in the estimated values of angle above a threshold angle are considered as a fake object w.r.t light source, and therefore, the image is taken as phony. Finally, the results demonstrate that the identification of forged parts is successfully done for light-based image forgery detection. Kumar et al. [39] presented a forgery detection technique based on estimations of multiple light source directions. This method uses a pixel patch from the image region to estimate the source light vector. The implementation is done for images where one and more light sources are available in the scene. This technique can identify image forgery in terms of elevation angle obtained from a source of light and surface normal. This technique is tested for both outdoor and indoor images under certain known parameters. The novelty of this technique is that photo manipulation detection is done using a multiple-light source detection. It produces robust performance as compared with state-of-the-art image forensic techniques by making certain assumptions about surface properties and illumination parameters.

Javed and Jalil [40] presented a novel approach for detecting suspicious images using deep learning. The technique transforms the image into the byte level to identify the forgery in real-time. Zhang et al. [41] utilized the joint probability density matrix [JPDM]; the technique after detecting the input image as forged uses JPDM to correlate within the discrete coefficients transform [DCT]. Islam et al. [33] presented a deep learning-based framework, namely, generative adversarial network (GAN) to show the forensic changes made within digital images. Initially, VGG-19 architecture was used to measure the deep features of a suspected sample. Then, two atrous spatial pyramid pooling (ASPP) operations were used to compute both contextual and cooccurrence keypoints, which were later combined to pass to the detection branch. Agarwal and Verma [42] presented a deep learning-based technique, namely, VGGNet along with the adaptive patch matching (APM) method was applied for CMFD. This approach is robust to various image transformation attacks like blurring and compression; however, the technique is suffering from high computational costs. Table 1 presents the critical analysis of the state-of-the-art CMFD methods.

3. Methodology

A detailed description of the proposed CMFD method for the detection of copy-move forgery is presented in this section. Firstly, the color image is converted to grayscale. Secondly, the image is distributed into circular blocks of pixels. In the third step, features are extracted from each circular block of the image by applying the LIOP descriptor [48]. In the fourth step, a comparison of blocks’ features is performed by applying Euclidean distance to highlight the forged regions of the image. Postprocessing is performed as the last step to make the detection results clearer. A visual presentation of the proposed CMFD method is exhibited in Figure 2.

3.1. Formulation of the Problem

In CMF, similarities of the tampered regions are always small against a predefined threshold. So, for tampered grayscale images, CMFD focuses on identifying two nonintersecting regions that are hole-less and have larger similarities.

For an image represented by , the forged image depends on the regions that is mathematically defined as follows: where where , a metric norm, represents the distance between source regions and destination regions of the forged region. The must be less than the similarity threshold , whereas is the translation vector, and is the corresponding threshold. Therefore,

Due to the postprocessing attacks on the images, detecting forgery while handling rotation and scaling attacks can be harder for rotation and scale-invariant descriptors. Hence, the proposed CMFD method also incorporates the mechanism for handling rotation and scale variances within the image. The postprocessing attacks of scaling and rotation on the forged image can be mathematically modeled using the following equations that serve as an image forgery model: where and denote the coordinates of the duplicated area of the image; , , denote the angle of rotation, a scaling factor, and flipped area of the forged image, respectively. Forger applies postprocessing attacks over an entire image like blurring and noise, to conceal the forgery effects, making it difficult to detect forgery. Hence, the proposed CMFD method aim is to detect the duplicated regions and to highlight the duplicated contents in the existence of various postprocessing operations.

3.2. Preprocessing and Division of the Image into Circular Blocks

After visually analyzing duplicated regions in the illumination domain, a standard color space conversion method is employed by the proposed CMFD method to convert each RGB forged image to a grayscale forged image. At first, image key points highlighting the distinct information of image content are located, and then, the corresponding feature vectors are captured. A feature vector is a collection of image statistics obtained from the keypoints’ local neighborhood. For key points and features to be effective, they should capture distinct positions in an image, and be robust against local geometric distortions, noise, illumination variations, etc. The proposed methodology utilizes LIOP descriptor for keypoint detection and description. For the sake of distinction, the keypoint-based techniques divide the image into subparts, and features are collected for each subpart separately, which are joined together at the end. Most of the existing methods use a pixel-based approach to extract image features. For example, SIFT used a grid as the feature extraction source region and 3 to 8 bins in one direction. The other direction, respectively, is utilized by gradient location-orientation histogram (GLOH) as a log-polar grid. These existing methods require a consistent orientation for each subpart of the image and descriptor construction per that specified orientation to make the descriptor rotation invariant. So, the accuracy and efficiency of these methods depend on the selection of subparts orientation. Due to these limitations, these methods are not robust to detect forgeries in postprocessed images. To avoid the orientation estimation for rotation-invariant features, the spin image technique is used, which further divides the image subparts into five circles/rings. However, its discriminative power is low, as it splits the image in one direction, i.e., radial direction, and is unable to handle angular direction. In the proposed CMFD method, the input grayscale image (forged image) is transformed into the circular block using the polar coordinate system that is defined using the following mathematical equation: where is the origin of the circle, is the radius, and the is the axis of the circle. As compared to grid-type region decomposition methods, it has low discriminative power. In grid-type region decomposition, all local pixels of an image subpart are sorted as per their intensities in ascending order. Then, subparts are further decomposed into ordinal bins as per their order of intensities. This region decomposition is invariant to postprocessing attacks like rotation and contains more information than ring-type region decomposition. The circular blocks are fed into the feature extraction step of the next section.

3.3. Local Intensity Order Pattern-Based Features

The local information of subparts of the image varies depending upon the methods used for feature description. For example, the spin image method develops a histogram based on the intensity change of local subparts of the image while SIFT and GLOH populate histograms based upon the gradient’s orientation. More recent methods focus on more robust features of local regions of the image, like LBP-based methods used histogram of local binary patterns which are centrally symmetric, and local ternary patterns (LTP) are also used to populate histogram which is a step ahead of LBP. LBP and LTP use the intensities of sample points that are centrally symmetric, so do not consider the relationship of sample points in the neighborhood. They also require locally consistent neighboring points in orientation to make rotation invariance better. However, these methods make them vulnerable to errors in orientation estimation. Keeping these facts into view, the local intensity order pattern (LIOP) can be used to overcome these limitations of existing methods for effective feature representation of the image contents. The proposed CMFD method uses LIOP features for the salient objects of the forged image which uses the intensity order of neighboring sampled points as the local information. It is robust to the rotation because its sampling is rotation-invariant and maps the consistent local orientation. Due to all these reasons, its expected discriminative power is high as compared with existing sparse representation-based methods like SIFT and SURF.

3.4. The Formation of LIOP Feature Vectors

In this step, the proposed CMFD method computed the LIOP descriptors from detected keypoints that are stored in ordinal bins and formed the feature descriptor from each circular block of the image. Mathematically, the LIOP descriptor can be defined as follows: where is the weighting function, which, upon any changes in rotation and monotonic intensity, increases the invariance of LIOP descriptors, and is the LIOP descriptor of a circular block of point (origin of a circular block). A LIOP point with a more distinct neighborhood is given a larger weight to make the descriptor more robust. The weighting function is mathematically defined as follows: where is the sign function and counts the distinct sample pairs and measures the intensity variations of neighboring sample points for a circular block of point . Let be an -dimensional vector of point in the local patch, with the intensities of neighboring sample points of point . The is mathematically expressed as follows: where and represents the intensity of neighboring sample point . The local patch is distributed into partitions where a LIOP represents every partition. The point has neighboring sample points, which are distributed equally along a circle having a radius . For creating rotation invariant samples, the first point is sampled to point from the local patch’s center along the radial direction, and the farther point out of the two radial direction points is chosen as the starting sample point. The points that are remaining are inspected in an anticlockwise way. The four neighboring points of , i.e., , , , and , remain consistent in all the patches that are rotated, i.e., , , , and , respectively. Due to the one-to-one relationship between the subset and the permutation, all the probable permutations in listed in an index table can be made by encoding BM’s subsets. To map -dimensional feature vector with permutation , a function is formulated over the index table; all elements are 0 besides the element, which is Equation (6). The is mathematically formulated as: where is the index of in the index table and . The extracted circular block-based features of each circular region of the image are now ready to get matched in the next step.

3.5. Feature Matching

As mentioned earlier, CMFD focuses on identifying duplicated regions that are (a) nonintersecting and (b) have a similarity count less than a defined threshold. To show the copied regions from the given query image, the resemblance among each circular block-based feature is computed using the Euclidean distance formula that is mathematically defined as follows: where , , , and are indicating the locations of the pixels of keypoints from two blocks. is a defined threshold with a global value of 0.7 for the proposed CMFD method. Therefore, the data points from the two circular block-based features are said to be similar if the distance among them is less than 0.6; otherwise, they are declared as unforged. The threshold value has a substantial impact on the matching process, as choosing a very small value of the threshold results in an increased rate of false matches, while using a large threshold value in the CMF process causes to eliminate the detection of forged areas.

The complete algorithm of the region duplication detection from digital images using the proposed CMFD method is as follows:

ImgIn ➔ The input image matrix
ImgGray ➔ Grayscale image
ImgBlocks ➔ A two-dimensional matrix where each block is represented by pixels
ImgFeatures ➔ Matrix representing feature space
EDist ➔ Euclidean distance between neighborhood features
D ➔ Region marker based on the threshold
ImgReconst ➔ Reconstructed image
Thresh ➔ Threshold determining if the Euclidean distance between feature spaces could mark the region
Step 1: ImgGray ← Grayscale(ImgIn) //Convert the RGB image into grayscale image
Step 2: ImgBlocks ← Divide(ImgGray,8) //divide the image into [] circular blocks
Step 3: FOREACH block in ImgBlocks
    ImgFeatures ← Extract LIOP features(block) //extract LIOP features from each circulate block
    END
Step 4: For : Size(ImgFeatures)-1
     EDist ← Euclidean(ImgFeatures[i], ImgFeatures[i+1])
      IF(EDist <= Thresh)
      D[i]=1
      ELSE
      D[i]=0
      END
    END
Step 5: ImgReconst ← Postprocessor(D)

4. Evaluation Measures, Experimental Results, and Discussions

4.1. Performance Evaluation Measures

For the proposed CMFD method, assessment measures, i.e., precision (), recall (), and -measure, are utilized to measure and compare the detection results of the proposed CMFD method with ground-truth images. Precision is the fraction of the percentage of predicted forged samples that are forged. (i.e., locate samples that are already manipulated). A recall is the fraction of the percentage of actual forged samples that are properly predicted forged (i.e., the number of returned samples being identified as manipulated from all the forged). -measure is a composite metric that is used to measure the accuracy of the proposed CMFD method as it employs both and parameters. These performance parameters are mathematically defined in the following equations. where indicates the total samples, which are accurately detected as forged, indicates the total misclassified samples as forged, and indicates the proportion of samples wrongly classified as original and forged.

4.2. Experimental Requirements and Results

The performance of the proposed method is evaluated using three standard datasets of the CMF that are CoMoFoD, KLTCI, and MICC-F220. The experiments are computed on a computer with the following hardware and software specifications; Microsoft Windows 8.1 (64-bit), MATLAB 2017b (64-bit), Intel Core i7 processor with 2.4 GHz, 500 GB hard disk drive, and 8 GB of RAM.

4.2.1. Performance Evaluation on Copy-Move Forgery Detection (CoMoFoD) Dataset

The CoMoFoD dataset contains 260 forged image sets in two categories [49]. The first category contains image sets, each with a dimension of pixels, and the second category contains 60 image sets, each with a dimension of pixels. These images have different forgery attacks like translation, splicing, rotation, distortion, and scaling. Different types of postprocessing methods are applied to all tempered and genuine samples, such as JPEG compression, noise adding, blurring, and color reduction. The experimental details of the proposed method on this dataset are reported in the following subsequent sections.

(1) Multiple CMFD Demonstrations. In this section, experimental details of the proposed CMFD method are analyzed by tampering with multiple forged areas of the image on the CoMoFoD dataset. The performance analysis of the proposed method in terms of visual results is presented in Figure 3. Table 2 presents its experimental details on different images of the CoMoFoD dataset in terms of performance parameters that are precision, recall, and -measure. After analyzing experimental details in this section, it can be concluded that the proposed CMFD method produces robust performance on the CoMoFoD dataset in the case of multiple forged areas of the image.

(2) Postprocessing Attack Demonstration. Postprocessing attacks on images include blurring, additive noise, compression, variations in brightness levels, color reduction, contrast adjustment, and rotational and scale invariance. Table 3 illustrates that the proposed CMFD method performs remarkably even in the presence of postprocessing attacks on the images of the CoMoFoD dataset. The visual results of the proposed CMFD method are shown in Figures 4 and 5 after applying different postprocessing attacks on the images of the CoMoFoD dataset, which also proves the robust performance of the proposed CMFD method.

The robustness of the proposed CMFD method is also tested by comparing its performance with state-of-the-art forgery detection methods. Table 4 presents experimental details in terms of precision, recall, and -measure parameters of the proposed CMFD method and its comparison with competitive forgery detection methods. After analyzing experimental details in Table 4, it can be concluded that the proposed CMFD method also produces robust performance as compared with its competitive methods.

4.2.2. Performance Evaluation on the Kodak Lossless True-Color Image (KLTCI) Dataset

The KLTCI dataset [52] contains 24 photographic-quality images of different subjects under various lighting conditions and various locations; the images are high-dimension with a resolution of pixels. The images of this dataset comprise several backgrounds and exterior conditions such as sea, sky, walls, and buildings. By using Adobe Photoshop, tempered images for the KLTCI dataset are manipulated. The various combinations of postprocessing attacks, i.e., translation, blurring, and rotation, along with CMF, are used to construct forged images. Copy-move regions in this dataset possess an approximate dimension that varies from pixels to pixels. Table 5 presents the performance analysis of the proposed CMFD method on the KLTCI dataset in the occurrence of several postprocessing attacks. After reviewing the experimental information provided in Table 5, it can be concluded that in the presence of various postprocessing attacks on the tampered images, the proposed method also exhibits a robust performance on the KLTCI dataset.

The visual results of the proposed CMFD method on different sample images of the KLTCI dataset are shown in Figure 6, which also proves its robust performance on this dataset. In Figure 6, the first to fourth columns show the original images, tampered images, detection results of the proposed CMFD method, and ground-truth images, respectively.

4.2.3. Performance Evaluation on the MICC-F220 Dataset

The MICC-F220 [53] dataset contains 220 images. The dataset contains a tampered image using a scale and rotational transformation. The tampered images included in this dataset have different sizes ranging from pixels to pixels. The analysis of the proposed CMFD method is evaluated using the MICC-F220 dataset, and its experimental results are compared with state-of-the-art CMFD methods that are Manu and Mehtre [54], Cozzolino et al. [51], and Thampi et al. [55]. The visual results of the proposed CMFD method conducted on the MICC-F220 dataset are shown in Figure 7. The comparison detail of the proposed CMFD method in terms of performance evaluation parameters is presented in Table 6. The experimental analysis in terms of visual and evaluation parameters on this dataset also proves the robustness of the proposed CMFD method.

4.2.4. Comparison of Computational Complexity

The performance analysis of the proposed CMFD method is also performed by considering its computational complexity as compared with competitor CMFD methods. Its computational complexity is reported on a desktop computer equipped with the following hardware and software resources: CPU: Intel Core [email protected] GHz, RAM: 8 GB, hard disk: 500 GB, Windows 10 operating system (64 bit), MATLAB 2017B (64 bit), VLFeat version 0.9.21-MATLAB library. Table 7 presents details of the computational complexity (time in seconds) of the proposed CMFD method and its comparison with competitor CMFD methods on the MICC-F220 dataset.

5. Conclusion and Future Directions

In this article, a novel method of copy-move forgery detection from digital images is presented. The proposed CMFD method divides the image into circular blocks and computes a rotation-invariant feature vector from each circular block. The similarity between each feature vector is computed by applying a Euclidean distance to detect forged image areas and to remove false matches. The LIOP descriptor assists the proposed CMFD method in effectively detecting the tampered areas of the image even in the presence of different postprocessing attacks and also enables it to effectively detect the single and multiple CMF areas of the image. The LIOP descriptor also assists the proposed CMF method to detect the tampered regions more efficiently as compared with SIFT and SURF features. The proposed CMFD method produces promising results on the CoMoFoD, KLTCI, and MICC-F220 datasets as compared to the state-of-the-art CMFD methods. There is room for improvement of the proposed CMFD method to detect small and extremely smooth, or blurred regions from the tampered images. The performance of the proposed CMFD method can also be analyzed by considering the deep learning-based method for CMF as well as other types of forgeries can be considered like image splicing.

Data Availability

The authors have used publicly available data to support the findings of this study that is included within the article.

Ethical Approval

This article does not contain any studies with human participants or animals performed by any of the authors.

Conflicts of Interest

All the authors declare no conflict of interest.

Authors’ Contributions

Each author has worked equally.

Acknowledgments

This research work was carried out as a joint research collaboration at the University of Engineering and Technology, Taxila, Pakistan; Politecnico di Milano, Milano; Sejong University, Seoul, South Korea; Captial University of Science and Technology (CUST), Islamabad; and University of Jeddah, KSA. The authors are thankful for this support.