Next Article in Journal
Towards Automated 3D Inspection of Water Leakages in Shield Tunnel Linings Using Mobile Laser Scanning Data
Previous Article in Journal
Toward Sub-Diffraction Imaging of Single-DNA Molecule Sensors Based on Stochastic Switching Localization Microscopy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Forgery Detection and Localization via a Reliability Fusion Map

1
School of Cyberspace, Hangzhou Dianzi University, Hangzhou 310018, China
2
Institute of Cyberspace Research, Zhejiang University, Hangzhou 310027, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(22), 6668; https://doi.org/10.3390/s20226668
Submission received: 22 October 2020 / Revised: 18 November 2020 / Accepted: 19 November 2020 / Published: 21 November 2020
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Moving away from hand-crafted feature extraction, the use of data-driven convolution neural network (CNN)-based algorithms facilitates the realization of end-to-end automated forgery detection in multimedia forensics. On the basis of fingerprints acquired by images from different camera models, the goal of this paper is to design an effective detector capable of completing image forgery detection and localization. Specifically, relying on the designed constant high-pass filter, we first establish a well-performing CNN architecture to adaptively and automatically extract characteristics, and design a reliability fusion map (RFM) to improve localization resolution, and tamper detection accuracy. The extensive results from our empirical experiments demonstrate the effectiveness of our proposed RFM-based detector, and its better performance than other competing approaches.

1. Introduction

As digital and other communications technologies advance, digital images, videos and audio files can be conveniently acquired from various devices, ranging from the conventional closed-circuit television cameras (CCTVs), digital cameras to other Internet of Things (IoT) devices with image, video and audio capturing capabilities (e.g., Ring Doorbell Camera). Modifying an image has also become easier, due to the availability of inexpensive image, video and audio (collectively referred to as multimedia) editing software. Implications of forged multimedia files, for example using re-sampling [1,2] or copy-moving [3,4], include ownership infringement or fraudulent activities. For example, as recently as Sep 2019, “the CEO of an unnamed UK-based energy firm believed he was on the phone with his boss, the chief executive of the firm’s German parent company, when he followed the orders to immediately transfer €220,000 (approx. $243,000) to the bank account of a Hungarian supplier” (https://www.forbes.com/sites/jessedamiani/2019/09/03/a-voice-deepfake-was-used-to-scam-a-ceo-out-of-243000/). This necessitates the need to design an effective and robust forensic detector with the capability of providing reliable digital evidence.
The study of both source identification and tampering detection is a relatively mature topic [5,6,7] for details. Image tampering detection targets processing techniques, such as object removing or adding. Object forgery detection approaches can be divided into three classes: (i) splicing detection: given two images, one can detect if a region of a source image has been spliced into a target image [8,9,10,11,12,13]; (ii) copy-moving forgery detection: given an image, one can identify if an object is copied-and-pasted from one to another location [14,15,16,17]; and (iii) object removal detection: given an image, one can detect if an object of the source image has been removed [18,19,20].
There has been a recent trend of moving away from conventional hand-crafted feature extraction to using convolution neural network (CNN)-based extractors. However, some primitive CNN-based forensic detectors are generally not practical for a number of reasons, for example in terms of the robustness of feature extraction, and the resolution of tampering localization. Therefore, there have been efforts to design a pre-processing layer to enhance the robustness of feature extraction [21,22,23], and fusing multiple detectors based possibility maps [24] and single CNN-based reliability maps [25,26] to improve the resolution of tampering localization.
There still remain several limitations in the aforementioned approaches. First, most existing pixel-wise tampering detectors adopt an independent patch-based strategy rather than using the correlated information among patches. This results in insufficient statistical information required for feature extraction, especially on the edge of a forged region. In other words, we should emphasize on neighbor patches’ characteristics to facilitate the determination of the authenticity of an inquiry patch (a principle we consider in this work). Furthermore, the absence of statistical characteristics over flat areas (clear sky, blue ocean, etc.) results in estimation ambiguity, and results in degraded detection performance. In that case, the texture of the image content becomes a decisive factor for enhancing detection accuracy. Besides, with the rapid development of image-editing software, the remnants left by manipulation operation have a behavior similar to its pristine version (i.e., tampering traces are hard to detect). Therefore, how to reduce the probability of detection mismatch and improve the resolution of localization (controlled by the smallest unit of detection) remains an open problem.
To address that challenge, in this paper, we propose a novel end-to-end framework to improve the accuracy of tampering detection and localization, mainly for composite images edited from different imaging sources. The main idea behind the proposed method is that camera model-related artifacts can be successfully extracted from a typical image acquisition pipeline, leading to that our proposed reliability fusion map (RFM)-based detector can capture subtle manipulation traces (see Figure 9 for illustration). By designing a pre-processing module, together with a feature extraction module containing CNN module equipped with content-texture module, a feature vector with initial detection (Figure 10d) is effectively generated. More importantly, we design a reliability fusion map (RFM) to improve the localization resolution (Figure 10e). The effectiveness of our proposed method (The source code is available on Github: https://github.com/grasses/Tampering-Detection-and-Localization) is experimentally verified compared with the prior arts [23,26].
The remainder of this paper is organized as follows. Section 2 reviews the related literature. In Section 3, we describe our proposed framework, consisting of a pre-processing stage (high-pass filter), a feature extraction stage (CNN module equipped with content-texture module), and a reliability fusion stage (binary map RFM). Section 4 presents the numerical results over the benchmark dataset, and a comparative performance evaluation. Finally, Section 5 concludes this paper.

2. State of the Art

A generic framework of tampering detection usually contains the following steps: pre-processing, feature extraction, and post-processing (see Figure 1). In general, low-level features are extracted in Stage 1; high-level features are extracted in Stage 2; Stage 3 plays a critical role in tampering detection and localization, that we mainly focus on in this paper. Next, let us generally review the relevant literature based on these three stages.

2.1. Pre-Processing Based Algorithms

Image pre-processing efforts have generally been put on how to manually design efficient constant convolution kernels, and meanwhile to train an effective feature extractor of capturing characteristics related to tampering traces. For instance, the research community has proposed constant filters to suppress the interference caused by edges and textures, and enhance the intrinsic features, such as using the median filter residual (MFR) [27], guided filtering for photo response non-uniformity noise (PRNU) [28], resampling detectors [29,30] and other forensic detectors based on steganalytic features like spatial rich model (SRM) [31]. It should be noted that the constant filter is good at accelerating convergence of a neural network, since the residual image obtained from a constant filter is content-independent.
Inspired by the aforementioned effective high-pass filter, some researchers utilized a pre-determined predictor to produce a series of residual pixels. Then, these residual pixels are exploited as low-level forensic features. High-level associations are formed by subsequent detection. For instance, Bayar and Stamm [22] combined a constant filter with a trainable convolutional filter in the pre-processing stage to enhance the robustness of detection. Subsequently, they used a new type of CNN layer (referred to as the constrained convolutional layer) for designing a universal detector [23]. Although this approach [23] reportedly achieved high detection accuracy, its theoretical performance for image tampering localization is still unknown. Moreover, each isolated patch-wise detection result is hardly analyzed together, leading to that the mismatched results of detection to some extent decrease the resolution of tampering localization (see Figure 9). However, in this paper, due to our proposed RFM algorithm, that limitation can be perfectly overcome.

2.2. Feature Extraction Based Algorithms

A number of feature extraction techniques have been proposed, such as those designed to distinguish camera fingerprints, leading to detection of camera model based tampered images. Ref. [32] proposed a CNN module to extract a noise residual, called noiseprint, which largely suppressed the scene content and enhanced camera model-related artifacts. Despite the promising results shown in [32], one has to keep in mind that the noiseprint can only be useful for camera model identification, but not for individual device identification. A large scale of feature extraction techniques leveraged other artifacts inherited in an image. By utilizing the information of chroma and saturation, Ref. [33] designed a Shallow Convolutional Neural Network (SCNN) to detect and localize the traces of low resolution tampered images. Ref. [34] investigated the features of manipulation especially artifacts near boundaries of manipulated regions. Then they proposed an encoder-decoder based network to exploit these traces. Some prior arts focused on designing the architecture of neural network to improve the manner of learning process and strengthen the effectiveness of feature extraction. Inspired by the mechanism of memory in human brain, Ref. [35] proposed a Ringed Residual U-Net (RRU-Net) to accelerate the convergence of the neural network. The RRU-Net was efficient in exploring the differences of image attributes between the pristine and tampered regions by using the contextual spatial information in an image. Ref. [36] proposed a densely connected CNN module to increase variations in the input of subsequent layers. The dense connectivity, which had better parameter efficiency than the traditional pattern, ensured the maximum information flow between layers in the network. Next, we will revisit some of the strategies proposed to improve resolution of tampering localization using high-level features.

2.3. Post-Processing Based Algorithms

In the stage of post-processing, one can utilize high-level features to obtain better localization resolution. The problem of tampering localization requires one to accurately specify forged region by minimizing the probability of patch-wise detection mismatch. In fact, tampering localization in a forged image is more difficult than merely binary classification between pristine and forged one.
Many prior works leveraged distinctive artifacts inherited in an image, for instance, based on sensor pattern noise [25,28,37], JPEG attributes [38,39], multiple techniques fusion [40,41,42,43]. Similar, the authors of [24] combined two existing forensic approaches (i.e., statistical feature-based and copy-moving forgery detectors) to obtain the tampering possibility map. Although such a method can deal with various manipulations, its usage in real-time scenario is limited due to its 18,157-dimensional high-level features.
CNN-based methods often employed one feature extractor coupled with confidence factors for detection. For instance, in [25], a two-tiered transfer learning-based approach was proposed for patch reliability estimation using camera model attribution, which achieved performance improvement in one single patch. However, the approach did not consider reliability of adjacent patches, and its theoretical performance on the whole image remains unknown. To mitigate the limitations, the authors in [26] used step-by-step clustering of camera-based CNN features. However, the localization resolution still needs to be improved. In addition, due to the extensive dependence of group constrained thresholds for filtering out nuisance noise, its robustness remains to be verified.
Existing approaches mainly focus on the generalized three stages in order to improve the performance of tampering detection and localization. During pre-processing stage, one accelerates the convergence of neural network and improves performance of feature extraction. In feature extraction stage, one utilizes an effective CNN to extract features characterizing tampering traces. In the post-processing stage, one reduces the mismatch result of detection, and improves the resolution of localization. It makes sense that different approaches have their unique advantages and limitations. Therefore, how to leverage the advantages of current arts for improving the accuracy of both detection and localization remains an ongoing challenge. In the following section, dependent of the powerful CNN, we will specifically present the design of an efficient RFM-based detector.

3. Proposed Method

The core idea behind our proposed method is that both tampering detection and localization are based on fingerprint discrimination among different camera models. Our proposed RFM-based detector is described below (see Figure 2): (i) pre-processing: we utilize a fixed high-pass filter to obtain full-size residual image, and then split the residual image into a set of 64 × 64 overlapped patches with stride of 32; (ii) feature extraction: we design the CNN module equipped with content-texture module, including each component for designing convolutional layer, fully-connected layer, and classification layer; (iii) reliability fusing: three significant factors are proposed to establish the binary map RFM for detecting tampered image and localizing forged region.

3.1. Pre-Processing

Let us assume that a pristine image is captured by an imaging device while its forged region is obtained from another. In order to remove interference from image content, a high-pass filter (see Equation (1)) formulated as:
F 0 = 1 12 1 2 2 2 1 2 6 8 6 2 2 8 12 8 2 2 6 8 6 2 1 2 2 2 1
is used in the stage of pre-processing to extract a residual image of each inquiry image. We remark that the high-pass filter is efficient in accelerating convergence of neural network, and its performance has been verified in [44,45,46]. Subsequently, it is proposed to split the residual image I into 64 × 64 patches. All patches from a pristine image are captured by the same camera. On the contrary, patches from a forged image contain more than one fingerprint generated by different cameras. Then, we define P i , j as the extracted patch, and i { 0 , N 1 1 } , j { 0 , N 2 1 } , N 1 × N 2 denotes the total number of patches extracted from I (see Figure 2).

3.2. Feature Extraction

The establishment of the proposed feature extraction involves two main stages, namely the CNN module and the content-texture module where texture quality is designed to quantify the perceived texture of each patch. In fact, it is worth noting that our proposed CNN module deals with the patch as the smallest calculation unit.

3.2.1. CNN Module

A typical CNN module consists of stacked convolutional layers, and fully connected layers, followed by a softmax classifier (or classification layer) (see Figure 3 and Table 1 for details). The stacked convolutional layers can be defined as follows:
f n P i , j = f pooling f activation f n 1 P i , j w n + b n ,
where a patch P i , j is fed into our CNN module, “*” means the convolution operation, f n · denotes an output of the n t h convolutional layer, and w n and b n are shared weights and bias parameter. f pooling · represents a pooling layer, which controls the representation dimension by reducing the amount of parameters and computation in the CNN module. It avoids the problem of overfitting. f activation · represents an activation function, aiming at activating effective units while suppressing invalid units.
Next, fully connected layers featured by the network parameters play an important role in the establishment of classification layer. The fully connected layer feeds the features, that are extracted from the convolutional layer, back to a typical softmax classifier. It is worth noting that each output of the node from the softmax classifier is a probability, serving as the discriminative factor for our classification. In the stage of backpropagation, the cross-entropy error function (namely loss funtion) is used to measure the distance between probability for each classification and original distribution, which can be defined as follows:
argmin Θ L y , y ^ ; Θ = i N y i × log y ^ i ,
where y ^ i denotes the probability for i-th classification; Θ represents the parameters of neural network. By minimizing the objective function L , the parameters of neural network is refined with Stochastic Gradient Descent (SGD) automatically. It should be noted that the goal of loss function in this paper is to discriminate among different camera models.
In this paper, we adopt the CNN architecture similar to our prior work [47]. Since the input data, referring to as patches, are not very large, the neural network should be good at analyzing difference between the pixel and its neighboring counterparts, and have a strong predictive ability to characterize feature maps. In general, a too-wide network architecture cannot fully learn the feature map; a too-deep network architecture might cause increment of the computational complexity. Hence, our proposed network is neither too deep nor too wide. In this context, we mainly focus on the design of fusion map for splicing detection and localization, but not for specific description of CNN module (the readers may refer to [47] for details).
Different our previous work [47] mainly analyzing the image features characterizing different source camera models, in this paper, we adopt a CNN architecture equipped with content-texture module, and leverage a reliability fusion map to refine extracted features for dealing with the problem of tampering detection and localization.

3.2.2. Content-Texture Module

When dealing with a low texture patch, the performance of the CNN module should be further enhanced. Inspired by the algorithm proposed in [37], we use the texture quality measure standard to define a patch texture, formulated as follows:
Q = 1 3 c R , G , B [ α × β ( μ c μ c 2 ) + ( 1 α ) ( 1 e γ σ c ) ] ,
where three parameters α , β and γ are used to assign the weights into μ c μ c 2 and 1 e γ σ c . μ c and σ c , c { R , G , B } respectively denote the mean and standard deviation of P i , j for each color channel. In our experiment, α = 0.7 , β = 4 and γ = l n ( 0.01 ) . Q i , j for each patch is normalized into the range 0 , 1 . As a decisive factor, texture quality suppresses ambiguous classification of CNN over the low-texture regions while further enhancing prediction accuracy in high-texture regions, leading to decreasing the mismatch of classifications.

3.3. Reliability Fusing

One cannot guarantee that all regions contain adequate statistical information for tampering localization, especially dealing with low-texture regions. In addition, the output result from our CNN module contains the probability vector for each camera model, meaning that it is more than just a binary (true or false) classification. The detection result of the adjacent patches may influence that of the central inspected patch. For instance, if the result of the patch generated by the CNN module has the large probability as a tampering sample while the results of its adjacent neighbors as pristine, it is reasonable that the probability of detection mismatch has increased. To achieve improvement in detection and localization accuracy, the reliability-fusing operation is thus proposed in this context. For clarity, we illustrate an example of the proposed RFM algorithm (see Figure 4). Let us give the specific description of RFM algorithm, involving three following factors:
  • Patch texture Q i , j . The parameter Q i , j can provide information about content texture of inquiry patch, which tends to be low for flat patches and high for patches with high variance. Since CNN module cannot perform in low-texture regions as well as in high-texture regions, let us accordingly decrease CNN confidence F i , j in low-texture regions.
  • CNN confidence F i , j . F i , j represents the output result of the CNN module extracted from P i , j , among which sum of all vectors equals to 1. Rather than truncating confidence F i , j by an empirical threshold, our proposed algorithm combines the CNN confidence for each patch, meaning that the algorithm accumulates the CNN confidence F i , j of adjacent patches around the inspected (or central) patch.
  • Density distribution ρ i , j . ρ i , j represents a tampering ratio of K adjacent patches. ρ i , j is proposed to remove the mismatched results generated by the CNN confidence F i , j . The larger ρ i , j indicates the more forged adjacent patches around the inspected patch.
Next, we will extend the specific reliability fusing procedure (RFM algorithm) to obtain the binary map RFM.

3.3.1. Fusing Q i , j and F i , j

Relying on Q i , j , overlapped adjacent patches, referring to ( P 0 , 0 , P 0 , 1 , P 1 , 0 , and P 1 , 1 ), jointly re-identify the central patch. Therefore, half of detection unit size with 32 × 32 is reduced (see Figure 4a), compared with the general clustering algorithm with 64 × 64 (see Figure 4b). Then, the formula is defined as follows:
R ^ i , j = a = 0 1 b = 0 1 Q i + a , j + b Q × F i + a , j + b ,
where F i + a , j + b represents the CNN confidence, and Q i + a , j + b is the adjacent patch texture. R ^ i , j denotes the reliability vector of the fused central patch P i , j , which is a re-estimation of the CNN confidence for four adjacent patches (see Figure 4a), relying on the assigned weights generated by Q. The reason why we choose four adjacent neighbors rather than only one used in existing methods such as [26] is twofold: (1) if only one nearest neighbor is considered, the localization accuracy may potentially decrease caused by incorrect classification; (2) The utilization of four adjacent neighbors effectively improves the localization resolution.

3.3.2. Fusing R ^ i , j and ρ i , j

We convert the reliability vector R ^ i , j into a tampering binary mask M ^ i , j { 0 , 1 } , based on the majority voting of the reliability vectors generated by neighboring patches. When M ^ i , j = 0 , P i , j is pristine; on the contrary, when M ^ i , j = 1 , P i , j is forged. Next, ρ i , j can be calculated using the following equation:
ρ i , j = M ^ K ,
where K is the number of adjacent patches for P i , j , and we set K as 8 to facilitate detection in practice. If ρ i , j is smaller than τ 1 , it is proposed to refine detected region in the mask by setting all inspected patches as pristine, which can be formulated as follows:
M ^ i , j = 0 if ρ i , j < τ 1 ,
where τ 1 [ 0 , 1 ] denotes a threshold. Note that when τ 1 = 0 , we do not take ρ i , j into consideration; when τ 1 = 1 , the inspected patch requires K forged adjacent patches. Then, we can generate the binary map RFM through M ^ i , j . For clarity, the visualization result of RFM is illustrated in Figure 9.

3.3.3. Designing Binary Classifier

To automatically realize the end-to-end detection, we introduce τ 2 to determine whether image I is forged or not by counting the number of forged patches:
I is pristine if μ M ^ τ 2 I is forged if μ M ^ > τ 2
where a threshold τ 2 [ 0 , 1 ] controls the number of forged patches in an inquiry image. μ M ^ denotes the averaged tampering rate of image I , which is calculated using the below equation:
μ M ^ = M ^ i , j N 1 × N 2 ,
where N 1 × N 2 denotes the total number of patches extracted from I .

4. Experimental Results

In order to comprehensively evaluate the performance of our proposed RFM-based detector, we focus on pre-processing effectiveness, binary tampering detection, and forgery localization. The results are compared with the competing state-of-the-art approaches. First, we will describe the database used in our evaluation.
We utilize the benchmark Dresden Database [48], which consists of more than 16,000 images from 26 different camera models depicting a total of 83 scenes. In our evaluation, we randomly selected 18 camera models from the Dresden Database, and split them into a training set D T , a validation set D V and an evaluation set D E .
Images both from dataset D T and D V were first divided into 64 × 64 overlapped patches. Then, we trained the CNN module in Section 3.2 in virtue of the Stochastic Gradient Descent [49]. We randomly selected 2700 images (150 images per model) as the training set D T , and another 1800 images (100 images per model) as the validation set D V . Meanwhile, we modified 500 images using the cross-model strategy from D V , and randomly chose another 500 images from D V as pristine samples, with a total of 1000 images (over 2,000,000 patches) as the evaluation set D E . In the following, we will describe the cross-model strategy.
The procedure of generating forged images is described in Algorithm 1. We first randomly select 500 images from nine camera models as group A, and 500 images from the remaining camera models as group B. Subsequently, we will select an image I t m p from group B to tamper a host image I r e v from group A. The next step is to generate a blank mask M with the same size of I r e v . Then, we crop a random rectangle region Q with the size of w × h ( w 128 , 1024 and h 128 , 1024 ) from I t m p , and splice it into a random location of I r e v as forged image I f o r g e . Finally, we update M to mark the tampering region, and respectively, save I f o r g e and M as forged image and ground truth mask.
Finally, it is proposed to validate our algorithm based on the trained CNN module. It should be noted that we use the same forged dataset in our experiments for fair comparison. We implement the experiments on a single Nvidia GPU card of type GeForce GTX 1070, with its built-in Deep Learning Tensorflow.
Algorithm 1: Procedure of generating forged images
Sensors 20 06668 i001

4.1. Pre-Processing Performance Evaluation

In the first evaluation, we intend to understand the knowledge hidden in the pre-processing stage. We experimentally compare our proposed high-pass filter (RFM-CNN for abbreviation), trainable pre-processing filter (Constrained-CNN) [23], and our previous work (SCI-CNN) [47] without pre-processing operation, to validate the effectiveness of pre-processing performance. It should be noted that RFM-CNN represents the key step of our proposed RFM-based detector, which only contains pre-processing and feature extraction stages. To this end, they were first trained with D T and then evaluated by D E .
Figure 5 depicts the training accuracy curves for our proposed RFM-CNN, SCI-CNN [47] and Constrained-CNN [23].
For fair comparison, the same CNN architecture is adopted in this experiment. It should be noted that the accuracy here is used for evaluating the classification performance of images from various camera models (i.e., patch-wise accuracy), different from the definition of accuracy of tampering localization in the following subsection. We observe that RFM-CNN had an average accuracy of over 90% using only about 3000 training steps, which achieved faster convergence than Constrained-CNN and SCI-CNN. Due to the constant pre-processing filter, the RFM-CNN framework was able to leverage the CNN to extract inherent characteristics of an image. Besides, it implies that the better-performed classification for identifying camera model undoubtedly leads to higher accuracy of tampering detection and localization.
As Figure 6 reports, we illustrate the detection visualization results between the proposed RFM-CNN and the other pre-processing strategies. We inserted a red bounding box labeling the tampering region. It should be noted that the pre-processing result of SCI-CNN is actually grayscale version of inspected color image, since the pre-processing operation was not adopted in that method. One can also observe that both RFM-CNN and constrained-CNN were capable of suppressing low-frequency content while enhancing high-frequency content. Moreover, according to magnitude of mismatch detection, RFM-CNN had a higher ability of feature extraction using constant filter, compared with Constrained-CNN and SCI-CNN. Therefore, from Figure 5 and Figure 6, one can conclude that the proposed RFM-CNN performs effectively in accelerating the convergence of neural network and assisting the CNN module to better extract features precisely.
Next, we analyze the importance of adopting the pre-processing stage prior to CNN. When extracting intrinsic features, it is required to suppress content-related features. Thus, it is proposed to enhance the effectiveness of the CNN equipped with pre-processing stage for capturing image intrinsic fingerprints. Moreover, an efficient pre-processing operation, referring to as an effective high-pass filter, can further improve the convergence and efficiency in feature extraction of CNN. For instance, an appropriate constrained filter has verified its effectiveness of improving detection performance (see [23]).

4.2. Tampering Detection

In this section, we presented the performance evaluation of the RFM algorithm on tampering detection. The proposed CNN was first trained using D T and then tested with D E . We adjusted thresholds τ 1 and τ 2 to obtain different results. Table 2 illustrates the detection accuracy (ACC), true positive rate (TPR) and false positive rate (FPR) of the RFM-based detector. Figure 7 describes the ROC curves under different τ 1 and τ 2 . In this experiment, the ACC denotes tampering detection accuracy (i.e., binary classification) for proposed RFM method, which can be formulated as follows:
A C C = T P + T N N ,
where T P denotes true positive and T N denotes true negative, N denotes the total number of images in D E . Besides, TPR can be formulated as follows:
T P R = T P T P + F N .
It should be noted that τ 1 plays an important role in reducing mis-classified patches. Additionally, τ 2 plays a critical role in determining the number of detected patches for identifying a forged image.
Table 2 describes the performance of our proposed RFM-based detector (i.e., an average ACC of 92.2%). As Table 2 illustrates, when τ 1 decreased from 0.6 to 0, the ACC decreased from 94.9% to 90.4%. In other words, the RFM in the fusing stage can effectively reduce mis-classification, and meanwhile refine tampering detection. Figure 7 describes the ROC curves obtained from different threshold τ 1 and τ 2 , where TPR achieves high values even at a very low FPR. Thus, the findings supported the fact that our detector can precisely identify forged images with a low mis-classification rate.
Moreover, we compared the proposed RFM-based detector with [26] and [23], where [26] focused on clustering CNN features and [23] had a trainable pre-processing filter (Constrained CNN). For a fair comparison, the same pre-trained CNN module was applied to our proposed method and the approach of [26]. Meanwhile, we added an additional experiment by adopting the RFM algorithm followed by the CNN output of [23] (see [23]+RFM in Figure 8). We used both ACC and TPR as the evaluation metrics to complete the comparison experiments.
Figure 8 presents the detection results of the RFM-based detector with various thresholds τ 1 and τ 2 , together with the other prior-art methods. Compared with methods proposed in [26] and [23], the RFM-based detector achieved the best accuracy of 94.9% when τ 1 = 0.6 . Additionally, when we adopted the RFM algorithm to refine CNN features of [23], both ACC and FPR gained a remarkable enhancement. The main reason is that [23] adopts the strategy based on each isolated patch without taking features of adjacent patches into consideration, while our proposed RFM algorithm reduces the mis-classified result caused by one single patch, and meanwhile improves the accuracy.

4.3. Tampering Localization

We then compared the performance of our RFM-based detector with [26] for tampering localization. The CNN module was trained with the set D T , and then verified using D E . For the evaluation metrics, we used both local and global detection accuracy. The local accuracy refers to the ratio of the number of detected forgery patches to that of all the forgery patches; the global accuracy refers to the ratio of the number of correctly-classified patches (both forgery and pristine patches) to that of all the patches (a full-size image). It is worth noting that the local accuracy only depends on tampering region, and serves as an evaluation metric to evaluate localization resolution. The global accuracy plays a critical role in evaluating the patch-wise detection performance.
Table 3 reports the results of tampering localization. It is observed that our RFM-based detector outperforms that of [26], with an average accuracy of over 90 % (local accuracy), better than around 70 % from [26]. That is, our proposed algorithm achieved significant improvement in the resolution of localization. Meanwhile, as Figure 9 illustrates the visualization results, the RFM-based detector had a higher resolution of localization, namely effective in locating the subtle tampering region. While our proposed RFM-based detector cannot perform as well as that of [26] in global accuracy. Nevertheless, Table 3 and Figure 9 empirically verify that our proposed RFM-based detector performs better in the resolution of localization.
A better insight on the result of each step can be demonstrated by a visual inspection of the examples of Figure 10. When only relying on the extracted features from CNN, one can observe that a large-scale mismatched patches labeled as dispersive colorized rectangles are scatted on the binary map (see Figure 10d). By adopting our proposed RFM algorithm, those mismatched patches can be filtered and refined (see Figure 10e), leading to more accurate tampering localization. It should also be noted that the tampering traces of examples in Figure 10 are hardly visually noticeable, which further highlights the powerful superiority of our proposed RFM-based detector.

5. Conclusions

The resolution of forgery localization is becoming more challenging for digital image forensics. Thus, in this paper, relying on CNN, we presented an RFM-based detector for authenticating a forged image and localizing tampering region. Specifically, in order to improve the accuracy of both tampering detection and localization resolution, we focused on the design of high-pass filter, the establishment of CNN architecture, and the construction of reliability fusion map, which mainly relies on patch texture, CNN confidence, and density distribution. Extensive evaluation results empirically demonstrated that our proposed RFM-based detector outperforms the prior arts in the resolution of localization.
However, the tampering technique is also advancing with the rapid development of image-editing software. Therefore, it is required to design an updated forensic detector for addressing the new challenge. Recently, a bunch of high-efficient detectors equipped with the new algorithms have been proposed to improve the performance of tampering detection and localization [50,51,52]. In our future work, we intend to further investigate the feature extractor characterizing the camera instance (not only focusing on the camera model) for widely tampering detection.

Author Contributions

Data curation, Y.W.; formal analysis, T.Q. and N.Z.; investigation, H.Y. and Y.W.; methodology, H.Y., M.X. and T.Q.; project administration, H.Y.; resources, M.X. and N.Z.; supervision, N.Z.; validation, Y.W.; writing—original draft, H.Y.; writing—review & editing, M.X. and T.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Cyberspace Security Major Program in the National Key Research and Development Plan of China under grant No. 2016YFB0800201, the Natural Science Foundation of China under grant No. 61702150, and 61803135, the Public Research Project of Zhejiang Province under grant No. LGG19F020015, the Key Research and Development Plan Project of Zhejiang Province under grant No. 2017C01065.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qiao, T.; Shi, R.; Luo, X.; Xu, M.; Zheng, N.; Wu, Y. Statistical model-based detector via texture weight map: Application in re-sampling authentication. IEEE Trans. Multimedia 2018, 21, 1077–1092. [Google Scholar] [CrossRef]
  2. Qiao, T.; Zhu, A.; Retraint, F. Exposing image resampling forgery by using linear parametric model. Multimed. Tools Appl. 2018, 77, 1501–1523. [Google Scholar] [CrossRef]
  3. Amerini, I.; Ballan, L.; Caldelli, R.; Del Bimbo, A.; Del Tongo, L.; Serra, G. Copy-move forgery detection and localization by means of robust clustering with J-Linkage. Signal Process. Image Commun. 2013, 28, 659–669. [Google Scholar] [CrossRef]
  4. Pan, X.; Lyu, S. Region duplication detection using image feature matching. IEEE Trans. Inf. Forensics Secur. 2010, 5, 857–867. [Google Scholar] [CrossRef]
  5. Zhao, Y.; Zheng, N.; Qiao, T.; Xu, M. Source camera identification via low dimensional PRNU features. Multimed. Tools Appl. 2019, 78, 8247–8269. [Google Scholar] [CrossRef]
  6. Qiao, T.; Retraint, F.; Cogranne, R.; Thai, T.H. Individual camera device identification from JPEG images. Signal Process. Image Commun. 2017, 52, 74–86. [Google Scholar] [CrossRef] [Green Version]
  7. Qiao, T.; Retraint, F. Identifying Individual Camera Device From RAW Images. IEEE Access 2018, 6, 78038–78054. [Google Scholar] [CrossRef]
  8. Chen, M.; Fridrich, J.; Goljan, M.; Lukás, J. Determining image origin and integrity using sensor noise. IEEE Trans. Inf. Forensics Secur. 2008, 3, 74–90. [Google Scholar] [CrossRef] [Green Version]
  9. Hsu, Y.F.; Chang, S.F. Camera response functions for image forensics: An automatic algorithm for splicing detection. IEEE Trans. Inf. Forensics Secur. 2010, 5, 816–825. [Google Scholar] [CrossRef] [Green Version]
  10. Zhao, X.; Wang, S.; Li, S.; Li, J. Passive image-splicing detection by a 2-D noncausal Markov model. IEEE Trans. Circuits Syst. Video Technol. 2015, 25, 185–199. [Google Scholar] [CrossRef]
  11. Salloum, R.; Ren, Y.; Kuo, C.C.J. Image Splicing Localization Using A Multi-Task Fully Convolutional Network (MFCN). J. Vis. Commun. Image Represent 2018, 51, 201–209. [Google Scholar] [CrossRef] [Green Version]
  12. Bunk, J.; Bappy, J.H.; Mohammed, T.M.; Nataraj, L.; Flenner, A.; Manjunath, B.; Chandrasekaran, S.; Roy-Chowdhury, A.K.; Peterson, L. Detection and localization of image forgeries using resampling features and deep learning. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1881–1889. [Google Scholar]
  13. Chen, C.; McCloskey, S.; Yu, J. Image Splicing Detection via Camera Response Function Analysis. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 5087–5096. [Google Scholar]
  14. Rao, Y.; Ni, J. A deep learning approach to detection of splicing and copy-move forgeries in images. In Proceedings of the 2016 IEEE International Workshop on Information Forensics and Security, Abu Dhabi, United Arab Emirates, 4–7 December 2016; pp. 1–6. [Google Scholar]
  15. Wu, Y.; Abd-Almageed, W.; Natarajan, P. Deep matching and validation network: An end-to-end solution to constrained image splicing localization and detection. In Proceedings of the 2017 ACM on Multimedia Conference, Silicon Valley, CA, USA, 23–27 December 2017; pp. 1480–1502. [Google Scholar]
  16. Cozzolino, D.; Poggi, G.; Verdoliva, L. Efficient dense-field copy–move forgery detection. IEEE Trans. Inf. Forensics Secur. 2015, 10, 2284–2297. [Google Scholar] [CrossRef]
  17. Soni, B.; Das, P.K.; Thounaojam, D.M. Copy-Move Tampering Detection based on Local Binary Pattern Histogram Fourier Feature. In Proceedings of the 7th International Conference on Computer and Communication Technology, Allahabad, India, 24–26 November 2017; pp. 78–83. [Google Scholar]
  18. Verdoliva, L.; Cozzolino, D.; Poggi, G. A feature-based approach for image tampering detection and localization. In Proceedings of the 2014 IEEE Information Forensics and Security (WIFS), Atlanta, GA, USA, 3–5 December 2014; pp. 149–154. [Google Scholar]
  19. Bappy, J.H.; Roy-Chowdhury, A.K.; Bunk, J.; Nataraj, L.; Manjunath, B. Exploiting spatial structure for localizing manipulated image regions. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4970–4979. [Google Scholar]
  20. Jin, X.; Su, Y.; Zou, L.; Zhang, C.; Jing, P.; Song, X. Video logo removal detection based on sparse representation. Multimed. Tools Appl. 2018, 77, 29303–29322. [Google Scholar] [CrossRef]
  21. Pibre, L.; Jérôme, P.; Ienco, D.; Chaumont, M. Deep Learning for Steganalysis Is Better Than a Rich Model with An Ensemble Classifier, and Is Natively Robust to the Cover Source-Mismatch. Available online: https://arxiv.org/abs/1511.04855 (accessed on 16 November 2015).
  22. Bayar, B.; Stamm, M.C. Augmented convolutional feature maps for robust cnn-based camera model identification. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 4098–4102. [Google Scholar]
  23. Bayar, B.; Stamm, M.C. Constrained convolutional neural networks: A new approach towards general purpose image manipulation detection. IEEE Trans. Inf. Forensics Secur. 2018, 13, 2691–2706. [Google Scholar] [CrossRef]
  24. Li, H.; Luo, W.; Qiu, X.; Huang, J. Image forgery localization via integrating tampering possibility maps. IEEE Trans. Inf. Forensics Secur. 2017, 12, 1240–1252. [Google Scholar] [CrossRef]
  25. Güera, D.; Zhu, F.; Yarlagadda, S.K.; Tubaro, S.; Bestagini, P.; Delp, E.J. Reliability map estimation for CNN-based camera model attribution. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 964–973. [Google Scholar]
  26. Bondi, L.; Lameri, S.; Güera, D.; Bestagini, P.; Delp, E.J.; Tubaro, S. Tampering detection and localization through clustering of camera-based CNN features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1855–1864. [Google Scholar]
  27. Kang, X.; Stamm, M.C.; Peng, A.; Liu, K.R. Robust median filtering forensics using an autoregressive model. IEEE Trans. Inf. Forensics Secur. 2013, 8, 1456–1468. [Google Scholar] [CrossRef] [Green Version]
  28. Chierchia, G.; Cozzolino, D.; Poggi, G.; Sansone, C.; Verdoliva, L. Guided filtering for PRNU-based localization of small-size image forgeries. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 6231–6235. [Google Scholar]
  29. Popescu, A.C.; Farid, H. Exposing digital forgeries by detecting traces of resampling. IEEE Trans. Signal Process. 2005, 53, 758–767. [Google Scholar] [CrossRef]
  30. Kirchner, M. Fast and reliable resampling detection by spectral analysis of fixed linear predictor residue. In Proceedings of the 10th ACM Workshop on Multimedia and Security, Oxford, UK, 22–23 September 2008; pp. 11–20. [Google Scholar]
  31. Qiu, X.; Li, H.; Luo, W.; Huang, J. A universal image forensic strategy based on steganalytic model. In Proceedings of the 2nd ACM Workshop on Information Hiding and Multimedia Security, Salzburg, Austria, 11 June 2014; pp. 165–170. [Google Scholar]
  32. Cozzolino, D.; Verdoliva, L. Noiseprint: A CNN-based camera model fingerprint. arXiv 2019, arXiv:1808.08396. [Google Scholar]
  33. Zhang, Z.; Zhang, Y.; Zhou, Z.; Luo, J. Boundary-based image forgery detection by fast shallow cnn. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2658–2663. [Google Scholar]
  34. Mazaheri, G.; Mithun, N.C.; Bappy, J.H.; Roy-Chowdhury, A.K. A Skip Connection Architecture for Localization of Image Manipulations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019; pp. 119–129. [Google Scholar]
  35. Bi, X.; Wei, Y.; Xiao, B.; Li, W. RRU-Net: The Ringed Residual U-Net for Image Splicing Forgery Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019; pp. 30–39. [Google Scholar]
  36. Chen, Y.; Kang, X.; Shi, Y.Q.; Wang, Z.J. A multi-purpose image forensic method using densely connected convolutional neural networks. J. REAL-TIME IMAGE PR. 2019, 16, 725–740. [Google Scholar] [CrossRef]
  37. Bondi, L.; Güera, D.; Baroffio, L.; Bestagini, P.; Delp, E.J.; Tubaro, S. A preliminary study on convolutional neural networks for camera model identification. J. Electron. Imaging 2017, 2017, 67–76. [Google Scholar] [CrossRef] [Green Version]
  38. Farid, H. Exposing digital forgeries from JPEG ghosts. IEEE Trans. Inf. Forensics Secur. 2009, 4, 154–160. [Google Scholar] [CrossRef]
  39. Bianchi, T.; Piva, A. Image forgery localization via block-grained analysis of JPEG artifacts. IEEE Trans. Inf. Forensics Secur. 2012, 7, 1003–1017. [Google Scholar] [CrossRef] [Green Version]
  40. Cozzolino, D.; Gragnaniello, D.; Verdoliva, L. Image forgery localization through the fusion of camera-based, feature-based and pixel-based techniques. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 5302–5306. [Google Scholar]
  41. Gaborini, L.; Bestagini, P.; Milani, S.; Tagliasacchi, M.; Tubaro, S. Multi-clue image tampering localization. In Proceedings of the 2014 IEEE International Workshop on Information Forensics and Security (WIFS), Atlanta, GA, USA, 3–5 December 2014; pp. 125–130. [Google Scholar]
  42. Korus, P.; Huang, J. Multi-scale fusion for improved localization of malicious tampering in digital images. IEEE Trans. Image Process. 2016, 25, 1312–1326. [Google Scholar] [CrossRef]
  43. Korus, P.; Huang, J. Improved tampering localization in digital image forensics based on maximal entropy random walk. IEEE Signal Process Lett. 2016, 23, 169–173. [Google Scholar] [CrossRef]
  44. Qian, Y.; Dong, J.; Wang, W.; Tan, T. Deep learning for steganalysis via convolutional neural networks. In Proceedings of the Media Watermarking, Security, and Forensics 2015, San Francisco, CA, USA, 9–11 February 2015; Volume 9409, p. 94090. [Google Scholar]
  45. Pibre, L.; Pasquet, J.; Ienco, D.; Chaumont, M. Deep learning is a good steganalysis tool when embedding key is reused for different images, even if there is a cover sourcemismatch. Electron. Imaging 2016, 2016, 1–11. [Google Scholar] [CrossRef] [Green Version]
  46. Tuama, A.; Comby, F.; Chaumont, M. Camera model identification with the use of deep convolutional neural networks. In Proceedings of the 2016 IEEE International Workshop on Information Forensics and Security (WIFS), Abu Dhabi, United Arab Emirates, 4–7 December 2016; pp. 1–6. [Google Scholar]
  47. Yao, H.; Qiao, T.; Xu, M.; Zheng, N. Robust Multi-Classifier for Camera Model Identification Based on Convolution Neural Network. IEEE Access 2018, 6, 24973–24982. [Google Scholar] [CrossRef]
  48. Gloe, T.; Böhme, R. The dresden image database for benchmarking digital image forensics. J. Digit. Forensic Pract. 2010, 3, 150–159. [Google Scholar] [CrossRef]
  49. Bottou, L. Large-scale machine learning with stochastic gradient descent. In Proceedings of the 19th International Conference on Computational Statistics (COMPSTAT), Paris, France, 22–27 August 2010; pp. 177–186. [Google Scholar]
  50. Fan, B.; Kong, Q.; Zhang, B.; Liu, H.; Pan, C.; Lu, J. Efficient nearest neighbor search in high dimensional hamming space. Pattern Recognit. 2020, 99, 107082. [Google Scholar] [CrossRef]
  51. Chen, B.; Tan, W.; Coatrieux, G.; Zheng, Y.; Shi, Y.Q. A serial image copy-move forgery localization scheme with source/target distinguishment. IEEE Trans. Multimedia 2020. [Google Scholar] [CrossRef]
  52. Fan, B.; Liu, H.; Zeng, H.; Zhang, J.; Liu, X.; Han, J. Deep Unsupervised Binary Descriptor Learning through Locality Consistency and Self Distinctiveness. IEEE Trans. Multimedia 2020. [Google Scholar] [CrossRef]
Figure 1. Generic framework for image tampering detection and localization, including image pre-processing, feature extraction, and post-processing.
Figure 1. Generic framework for image tampering detection and localization, including image pre-processing, feature extraction, and post-processing.
Sensors 20 06668 g001
Figure 2. Flowchart of our proposed classifier.
Figure 2. Flowchart of our proposed classifier.
Sensors 20 06668 g002
Figure 3. Architecture of the convolution neural network (CNN) module, with 13 conventional layers, three fully-connected layers and a softmax layer.
Figure 3. Architecture of the convolution neural network (CNN) module, with 13 conventional layers, three fully-connected layers and a softmax layer.
Sensors 20 06668 g003
Figure 4. Illustration of the reliability fusion map (RFM) algorithm pipeline (a), and step by step clustering approach of [26] (b). “px” is the abbreviation of “pixel”.
Figure 4. Illustration of the reliability fusion map (RFM) algorithm pipeline (a), and step by step clustering approach of [26] (b). “px” is the abbreviation of “pixel”.
Sensors 20 06668 g004
Figure 5. Accuracy curves on training dataset ( D T ) for Constrained-CNN [23], SCI-CNN [47] and RFM-CNN proposed in this work.
Figure 5. Accuracy curves on training dataset ( D T ) for Constrained-CNN [23], SCI-CNN [47] and RFM-CNN proposed in this work.
Sensors 20 06668 g005
Figure 6. Tampering localization with different pre-processing stages: (a) forged image; (e) ground truth; (b) SCI-CNN denoting grayscale input image without pre-processing operation; (c) Constrained-CNN; (d) RFM-CNN with pre-processing operation; (fh) visualization results generated by different methods.
Figure 6. Tampering localization with different pre-processing stages: (a) forged image; (e) ground truth; (b) SCI-CNN denoting grayscale input image without pre-processing operation; (c) Constrained-CNN; (d) RFM-CNN with pre-processing operation; (fh) visualization results generated by different methods.
Sensors 20 06668 g006
Figure 7. ROC curves of tampering detection results using our RFM-based detector with various thresholds τ 1 and τ 2 .
Figure 7. ROC curves of tampering detection results using our RFM-based detector with various thresholds τ 1 and τ 2 .
Sensors 20 06668 g007
Figure 8. Accuracy (ACC), true positive rate (TPR) of proposed method with various thresholds τ 1 , τ 2 (on the left of the gray dashed line) and the other competing algorithms (on the right of the gray dashed line). Blue and orange dashed lines denote the best ACC and TPR results of our proposed RFM-based detector, respectively.
Figure 8. Accuracy (ACC), true positive rate (TPR) of proposed method with various thresholds τ 1 , τ 2 (on the left of the gray dashed line) and the other competing algorithms (on the right of the gray dashed line). Blue and orange dashed lines denote the best ACC and TPR results of our proposed RFM-based detector, respectively.
Sensors 20 06668 g008
Figure 9. Comparison of localization performance between our RFM-based detector and the algorithm of [26].
Figure 9. Comparison of localization performance between our RFM-based detector and the algorithm of [26].
Sensors 20 06668 g009
Figure 10. Tampering localization using our proposed RFM-based detector; from left to right: (a) forged image, (b) ground truth, (c) pre-processing result, (d) detection result without RFM (only relying on feature extraction), and (e) detection result with RFM (by adopting post-processing procedure).
Figure 10. Tampering localization using our proposed RFM-based detector; from left to right: (a) forged image, (b) ground truth, (c) pre-processing result, (d) detection result without RFM (only relying on feature extraction), and (e) detection result with RFM (by adopting post-processing procedure).
Sensors 20 06668 g010
Table 1. Configuration of each convolutional layer in Figure 3.
Table 1. Configuration of each convolutional layer in Figure 3.
IDInput SizeConfigurationType
conv 164 × 64-3stride = 2, ksize = 8 × 8conv + ReLU
conv 232 × 32-16stride = 1, ksize = 8 × 8conv + ReLU
conv 332 × 32-32stride = 2, ksize = 6 × 6conv + ReLU
conv 416 × 16-48stride = 1, ksize = 6 × 6conv + ReLU + maxpool
conv 516 × 16-64stride = 1, ksize = 3 × 3conv + ReLU
conv 616 × 16-128stride = 2, ksize = 3 × 3conv + ReLU
conv 78 × 8-256stride = 1, ksize = 3 × 3conv + ReLU + maxpool
conv 88 × 8-512stride = 2, ksize = 3 × 3conv + ReLU
conv 98 × 8-1024stride = 2, ksize = 3 × 3conv + ReLU + maxpool
conv 104 × 4-512stride = 1, ksize = 1 × 1conv + ReLU
conv 114 × 4-256stride = 1, ksize = 1 × 1conv + ReLU
conv 124 × 4-128stride = 2, ksize = 1 × 1conv + ReLU
conv 131 × 1-64stride = 2, ksize = 1 × 1conv + ReLU + maxpool
Table 2. Results of tampering detection with various thresholds τ 1 and τ 2 .
Table 2. Results of tampering detection with various thresholds τ 1 and τ 2 .
ThresholdACCTPRFPR
τ 1 = 0.0 τ 2 = 0.0150.9040.8280.020
τ 1 = 0.4 τ 2 = 0.0150.9420.9100.026
τ 1 = 0.6 τ 2 = 0.0150.9490.9420.044
τ 1 = 0.6 τ 2 = 0.0120.8920.7920.008
Average0.9220.8680.025
Table 3. Tampering localization comparison between our RFM-based detector and the algorithm of [26].
Table 3. Tampering localization comparison between our RFM-based detector and the algorithm of [26].
MethodThresholdLocal AccuracyGlobal AccuracyResolution
RFM-based τ 1 = 0.40.9050.954 32 × 32
RFM-based τ 1 = 0.60.9070.955 32 × 32
[26] λ dist = 0.7, λ conf = 0.20.7120.982 64 × 64
[26] λ dist = 0.7, λ conf = 0.00.7340.983 64 × 64
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yao, H.; Xu, M.; Qiao, T.; Wu, Y.; Zheng, N. Image Forgery Detection and Localization via a Reliability Fusion Map. Sensors 2020, 20, 6668. https://doi.org/10.3390/s20226668

AMA Style

Yao H, Xu M, Qiao T, Wu Y, Zheng N. Image Forgery Detection and Localization via a Reliability Fusion Map. Sensors. 2020; 20(22):6668. https://doi.org/10.3390/s20226668

Chicago/Turabian Style

Yao, Hongwei, Ming Xu, Tong Qiao, Yiming Wu, and Ning Zheng. 2020. "Image Forgery Detection and Localization via a Reliability Fusion Map" Sensors 20, no. 22: 6668. https://doi.org/10.3390/s20226668

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop