Next Article in Journal
Dynamic Hand Gesture Recognition for Smart Lifecare Routines via K-Ary Tree Hashing Classifier
Next Article in Special Issue
Minimizing Maximum Feature Space Deviation for Visible-Infrared Person Re-Identification
Previous Article in Journal
Smooth Bend Structures Using Hybrid Compensation for Common-Mode Noise Reduction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Scale-Adaptive Deep Matching Network for Constrained Image Splicing Detection and Localization

Beijing Electronic Science and Technology Institute, Beijing 100070, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(13), 6480; https://doi.org/10.3390/app12136480
Submission received: 15 May 2022 / Revised: 12 June 2022 / Accepted: 16 June 2022 / Published: 26 June 2022
(This article belongs to the Special Issue Recent Applications of Computer Vision for Automation and Robotics)

Abstract

:
Constrained image splicing detection and localization (CISDL) is a newly formulated image forensics task that aims at detecting and localizing the source and forged regions from a series of input suspected image pairs. In this work, we propose a novel Scale-Adaptive Deep Matching (SADM) network for CISDL, consisting of a feature extractor, a scale-adaptive correlation module and a novel mask generator. The feature extractor is built on VGG, which has been reconstructed with atrous convolution. In the scale-adaptive correlation computation module, squeeze-and-excitation (SE) blocks and truncation operations are integrated to process arbitrary-sized images. In the mask generator, an attention-based separable convolutional block is designed to reconstruct richer spatial information and generate more accurate localization results with less parameters and computation burden. Last but not least, we design a pyramid framework of SADM to capture multiscale details, which can increase the detection and localization accuracy of multiscale regions and boundaries. Extensive experiments demonstrate the effectiveness of SADM and the pyramid framework.

1. Introduction

With the widespread availability of affordable acquisition devices such as smartphones and the easy use of powerful image editing software, we can easily modify digital images without leaving any perceptible artifacts [1]. Malicious tampered images can distort the truth in news reports, or destroy someone’s reputation and privacy, leading to potentially devastating consequences [2]. Digital image forensics intends to verify the authenticity of digital images and provide automatic tools to detect image manipulation [3]. In conventional image forensics methods, image manipulation always causes high-level or low-level inconsistencies, which can determine whether images have been tampered with. However, these image forensics techniques only investigate a single image, and the information provided by the single image is limited [1,4,5]. So, it is difficult to identify fake images accurately with the improved image editing techniques. Otherwise, these conventional image forensics methods do not provide the source of the forged area or the specific tampering process, which reduces the persuasion of the detection results in practical application.
To tackle these limitations, constrained image splicing detection and localization (CISDL) was proposed to find forged regions and corresponding source regions in a pair of candidate images by comparing pixel-level features [6,7]. Theoretically, with a probe image P and a potential donor image D, CISDL aims to determine whether P has spliced regions from D. In [6], Wu et al. designed a deep matching and validation network (DMVN) for CISDL. It contains four modules: a feature extractor, an inception-based mask deconvolution module, a visual consistency validator module and a Siameselike module. DMVN is the first method to address CISDL task and uses visual comparison to show the localization performance. In [8], Ye et al. proposed a feature pyramid deep matching and localization network (FPLN), which can detect and localize small spliced regions by fusing pyramidal feature maps with different resolution [1]. However, the loss of spatial information limits the discriminative ability and localization accuracy of the splicing model. In [7], Liu et al. proposed a new deep matching network for CISDL named DMAC, using atrous convolution to generate two high-quality candidate masks. Additionally, they employ a detection network and a discriminative network to optimize the pretrained DMAC further. Though this framework achieves remarkable improvement than DMVN, the discriminative ability of DMAC and DMVN for simple scalar product is still limited in correlation computation. Liu et al. proposed an encoder–decoder architecture based on an attention-aware mechanism in [9], which is named as AttentionDM to mitigate these problems. AttentionDM generates fine-grained masks by building a decoder with atrous spatial pyramid pooling (ASPP) [10]. A channel attention block is added to the correlation computation module to emphasize channelwise informative features. However, AttentionDM can only process fixed-size images for the restriction of correlation computation module and cannot be applied to actual multiscale target detection. In a word, there are still some challenges hindering the development of CISDL: (1) fixed-size image processing; (2) degraded spatial information of extracted features; (3) multiscale objects.
In this paper, we propose a novel Scale-Adaptive Deep Matching network (SADM), as shown in Figure 1. In the proposed method, we make three improvements: First, a scale-adaptive correlation computation module is proposed based on squeeze-and-excitation (SE) blocks [11] and truncation operations to deal with arbitrary-sized images. Second, an attention-based separable convolutional block is constructed in mask generator to recover spatial information further. The attention-based separable convolutional block is composed of depthwise separable convolution and spatial attention, that depthwise separable convolution can help to improve efficiency as well as reduce the model parameters. Meanwhile, spatial attention can help recover the spatial information. Last but not least, we propose a pyramid framework of SADM to address the problem of insufficient edge detection and multiscale object detection.
In summary, our contributions are:
A novel scale-adaptive deep matching network (SADM) is constructed based on a newly designed scale-adaptive correlation computation module.
A novel mask generator is designed by attention-based separable convolutional blocks, consisting of depthwise separable convolution and spatial attention. This kind of hybridized combination is helpful for spatial information reconstruction with low computational complexity.
A pyramid framework of SADM (PSADM) is proposed to process the multiscale objects.
Extensive experiments prove the excellent performance of SADM and PSADM.
This paper is organized as follows. Section 2 describes the related work. Section 3 introduces SADM and the pyramid version of SADM. Experimental results and visual comparisons are presented in Section 4. In Section 5, we make a conlusion.

2. Related Work

2.1. Image Forensics

Digital image forensics [12,13] has two main tasks: image forgery localization and detection. The purpose of image forgery detection is to determine whether an image has been forged, while image forgery localization is to mark the forged area on the forged image. The main primary of tampering detection technology is that the statistical features inherently in digital image acquisition are inevitably disturbed by tampering operations. That manipulated images can be detected and distinguished by analyzing these features. In practical image forensics applications, users are concerned more about which areas of the image rather than whether an image has been tampered with, making tampering localization an important research topic in image forensics.
In recent years, deep learning algorithms [14,15,16] represented by convolutional neural networks [17,18,19], recurrent neural networks [20] and generative adversarial networks have been widely used in many fields such as image classification [21,22], object detection [23], semantic segmentation [24,25], image retrieval [26], scene understanding [27], etc. and have made a leap forward compared with traditional methods. Given the outstanding performance of deep learning algorithms in computer vision, researchers have also tried to adopt deep learning algorithms to solve some problems in image forensics. However, in terms of tampering localization, both traditional methods and the recent emergence of deep learning feature-based methods are still difficult to achieve industrial application [28]. In this work, we focus on the visual features of tampered images and the statistical inconsistency between pixels to investigate the image tampering localization. Besides, we also try to explore effective and robust image tampering localization methods by using deep learning algorithms for homologous and heterologous tampering.

2.2. CISDL

Constrained image splicing detection and localization, proposed for image forensic task, plays a crucial part in constructing a germline genesis map of an image by dense matching. This task aims to afford the source image of the tampered region and the corresponding region by finding high similarity of corresponding regions over long distances. Studying CISDL has great display significance in improving the accuracy of image stitching detection and localization, etc. The three methods to address this task are as follows: Wu et al. first proposed a method to address CISDL specifically [6]. They apply a deep convolutional neural network architecture to deep matching and validation network (DMVN) with two input images. DMVN is composed of a feature extractor with convolutional neural network (CNN), a inception-based mask deconvolution module, a visual consistency validator module and a siameselike module producing a probability value. The value indicates the likelihood between donor image and query image, as for the spliced region. Yet, in [6], it explained the localization performance by visual comparison and didn’t evaluate quantitatively. And DMVN also performs poorly in detecting accurate boundaries and small regions for comparing with high-level resolution feature maps of VGG [10] merely. Liu et al. then proposed a novel framework for CISDL with adversarial learning [7]. The proposed framework contains a deep matching network based on atrous convolution (DMAC), a detection network and a discriminative network. The detection network and the discriminative network can optimize the masks generated from DMAC adversarially. Compared to DMVN, the DMAC network accomplishes real-time behavior in a scene with a large crowd of manipulated images by fully end-to-end architecture. Instead of the adversarial learning framework of DMAC, Liu et al. proposed an attention-aware deep matching network for CISDL, named AttentionDM [9]. Overall, AttentionDM employs an encoder–decoder architecture to generate fine-grained masks. AttentionDM contains a feature extractor by employing VGG16 with atrous convolution, an attention-aware correlation computation module and mask generator module with ASPP blocks to generate the fine-grained mask. Significant progress has been made that the DMVN network has sufficient ability to detect small tampered areas and locate the edges of tampered areas. However, both DMAC and AttentionDM are sensitive to changes in image size and can only manufacture fixed images. Besides, many researched images are uncorrelated or unforged in the actual application, which all cause serious problems. In a sense, when combining the query and donor images, a spliced image can be rendered as a copy-move detection task. So, when CISDL was first proposed, Wu et al. compared against baseline algorithms from the state-of-the-art copy-move detection algorithms and used precision, recall and F1-score to detect spliced image correctly [6]. Besides, Ref. [29] also applies DMVN to judge whether an image has copy-move regions: they feed DMVN the pair of (X1, X2), which is split along X’s longer axis. And they judge the X contains copy-move regions (one in X1 and the other in X2) by whether DMVN finds a splice. If not, they then split X1 and X2 into halves again. Above all, constrained image splicing detection and localization methods are mutually linked to copy-move detection method.

3. Method

Our method has three components: the feature extractor, the scale-adaptive correlation computation module and the mask generator. The feature extractor employs VGG16, but removes the maxpooling operation while using atrous convolution in the last convolutional block. In Figure 1, from the feature extractor, three feature maps are generated with the same size. The scale-adaptive correlation computation module adopts SE blocks and truncation operations to break the limit of image size. The mask generator designs an attention-based separable convolutional block based on depthwise separable convolution and spatial attention. Depthwise separable convolution can effectively improve the deep matching, and spatial attention can restore spatial information. In terms of the overall structure, we employ channel attention first, learning ‘what’ in the channel axes, and spatial attention second, learning ‘where’ in the spatial axes, for which can blend cross-channel and spatial information [30]. Additionally, the pyramid version of SADM is proposed to make full use of multiscale information to improve the ability of multiscale objects location.

3.1. Feature Extractor with Atrous Convolution

The structure and parameter settings of the feature extractor, a transformed version of VGG16, are shown in Figure 2. Our feature extractor consists of five blocks. The first two blocks contain two convolutional layers and one maxpooling operation, respectively. The third block includes three convolutional layers and one maxpooling operation. To generate three feature maps of the same size, only three convolutional layers are used in the last two blocks without maxpooling operation. In general, three points differ from VGG16:
(1)
Removing maxpooling operations in the fourth and fifth convolutional blocks. This change makes the final size of the feature map from W / 32 × H / 32 to W / 8 × H / 8 and upgrades image resolution [10].
(2)
Adding atrous convolution at the fifth block of VGG16 [10]. Atrous convolution generalizes standard convolution operation, which utilizes the numerical value of parameters to control the resolution of the feature map for freedom. Specifically, by adjusting the filter’s field-of-view, atrous convolution can collect multiscale information. Atrous convolution is calculated as:
y ( i c , j c ) = k 1 , k 2 w ( k 1 , k 2 ) × x ( i c + r c k 1 , j c + r c k 2 )
where y(i c , j c ) denotes the output of the atrous convolution of a 2-D input of signal x( i c , i c ). Besides, k 1 , k 2 ∈ [ f l( K 2 ),fl( K 2 )], w( k 1 , k 2 ) denotes a K × K filter, atrous rate r c means the sampling stride of the input signal. And the atrous of fifth block is set with r c = 2.
(3)
Skip architecture. Due to removing the maxpooling operation of VGG16’s last two blocks [10], F n ( 1 ) and F n ( 2 ) (n ∈ {3, 4, 5}) are produced with the same size. A high-level feature contains rich semantic information, while a low-level feature contains detailed spatial information. Next, F 3 ( k ) , F 4 ( k ) , F 5 ( k ) (k ∈ {1, 2}) are fed into the scale-adaptive correlation computation module and the mask generator.

3.2. Scale-Adaptive Correlation Computation

As for previous studies, they can not handle arbitrary-sized images. Such as the process of matching response using cross-correlation in DMVN, it uses all pixels of the feature map relative to the size of the image. Similarly, DMAC and AttentionDM continued to use this approach. So, they are all restricted by the size of images. Therefore, distinctly accounting for image scale is essential for improving the model’s ability to process low-resolution and high-resolution images.
In order to address this defect, Liu et al. proposed a sliding window-based matching strategy to process high-resolution images, but it causes high computational complexity [9]. In this paper, we employ correlation computation with SE blocks and truncation operations to boost our model’s ability to process arbitrary-sized images. F 3 ( k ) , F 4 ( k ) , F 5 ( k ) (k ∈ {1, 2}) are extracted from the feature extractor, and then each feature map has to go through the L2-normalization, SE blocks and truncation operations. At last, we utilize ReLU and L2-normalization to produce the final two correlation maps.
(1)
SE blocks
For each channel of the convolution feature, SE blocks model their interdependencies explicitly. Specifically, SE blocks take each channel of feature map as a feature detector, then they utilize global information to emphasize informative features or suppress unuseful ones selectively. Before the SE blocks, L2-normalization is conducted:
f ¯ ( k ) ( i k , j k ) = f ( k ) ( i k , j k ) f ( k ) ( i k , j k ) 2
where F ( k ) R h × w × c , k ∈{1, 2}, f ( 1 ) ( i 1 , j 1 ) ∈ F ( 1 ) , f ( 2 ) ( i 2 , j 2 ) ∈ F ( 2 ) . And then, we can get two feature maps after standardization, F ¯ ( 1 ) , F ¯ ( 2 ) . Next, SE blocks are applied to recalibrate informative features, making an improvement on discriminating features. Based on these above, our SE block has three steps: First, using global average pooling to exploit contextual information. We denote z ( k ) R C as channelwise statistics, and denote f ( k ) as homologous channel dimensions feature, H × W, the cth element of z ( k ) is computed with:
z c ( k ) = F s q ( f ¯ c ( k ) ) = 1 H × W i = 1 H j = 1 W f ¯ c ( k ) ( i , j )
This operation, global average pooling, is the simplest aggregation technique to collect local descriptors that can express the whole image. The second step is to capture channelwise dependencies, which consists of a ReLU and a sigmoid activation as a gating mechanism:
s = F e x ( z ( k ) , W ) = σ ( W 2 δ ( W 1 ( z ( k ) ) )
where δ refers to the ReLU function, W 1 R C r × C and W 2 R C × C r . The third step is to weight the characteristics of each channel, with the weight obtained above. The final output is computed with:
F ¨ ( k ) = F s c a l e ( f ¯ c ( k ) , s c )
where F ¨ ( k ) = [ f ¨ 1 ( k ) , f ¨ 2 ( k ) ,⋯, f ¨ c ( k ) ] and F s c a l e ( f ¯ c ( k ) , s c ) indicate the channelwise multiplication between the scalar s c and the feature map f ¯ ( k ) R H × W .
(2)
Correlation-computation with truncation operations
The process of this part is shown in Figure 3. F ¨ ( 1 ) and F ¨ ( 2 ) refer to the feature maps computed from the SE blocks, f ¨ ( 1 ) ( i 1 , j 1 ) ∈ F ¨ ( 1 ) , f ¨ ( 2 ) ( i 2 , j 2 ) ∈ F ¨ ( 2 ) . The correlation maps are computed by:
C ( 12 ) ( i 12 , j 12 , m 12 ) = f ¨ ( 1 ) ( i 1 , j 1 ) T f ¨ ( 2 ) ( i 2 , j 2 )
where C ( 12 ) is obtained by comparing f ¨ ( 1 ) ( i 1 , j 1 ) ∈ F ¨ ( 1 ) and f ¨ ( 2 ) ( i 2 , j 2 ) ∈ F ¨ ( 2 ) . Since the corresponding f ¨ ( 1 ) ( i 1 , j 1 ) T are not known in advance and most fraction of the features are irrelative, we sort the C ( 12 ) ( i 12 , j 12 , m 12 ) along the (h × w) channels, and select top-T values:
C ^ ( 12 ) ( i 12 , j 12 , 1 : T ) = Top _ T ( Sort ( C ( 12 ) ( i 12 , j 12 , : ) ) )
If apply a curve to show C ^ ( 12 ) ( i 12 , j 12 ), it is supposed to be a monotonic decreasing curve. Once an abrupt drop appeared, it means C ^ ( 12 ) ( i 12 , j 12 ) has matched regions. So, the T channels should include the most drops. Due to the operation of the top-T selection, our network was given the capability of accepting arbitrary-sized images. We summarize above correlation computation as:
C ^ ( 12 ) = Corr ( F ¨ ( 1 ) , F ¨ ( 2 ) )
This is an example of handling a correlation map between two feature maps. After input F 3 ( k ) , F 4 ( k ) , F 5 ( k ) , the scale-adaptive correlation computation module procedure can be concluded as Algorithm 1.
Algorithm 1: Scale-Adaptive Correlation Computation
  Require: Image I ( 1 ) and I ( 2 )
    1: “Hierarchical features extraction:”
    2:  F 3 ( 1 ) , F 4 ( 1 ) , F 5 ( 1 ) = Encoder( I ( 1 ) )
    3:  F 3 ( 2 ) , F 4 ( 2 ) , F 5 ( 2 ) = Encoder( I ( 2 ) )
    4: “Attention weighted feature maps generation for hierarchical features:”
    5: for c = 3 to 5 do
    6:  “L2 normalization of Equation (2)
    7:   F ¯ n ( 1 ) = L2_norm(F n ( 1 ) )
    8:   F ¯ n ( 2 ) = L2_norm(F n ( 2 ) )
    9:  “Refer to Equations (3)–(5)
  10:   F ¨ n ( 1 ) = F s c a l e ( F ¯ m ( 1 ) , F s q ( f c ( 1 ) )
  11:   F ¨ n ( 2 ) = F s c a l e ( F ¯ n ( 2 ) , F s q ( f c ( 2 ) )
  12: end for
  13: “Correlation computation based on hierarchical attention weighted feature maps:”
  14: for c = 3 to 5 do
  15:  “Refer to Equation (8) based on Equations (6) and (7)
  16:   C ^ n ( 12 ) = Corr ( F ¨ n ( 1 ) , F ¨ n ( 2 ) )
  17:   C ^ n ( 11 ) = Corr ( F ¨ n ( 1 ) , F ¨ n ( 1 ) )
  18:   C ^ n ( 21 ) = Corr ( F ¨ n ( 2 ) , F ¨ n ( 1 ) )
  19:   C ^ n ( 22 ) = Corr ( F ¨ n ( 2 ) , F ¨ n ( 2 ) )
  20:  “Concatenate correlation maps
  21:   C n ( 1 ) = { C ^ n ( 12 ) , C ^ n ( 11 ) }
  22:   C n ( 2 ) = { C ^ n ( 21 ) , C ^ n ( 22 ) }
  23: end for
  24: “Concatenate hierarchical correlation maps
  25:  C ( 1 ) = { C 3 ( 1 ) , C 4 ( 1 ) , C 5 ( 1 ) }
  26:  C ( 2 ) = { C 3 ( 2 ) , C 4 ( 2 ) , C 5 ( 2 ) }
  27: “ReLU and L2 normalization
  28:  C ¯ ( 1 ) = L2_norm ( max ( C ( 1 ) , 0 ) )
  29:  C ¯ ( 2 ) = L2_norm ( max ( C ( 2 ) , 0 ) )
Ensure: Correlation maps C ¯ ( 1 ) and C ¯ ( 2 ) of I ( 1 ) and I ( 2 )
In Algorithm 1, each group of the feature map, extracted from the same layer of feature extractor, will through the operation of L2-normalization and SE blocks. And then, for each two feature maps from the same layer, we utilize correlation computation and truncation operations to get two pairs of correlation maps, i.e., C ¯ n ( 12 ) , C ¯ n ( 11 ) and C ¯ n ( 21 ) , C ¯ n ( 22 ) are produced from F ¨ n ( 1 ) and F ¨ n ( 2 ) (n ∈ {1, 2, 3}). Next, concatenating each group of correlation maps, we get two correlation maps, i.e., C n ( 1 ) and C n ( 2 ) (n ∈ {1, 2, 3}). Last, correlation maps C ( 1 ) , C ( 2 ) are generated by concatenating C n ( 1 ) and C n ( 2 ) respectively. Since the associated areas should have the same symbol and should all be positive, we employ ReLU to convert negative values to zero. And then apply L2-normalization to access the normalized correlation maps as C ¯ ( 1 ) , C ¯ ( 2 ) . Next, the process of handling these correlation maps will be shown in the next section.

3.3. Mask Generator Based on Attention-Based Separable Convolutional Module

Our mask generator integrates ASPP blocks, maxpooling layers and attention-based separable convolutional blocks to generate high-resolution masks. The architecture and parameter settings are shown in Figure 4. It consists of an ASPP block, three upsampling layers, three attention-based separable convolutional blocks and a 1 × 1 convolution to reduce channels at last. Additionally, each attention-based separable convolutional block is followed by an L2-normalization layer.
Briefly, ASPP contains several atrous convolution for conducting multiscale objects. As shown in Figure 5, the attention-based separable convolutional block is applied to generate fine-grained masks, including two layers of depthwise separable convolution and one layer of spatial attention. Depthwise separable convolution can improve the speed and accuracy for deep matching to locate and discriminate regions with less parameters and computation burden. Spatial attention can assign greater weights to critical sections so that the model can focus more attention on it. Qualitatively, this mask generator shows improvement in its ability to address spatial information problems as well as detect edges and small areas. And quantitatively, it reduces computational complexit with less model parameters. In summary, our SADM typically employs attention-based separable convolutional blocks to gain a good tradeoff between localization and detection performance and computation complexity.
(1)
ASPP.
Before the ASPP block, the correlation computation module produces a feature tensor of size W / 8 × H / 8 × 96. ASPP is the first part of the mask generator with three parallel layers of atrous convolution to capture multiscale features, setting atrous rates with [6, 12, 18]. Then the feature maps are concatened into a 1 × 1 convolution to reduce channels.
(2)
Attention-based separable convolutional block.
The mask generator uses attention-based separable convolutional block for three times, corresponding to the three times of upsampling operations. The design of attention-based separable convolutional block is shown in Figure 5. It employs depthwise separable convolution for two times first and a spatial attention second, utilizing L2-normalization, ReLU among them.
Depthwise separable convolution. The depthwise separable convolution is shown in Figure 6. Motivated by the architecture of [31], we adopt a variant of depthwise separable convolution for our mask generator. Different from ecumenical depthwise separable convolution, we utilize the 1 × 1 pointwise convolution followed by the 3 × 3 depthwise convolution for each input channel and concatenate into the subsequent layers. We apply it to mask generator to reduce the model parameters uttermost. Similar to the Xception network [31], we employ L2-normalization and ReLU between pointwise convolution and depthwise convolution.
Spatial attention. Spatial attention can capture the spatial dependencies and produce more powerful pixel-level characterization, helping recover the spatial detail information effectively. The detail of spatial attention is shown in Figure 6. Let P denote the feature map input to spatial attention blocks, P(i, j) denotes a c-dimensional descriptor at at (i, j). Noted that P R h × w × c , i ∈ [1, h], j ∈ [1, w], h and w indicate the height and width of the feature map and h = w in our work. Prior to reinforce P using spatial attention, we employ L2-normalization and ReLU to modify it. Then, the first step is to transform P into two feature spaces by the 1 × 1 convolution layer, f(P) = PW f + b f and g(P) = PW g + b g . The similarity between f(P) and each g(P) is calculated as follows:
s ( x , y ) = f ( P ( x ) ) T g ( P ( y ) )
In order to normalize these weights, we use a softmax function:
β ( x , y ) = e x p ( s ( x , y ) ) y e x p ( s ( x , y ) )
when predicting the xth region, β ( x , y ) denotes the extent that the model attends to the yth location, x, y ∈ [1, h × w]. The final attention is implemented as a 1 × 1 convolutional layer and computed as:
o ( x ) = y β x y h ( P ( y ) )
In the above equation, h ( P ) = P W h + b h . And W f R c × c 8 , W h R c × c , W g R c × c 8 , b f R c 8 , b g , h R c 8 , b h R c . After attention reinforced, the feature maps is calculated as:
F ¨ = A t t e n ( P ) = λ O + P
where O = { o ( 1 ) , o ( 2 ) ,⋯, o ( h × w ) }, λ represents a scale parameter that can be initialized as zero and gradually learn a proper value.

3.4. Pyramid Version of SADM

The localization of object boundaries is vital for tampering detection. Specially, when multiscale objects appear in the image, it will increases the probe sophistication of tampered regions. Score maps play a crucial role in the CISDL task as it can reliably predict the presence and rough positioning of objects. But, score maps are not sensitive to pinpointing the exact outline of the tampered area. In this regime, explicitly accounting for object boundaries across different scales makes an essential contribution to CISDL’s successful handling of large and small objects. There are two main types of study designs to solve multiscale objects prediction challenge. The first approach is to train the model on a dataset that adapts to certain types of transformations, such as shift, rotation, scale, luminance and deformation changes. By applying multiple parallel filters with different rates, the second approach is to harness ASPP to exploit multiscale features. These two approaches have displayed an excellent capacity to represent scale.
In this paper, we employ an alternative method, called the pyramid version of SADM (PSADM), to handle this problem. First, feeding multiple rescaled versions of the original image to parallel module branches with the same parameters. Second, every scale score maps are bilinear interpolated to the original image resolution, converting image classification networks into dense feature extractors without learning any additional parameters, making the training speed of CNN in practice faster. Bilinear interpolation requires two linear transformations, the first one on the X-axis:
f ( M 1 ) = x 2 x x 2 x 1 f ( M 11 ) x x 1 x 2 x 1 f ( N 21 )
f ( M 2 ) = x 2 x x 2 x 1 f ( N 12 ) x x 1 x 2 x 1 f ( N 22 )
where M 1 = ( x , y 1 ) , M 2 = ( x , y 2 ) , N 11 = ( x 1 , y 1 ) , N 21 = ( x 2 , y 1 ) , N 12 = ( x 1 , y 2 ) , N 22 = ( x 2 , y 2 ) . Then find the target point in the region by another linear transformation:
f ( x , y ) = y 2 y y 2 y 1 f ( M 1 ) y y 1 y 2 y 1 f ( M 2 )
Third, fuse them by taking the average response across scales for each position separately. Finally, a new score map is got. The discussion in the experimental section shows that the pyramid version with scale = {384, 512, 640} achieves the best performance.

4. Experiment

In this section, we demonstrate the superiority of our model from an experimental point of view. In the task of CISDL, the most challenging are (1) processing arbitrary-sized images; (2) finding spliced regions as well as pinpointing their exact boundaries under various transformations; and (3) processing the multiscale object in the same image. Based on the above challenges, the enhanced effect of the model on tampering area localization and detection was verified by visual comparisons and quantitative results for the three improvements.

4.1. Benchmark Datasets and Compared Methods

To evaluate the effectiveness of SADM and PSADM, we conducted localization and detection experiments of tampered regions. According to the characteristics of each dataset, we evaluate the localization performance on the generated dataset from MS COCO and verify that SADM outperforms all the previous approaches. Then, the paired CASIA dataset and Media Forensics Challenge 2018 (MFC2018) dataset are tested to demonstrate the superiority of the proposed model in terms of detection performance.
(1)
The generated datasets from MS COCO.
The MS COCO dataset consists of 82,783 training images and 40,504 testing images. It provides object annotations for abundant images and can generate accurate ground truth masks by the enormous number of training pairs, which is suitable for localization performance experiments. After synthesis of the MS COCO dataset, thethe generated dataset’s training sets consist of 1035,255 training pairs, of which one-third were foreground pairs, one-third were background pairs, and one-third were negative sample pairs. The generated dataset’s test sets were divided into three main groups, namely the Difficult set (1–10%), the Normal set (10–25%) and the Easy set (25–50%). We adopt the pixel-level IoU (Intersection over Union), NMM (Nimble Mask Metric) and MCC (Matthews Correlation Coefficient) of the tampered regions and the average indicator of all the tested image pairs to evaluate the localization performance.
(2)
The paired CASIA dataset.
The CASIA TIDEv2.0 dataset is initially designed for the image of copy-move and the classic splicing detection problem [6]. The new paired CASIA dataset we used consists of 3642 positive samples and 5000 negative samples selected from the CASIA TIDEv2.0 dataset by CISDL. Due to the lack of ground truth masks, this new paired CASIA dataset only focuses on estimating the splicing detection performance [6]. We adopt F1-score, precision and recall to evaluate the detection performance.
(3)
MFC2018.
The Media Forensics Challenge 2018 (MFC2018) dataset has 16,673 negative image pairs and 1327 positive image pairs. MFC2018 is a challenging dataset, collected particularly for two problems of the CISDL task: quantitative evaluation of detection performance and localization performance, using a large number of negative image pairs and ground truth by visual comparison. We use AUC, an official metric, to quantify the capability to distinguish between the two categories and employ the EER (Equal Error Rate) score to evaluate false-alarm images.

4.2. Training Procedure and Testing Set

This model is trained with a 1 × 10−6 base learning rate for 3 epochs and 18 samples for each batch size. To calculate the model loss, we use the BCEWithLogitsLoss function. Meanwhile, we employ the Adadelta optimizer to optimize model parameters. Additionally, our feature extractor uses a migration learning approach by using a pre-trained model of VGG16 [10]. Since the scale of objects in the image is arbitrary, a fixed size perceptual field will limit the localization and detection effect, and ability to recognize object boundaries. To address these problems, we use diverse groups of different scale images to find the best formal of our model. The scale of each input image is experimented with [384, 512, 640], [448, 512, 576], respectively, called as PSADM. We adopt postfix to annotate different strategies, such as “ [384, 512, 640]” denotes the rescaled version of original images, and the size of original images are identified by “256/384/512”.
Experiments demonstrate that the localization performance is improved to a new level after adopting this strategy.

4.3. Experimental Results

4.3.1. The Generated Datasets from MS COCO

(1)
Parametric test.
Truncation error. We adopted a truncation operation in the scale-adaptive correlation calculation module to process arbitrary-sized images. After sorting the response values for the feature maps generated from SE-blocks, we tested the response values for the first 32, 128 and 256 channels. The test results are shown in Table 1. The model has the best localization performance when the response values of the first 32 channels are taken. As the response value decreases, the feature it represents becomes weaker. Therefore taking too many of these response values is not conducive to feature extraction.
The comparision of L1-normalization and L2-normalization. Based on the determined parameters of the truncation operation, we tested the types of normalization used in the model, including L1-normalization and L2-normalization, as shown in Table 2. There is little difference in the localization effect between using L1-normalization and L2-normalization, with L2-normalization showing slightly better results. Combining the above two points, SADM indicates that the truncation operation takes the first 32 channels and uses L2-normalization in the subsequent article.
(2)
Localization performance.
Table 3 shows the localization performance on the generated datasets. As for processing the image of 256 × 256, the SADM significantly improves localization capability compared with previous models such as AttentionDM. In [7], Liu et al. employed a sliding window strategy to compensate for the defect of only being able to handle fixed-size images. Yet, it consumes more computing time. To compare with the sliding window strategy, we evaluate the effectiveness of the scale-adaptive correlation computation module on the same dataset. The results of handling 256 × 256, 384 × 384, 512 × 512 are shown in Table 3. As we can see, it is obvious that IoU, MCC and NMM are rising, which thoroughly explains the strong ability of SADM to process arbitrary-sized images. To further improve the capability of processing large-scale images, many kinds of parameters of the PSADM have been experimented with, as shown in Table 3. Under various considerations, the PSADM-512-[384, 512, 640] achieves superior performance on recognizing small regions and pinpointing their exact outline. And then, we will test the detection ability of this version.
(3)
Complexity analyses.
Table 4 lists the testing time, parameters numbers and implemented frameworks. All experiments were conducted on a machine with Intel(R) Core (TM) i7-5930K CPU @ 3.50 GHz, 64 GB RAM and a single GPU (TITAN X). As shown in Table 4, the number of the trainable parameters for SADM is 15,407,276, slightly more than the number of parameters for DMAC but less than AttentionDM. Although the testing time of SADM is 0.0298 s, marginally lower than that of DMAC, it achieves significant improvements in localization and detection.

4.3.2. The Paired CASIA Dataset

In [9], AttentionDM was compared with DMVN and DMAC. In the comparison on CASIA, the previous scores are from [9]. In this paper, we calculate the average score of the tampered area as the detection score (here called tampering probability). In other words, we calculate the average score of the detected regions with {s ( k ) |k = 1, 2} for each generated mask firstly. And then the mean value, (s ( 1 ) + s ( 2 ) ) / 2, is computed as the final forged probability. As shown in Table 5, SADM has a very high Precision (nearly 100%) and a little low Recall value. For the size of 256 × 256, SADM improves the precision from 92.88% to 99.01%, while the recall score only decreases by 3.87%. F1-score has been further improved, which is a good tradeoff between Precision and Recall. As for the large-size such as 512 × 512, these indicators are further enhanced. In addition, the detection performance of large-scale images is further enhanced by the processing of PSADM. Visual comparisons, afforded in Figure 7, show that SADM achieves very good performance. It has a high level of competence in detecting small regions and accurate boundaries. Furthermore, it is also sensitive to the arbitrary-sized images as well as robust to transformation and rotation changes.

4.3.3. MFC2018

Since MFC2018 is definitely collected for the CISDL task, we compare SADM, PSADM with DMVN, DMAC and AttentionDM on MFC2018. The MFC2018 challenge provides the evaluation codes, AUC and EER scores, which are calculated and shown in Table 6 [36]. Compared with AttentionDM [9], SADM achieves the lowest EER score for 256 × 256 images with an AUC decreased by 0.014, which far exceeded the previous CISDL approach. As for large-scale images, SADM can also achieve higher AUC and lower EER. Additionally, PSADM further enhance the effect of processing large-scale images, with AUC increased by 0.0037 and EER decreased by 0.0019 based on “SADM-512”. Figure 8 provides visual comparisons. Obviously, SADM has a better capability to detect small regions and accurate boundaries than previous methods.

5. Conclusions

In this paper, we propose a Scale-Adaptive Deep Matching network (SADM) for CISDL, and a pyramid version of SADM (PSADM). SADM consists of three components: the feature extractor, the scale-adaptive correlation computation module and the mask generator. The correlation-computation with truncation operations are proposed to deal with arbitrary-sized images. The mask generator is designed to reconstruct the spatial information of an image and generate fine-grained mask without requiring more computational complexity. The PSADM is applied to improve the ability of multiscale object detection and matching. Experimental results show that the proposed method can achieve better performance than the state-of-the-art methods on the publicly available datasets.

Author Contributions

Conceptualization, S.X. and Y.L.; methodology, S.L. and Y.L.; software, S.L. and Y.L.; validation, S.L.; formal analysis, S.L., Y.L. and C.X.; investigation, S.L.; data curation, S.L.; writing—original draft preparation, S.L.; writing—review and editing, S.X., S.L., Y.L., C.X. and N.G.; visualization, S.L.; funding acquisition, Y.L. and C.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China grant number 62102010 and 62002003.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data generated or analyzed during this study are included in this published article.

Acknowledgments

This work was supported by NSFC under 62102010 and 62002003.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CISDLConstrained image splicing detection and localization
DMVNDeep matching and validation network
FPLNFeature pyramid deep matching and localization network
DMACDeep matching network based on atrous convolution
DMAC-advAdversarial learning framework of DMAC
AttentionDMAttention-aware deep matching network
SADMScale-adaptive Deep Matching network
PSADMPymaid Scale-adaptive Deep Matching network

References

  1. Li, C.; Ma, Q.; Xiao, L.; Li, M.; Zhang, A. Image splicing detection based on Markov features in QDCT domain. Neurocomputing 2017, 228, 29–36. [Google Scholar] [CrossRef]
  2. Tang, W.; Li, B.; Tan, S.; Barni, M.; Huang, J. CNN-based adversarial embedding for image steganography. IEEE Trans. Inf. Forensics Secur. 2019, 14, 2074–2087. [Google Scholar] [CrossRef] [Green Version]
  3. Matern, F.; Riess, C.; Stamminger, M. Gradient-based illumination description for image forgery detection. IEEE Trans. Inf. Forensics Secur. 2019, 15, 1303–1317. [Google Scholar] [CrossRef]
  4. Cozzolino, D.; Verdoliva, L. Noiseprint: A CNN-Based Camera Model Fingerprint. IEEE Trans. Inf. Forensics Secur. 2020, 15, 144–159. [Google Scholar] [CrossRef] [Green Version]
  5. Liu, Y.; Guan, Q.; Zhao, X.; Cao, Y. Image Forgery Localization Based on Multi-Scale Convolutional Neural Networks. In Proceedings of the 6th ACM Workshop on Information Hiding and Multimedia Security, Innsbruck, Austria, 20–22 June 2018; pp. 85–90. [Google Scholar]
  6. Wu, Y.; Abd-Almageed, W.; Natarajan, P. Deep Matching and Validation Network: An End-to-End Solution to Constrained Image Splicing Localization and Detection. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 1480–1502. [Google Scholar]
  7. Liu, Y.; Zhu, X.; Zhao, X.; Cao, Y. Adversarial Learning for Constrained Image Splicing Detection and Localization Based on Atrous Convolution. IEEE Trans. Inf. Forensics Secur. 2019, 14, 2551–2566. [Google Scholar] [CrossRef]
  8. Ye, K.; Dong, J.; Wang, W.; Peng, B.; Tan, T. Feature Pyramid Deep Matching and Localization Network for Image Forensics. In Proceedings of the 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Honolulu, HI, USA, 12–15 November 2018; pp. 1796–1802. [Google Scholar] [CrossRef]
  9. Liu, Y.; Zhao, X. Constrained Image Splicing Detection and Localization With Attention-Aware Encoder-Decoder and Atrous Convolution. IEEE Access 2020, 8, 6729–6741. [Google Scholar] [CrossRef]
  10. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  11. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; IEEE Computer Society: Los Alamitos, CA, USA, 2018; pp. 7132–7141. [Google Scholar] [CrossRef] [Green Version]
  12. Zeng, H.; Kang, X.; Peng, A. A multi-purpose countermeasure against image anti-forensics using autoregressive model. Neurocomputing 2016, 189, 117–122. [Google Scholar] [CrossRef]
  13. Xiao, D.; Wang, Y.; Xiang, T.; Bai, S. High-payload completely reversible data hiding in encrypted images by an interpolation technique. Frontiers Inf. Technol. Electron. Eng. 2017, 18, 1732–1743. [Google Scholar] [CrossRef]
  14. Yang, F.; Zhang, W.; Tao, L.; Ma, J. Transfer Learning Strategies for Deep Learning-based PHM Algorithms. Appl. Sci. 2020, 10, 2361. [Google Scholar] [CrossRef] [Green Version]
  15. Zhang, S.X.; Zhu, X.; Yang, C.; Wang, H.; Yin, X.C. Adaptive Boundary Proposal Network for Arbitrary Shape Text Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 1305–1314. [Google Scholar]
  16. Li, C.; Wei, F.; Dong, W.; Wang, X.; Yan, J.; Zhu, X.; Liu, Q.; Zhang, X. Spatially Regularized Streaming Sensor Selection. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 3871–3879. [Google Scholar]
  17. Miah, A.S.M.; Shin, J.; Hasan, M.A.M.; Rahim, M.A. BenSignNet: Bengali Sign Language Alphabet Recognition Using Concatenated Segmentation and Convolutional Neural Network. Appl. Sci. 2022, 12, 3933. [Google Scholar] [CrossRef]
  18. Alshowaish, H.; Al-Ohali, Y.; Al-Nafjan, A. Trademark Image Similarity Detection Using Convolutional Neural Network. Appl. Sci. 2022, 12, 1752. [Google Scholar] [CrossRef]
  19. Zhu, X.; Li, Z.; Lou, J.; Shen, Q. Video super-resolution based on a spatio-temporal matching network. Pattern Recognit. 2021, 110, 107619. [Google Scholar] [CrossRef]
  20. Nam, J.; Kang, J. Classification of Chaotic Signals of the Recurrence Matrix Using a Convolutional Neural Network and Verification through the Lyapunov Exponent. Appl. Sci. 2021, 11, 77. [Google Scholar] [CrossRef]
  21. Li, C.; Zhen, T.; Li, Z. Image Classification of Pests with Residual Neural Network Based on Transfer Learning. Appl. Sci. 2022, 12, 4356. [Google Scholar] [CrossRef]
  22. Zhu, X.; Li, Z.; Li, X.; Li, S.; Dai, F. Attention-aware perceptual enhancement nets for low-resolution image classification. Inf. Sci. 2020, 515, 233–247. [Google Scholar] [CrossRef]
  23. Tang, C.; Ling, Y.; Yang, X.; Jin, W.; Zheng, C. Multi-View Object Detection Based on Deep Learning. Appl. Sci. 2018, 8, 1423. [Google Scholar] [CrossRef] [Green Version]
  24. Zhang, S.X.; Zhu, X.; Chen, L.; Hou, J.B.; Yin, X.C. Arbitrary Shape Text Detection via Segmentation with Probability Maps. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 1. [Google Scholar] [CrossRef]
  25. Li, P.; Xia, H.; Zhou, B.; Yan, F.; Guo, R. A Method to Improve the Accuracy of Pavement Crack Identification by Combining a Semantic Segmentation and Edge Detection Model. Appl. Sci. 2022, 12, 4714. [Google Scholar] [CrossRef]
  26. Dang, T.V.; Yu, G.H.; Kim, J.Y. Revisiting Low-Resolution Images Retrieval with Attention Mechanism and Contrastive Learning. Appl. Sci. 2021, 11, 6783. [Google Scholar] [CrossRef]
  27. Gu, Y.; Wang, Y.; Li, Y. A Survey on Deep Learning-Driven Remote Sensing Image Scene Understanding: Scene Classification, Scene Retrieval and Scene-Guided Object Detection. Appl. Sci. 2019, 9, 2110. [Google Scholar] [CrossRef] [Green Version]
  28. Korus, P. Digital image integrity—A survey of protection and verification techniques. Digit. Signal Process. 2017, 71, 1–26. [Google Scholar] [CrossRef]
  29. Wu, Y.; Abd-Almageed, W.; Natarajan, P. BusterNet: Detecting Copy-Move Image Forgery with Source/Target Localization. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; pp. 170–186. [Google Scholar]
  30. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–19. [Google Scholar]
  31. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Los Alamitos, CA, USA, 2017; pp. 1800–1807. [Google Scholar]
  32. Christlein, V.; Riess, C.; Jordan, J.; Riess, C.; Angelopoulou, E. An Evaluation of Popular Copy-Move Forgery Detection Approaches. IEEE Trans. Inf. Forensics Secur. 2012, 7, 1841–1854. [Google Scholar] [CrossRef] [Green Version]
  33. Luo, W.; Huang, J.; Qiu, G. Robust Detection of Region-Duplication Forgery in Digital Image. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; pp. 746–749. [Google Scholar] [CrossRef]
  34. Ryu, S.J.; Lee, M.J.; Lee, H.K. Detection of Copy-Rotate-Move Forgery Using Zernike Moments. In Information Hiding; Springer: Berlin/Heidelberg, Germany, 2010; pp. 51–65. [Google Scholar]
  35. Cozzolino, D.; Poggi, G.; Verdoliva, L. Efficient dense-field copy–move forgery detection. IEEE Trans. Inf. Forensics Secur. 2015, 10, 2284–2297. [Google Scholar] [CrossRef]
  36. Guan, H.; Kozak, M.; Robertson, E.; Lee, Y.; Yates, A.N.; Delgado, A.; Zhou, D.; Kheyrkhah, T.; Smith, J.; Fiscus, J. MFC Datasets: Large-Scale Benchmark Datasets for Media Forensic Challenge Evaluation. In Proceedings of the 2019 IEEE Winter Applications of Computer Vision Workshops (WACVW), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 63–72. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed SADM. The probe P and potential donor image D are input to the network. Three groups of feature maps with the same size are generated from the feature extractor. They are processed with the scale-adaptive correlation computation module composed of L2-Normalization, SE block, correlation-computation and ReLU. It generates two correlation maps, which are fed into the mask generator with attention-based separable convolutional blocks to generate two fine-grained masks Pm and Dm.
Figure 1. Overview of the proposed SADM. The probe P and potential donor image D are input to the network. Three groups of feature maps with the same size are generated from the feature extractor. They are processed with the scale-adaptive correlation computation module composed of L2-Normalization, SE block, correlation-computation and ReLU. It generates two correlation maps, which are fed into the mask generator with attention-based separable convolutional blocks to generate two fine-grained masks Pm and Dm.
Applsci 12 06480 g001
Figure 2. Parameter settings of feature extractor. “3 × 3” represents the kernel size of convolutional layers, “64”, “128” and “512” stand for the number of filters, and “AC” indicates the default setting, r s = 2 for atrous convolutional layers.
Figure 2. Parameter settings of feature extractor. “3 × 3” represents the kernel size of convolutional layers, “64”, “128” and “512” stand for the number of filters, and “AC” indicates the default setting, r s = 2 for atrous convolutional layers.
Applsci 12 06480 g002
Figure 3. Truncation operations of Scale-Adaptive Correlation Computation module. F ¨ ( 1 ) , F ¨ ( 2 ) are generated from SE blocks. “Top_32 ” means that we sort each correlation map along the H / 8 × W / 8 channels to select Top_32 values.
Figure 3. Truncation operations of Scale-Adaptive Correlation Computation module. F ¨ ( 1 ) , F ¨ ( 2 ) are generated from SE blocks. “Top_32 ” means that we sort each correlation map along the H / 8 × W / 8 channels to select Top_32 values.
Applsci 12 06480 g003
Figure 4. Parameter settings of mask generator architecture. “3 × 3” and “[1, 3]” indicate the kernel size used in convolution layers. “AC” means the atrous convolution, and “6”, “12”, “18” stand for the r s of atrous convolution. “480”, “96”, “48”, “16” represent the input or output channels of each layer.
Figure 4. Parameter settings of mask generator architecture. “3 × 3” and “[1, 3]” indicate the kernel size used in convolution layers. “AC” means the atrous convolution, and “6”, “12”, “18” stand for the r s of atrous convolution. “480”, “96”, “48”, “16” represent the input or output channels of each layer.
Applsci 12 06480 g004
Figure 5. Attention-based separable convolutional block, where “m” and “n” denote the number of input and output channles of depthwise separable convolution, respectively. The “n” also represents the number of input and output channels of spatial attention.
Figure 5. Attention-based separable convolutional block, where “m” and “n” denote the number of input and output channles of depthwise separable convolution, respectively. The “n” also represents the number of input and output channels of spatial attention.
Applsci 12 06480 g005
Figure 6. Depthwise separable convolution and spatial attention. (a) The structure of depthwise separable convolution and (b) spatial attention. “(m,n) @ [1,3]” denotes the depthwise separable convolution using 1 × 1 depthwise convolution for m times and 3 × 3 point convolution for n times. “(n, n) @ [1,1]” represents the number of input and output channels of convolution layers and “[1, 1]” means all convolution filters are size of 1 × 1.
Figure 6. Depthwise separable convolution and spatial attention. (a) The structure of depthwise separable convolution and (b) spatial attention. “(m,n) @ [1,3]” denotes the depthwise separable convolution using 1 × 1 depthwise convolution for m times and 3 × 3 point convolution for n times. “(n, n) @ [1,1]” represents the number of input and output channels of convolution layers and “[1, 1]” means all convolution filters are size of 1 × 1.
Applsci 12 06480 g006
Figure 7. Visual comparisons on the paired CASIA dataset.
Figure 7. Visual comparisons on the paired CASIA dataset.
Applsci 12 06480 g007
Figure 8. Visual comparisons on the MFC2018 dataset.
Figure 8. Visual comparisons on the MFC2018 dataset.
Applsci 12 06480 g008
Table 1. Truncation error of SADM.
Table 1. Truncation error of SADM.
MethodDiffcultNormalEasy
IoUMCCNMMIoUMCCNMMIoUMCCNMM
SADM32-L20.77590.81280.51290.90400.82880.82650.96210.96160.9410
SADM128-L20.76490.81150.50570.89810.82600.82660.94500.95630.9452
SADM256-L20.76020.82400.49870.89370.82440.81660.94340.94880.9246
Table 2. The comparision of L1-Normalization an L2-Normalization.
Table 2. The comparision of L1-Normalization an L2-Normalization.
MethodDiffcultNormalEasy
IoUMCCNMMIoUMCCNMMIoUMCCNMM
SADM32-L10.77020.83400.51670.89370.92440.81660.94840.96880.9246
SADM32-L20.77590.81280.51290.90400.82880.82650.96210.96160.9410
Table 3. Localization performance comparison on the generated datasets from MS COCO of SADM and PSADM.
Table 3. Localization performance comparison on the generated datasets from MS COCO of SADM and PSADM.
MethodDiffcultNormalEasy
IoUMCCNMMIoUMCCNMMIoUMCCNMM
DMVN [6]0.27220.3533−0.43820.68180.75700.40420.81980.85440.6770
DMAC [7]0.51140.63080.03350.82790.88150.68400.92220.93950.8685
DMAC-adv [7]0.54330.65840.10260.83170.88330.68770.92370.94110.8655
AttentionDM [9]0.72280.81080.47930.89800.93200.82530.96020.96030.9388
SADM0.77590.81280.51290.90400.82880.82650.96210.96160.9410
SADM-3840.78280.84420.50600.89910.93210.84010.96210.97070.9495
SADM-5120.78290.85210.58570.90420.92230.92810.96250.96630.9552
PSADM-avg-[384, 512, 640]0.80890.87460.60700.92470.95340.94030.97380.99080.9761
PSADM–avg-[448, 512, 576]0.78630.85230.58590.91110.93610.93970.96260.97960.9588
PSADM-max-[384, 512, 640]0.79310.86320.59610.89930.95690.92990.98290.98250.9632
PSADM-max-[448, 512, 576]0.78490.85140.57930.89760.93470.92870.95980.96930.9605
Table 4. Time complexity analyses.
Table 4. Time complexity analyses.
MethodTime/sParametersFramework
DMVN [6]0.296810,473,788Keras/Theano
DMAC/DMAC-adv [7]0.028814,920,520PyTorch
AttentionDM [9]0.030615,758,162PyTorch
SADM0.029815,407,276PyTorch
Table 5. Comparisons on the paired CASIA dataset.
Table 5. Comparisons on the paired CASIA dataset.
MethodPrecisionRecallF1-Score
Christlein et al. [32]0.51640.82920.6364
Luo et al. [33]0.99690.53530.6966
Ryu et al. [34]0.96140.58590.7309
Cozzolino et al. [35]0.98970.63340.7725
DMVN-loc [6]0.91520.79180.8491
DMVN-det [6]0.94150.79080.8596
DMAC [7]0.92550.86680.8952
DMAC-adv [7]0.96570.85760.9085
AttentionDM [9]0.92880.92040.9246
SADM-2560.97410.87580.9263
SADM-3840.98350.87710.9314
SADM-5120.99030.87780.9316
PSADM-512-[384, 512, 640]0.99320.89240.9329
Table 6. Comparisons on MFC2018.
Table 6. Comparisons on MFC2018.
MethodAUCEER
DMAC [7]0.75420.3123
DMAC-adv [7]0.75110.3093
AttentionDM [9]0.79220.2756
SADM-2560.77820.2527
SADM-3840.78250.2490
SADM-5120.79450.2470
PSADM-512-[384, 512, 640]0.79820.2489
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, S.; Lv, S.; Liu, Y.; Xia, C.; Gan, N. Scale-Adaptive Deep Matching Network for Constrained Image Splicing Detection and Localization. Appl. Sci. 2022, 12, 6480. https://doi.org/10.3390/app12136480

AMA Style

Xu S, Lv S, Liu Y, Xia C, Gan N. Scale-Adaptive Deep Matching Network for Constrained Image Splicing Detection and Localization. Applied Sciences. 2022; 12(13):6480. https://doi.org/10.3390/app12136480

Chicago/Turabian Style

Xu, Shengwei, Shanlin Lv, Yaqi Liu, Chao Xia, and Nan Gan. 2022. "Scale-Adaptive Deep Matching Network for Constrained Image Splicing Detection and Localization" Applied Sciences 12, no. 13: 6480. https://doi.org/10.3390/app12136480

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop