Next Article in Journal
Research on Progress of Blockchain Consensus Algorithm: A Review on Recent Progress of Blockchain Consensus Algorithms
Next Article in Special Issue
A Vote-Based Architecture to Generate Classified Datasets and Improve Performance of Intrusion Detection Systems Based on Supervised Learning
Previous Article in Journal
Topology Inference and Link Parameter Estimation Based on End-to-End Measurements
Previous Article in Special Issue
Models versus Datasets: Reducing Bias through Building a Comprehensive IDS Benchmark
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Framework of Cross-Domain and Model Adversarial Attack against Deepfake

College of Information and Cyber Security, People’s Public Security University of China, Beijing 100038, China
*
Author to whom correspondence should be addressed.
Future Internet 2022, 14(2), 46; https://doi.org/10.3390/fi14020046
Submission received: 21 December 2021 / Revised: 28 January 2022 / Accepted: 28 January 2022 / Published: 29 January 2022
(This article belongs to the Special Issue Machine Learning Integration with Cyber Security)

Abstract

:
To protect images from the tampering of deepfake, adversarial examples can be made to replace the original images by distorting the output of the deepfake model and disrupting its work. Current studies lack generalizability in that they simply focus on the adversarial examples generated by a model in a domain. To improve the generalization of adversarial examples and produce better attack effects on each domain of multiple deepfake models, this paper proposes a framework of Cross-Domain and Model Adversarial Attack (CDMAA). Firstly, CDMAA uniformly weights the loss function of each domain and calculates the cross-domain gradient. Then, inspired by the multiple gradient descent algorithm (MGDA), CDMAA integrates the cross-domain gradients of each model to obtain the cross-domain perturbation vector, which is used to optimize the adversarial example. Finally, we propose a penalty-based gradient regularization method to pre-process the cross-domain gradients to improve the success rate of attacks. CDMAA experiments on four mainstream deepfake models showed that the adversarial examples generated from CDMAA have the generalizability of attacking multiple models and multiple domains simultaneously. Ablation experiments were conducted to compare the CDMAA components with the methods used in existing studies and verify the superiority of CDMAA.

1. Introduction

Deepfake [1] constructs generator models based on generative adversarial networks (GANs) to forge images. Receiving real images as input, the deepfake model can output fake images by, for example, changing hair color. Deepfake has played an important role in the entertainment and culture industry, bringing many conveniences to life and work. However, malicious users may take advantage of this technology to produce fake videos and news, misleading face recognition systems and seriously disrupting the social order [2,3].
In order to cope with deepfake tampering with images, a large number of studies focus on constructing deepfake detection models [4,5,6,7,8,9,10], which can detect whether an image is faked. However, this technology can only ensure the authenticity of the image, but it cannot guarantee the integrity of the image. Moreover, even if an image is confirmed a fake, negative impacts are caused on the people concerned or on society because the image has already been widely circulated. More direct interventions should therefore be taken to ensure that images are not tampered with though deepfake from the source.
Some studies propose the use of the adversarial attack [11] to interfere with the work of the deepfake model. The main idea of the adversarial attack is adding a perturbation, imperceptible to the naked eye, to the original example, generating the adversarial example, which can mislead the deep learning models to produce a quite different output. The adversarial attack was originally used to destroy security systems such as face recognition, which posed a huge challenge to the security of deep learning models. However, if the object of an adversarial attack is turned into a malicious model such as through deepfake, the meaning of the adversarial attack becomes dramatically opposite: to disrupt the normal operation of malicious models to guarantee information security. As shown in Figure 1, the specific operation involves adding a perturbation, imperceptible to the naked eye, to an image when users post the image online so that when the attacker obtains the image, the fake version generated through deepfake will have obvious distortions or deformations, which can be easily identified as forgery.
In the current study, however, the generalization of the adversarial attack against the deepfake model is very limited: an adversarial example generated for a specific deepfake model is unable to produce equal attack effect on other models [12]; furthermore, even in the same model, an adversarial example generated in a particular domain cannot achieve effective attack in other domains [13] (by setting the corresponding conditional variables, deepfake models can generate multiple domains of forged images, such as the hair color or the gender of the face image). Without the knowledge of what deepfake model will be employed or what conditional variables will be set to tamper with images, the adversarial attack methods currently studied have great limitations in practice.
In order to improve the generalization of the adversarial attack, that is, to generate the adversarial samples in each domain of multiple models of given images, this paper proposes a framework of Cross-Domain and Model Adversarial Attack (CDMAA): Any gradient-based adversarial example generation algorithm can be used for an adversarial attack, such as the I-FGSM [14]. In the backpropagation phase, the algorithm uniformly weights the loss function with different condition variables in the model to extend the generalization of the adversarial example between various domains. The Multiple Gradient Descent Algorithm (MGDA) [15] is used to calculate the weighted sum of the gradients of each model to ensure the generalization of adversarial examples between various models. Finally, we propose a penalty-based gradient regularization method to further improve the success rate of adversarial attacks. CDMAA can expand the attack range of the generated adversarial example and ensure that the images are not tampered with and forged by multiple deepfake models.

2. Related Work

According to the category of model input, some deepfake models input random noise to synthesize images which were entirely non-existent before [16], such as ProGAN [17], StyleGAN [18], etc. Other deepfake models input real images to achieve the image translation from domain to domain. For example, StarGAN [19], AttGAN [20] and STGAN [21] can translate the facial images in domains by setting different conditional variables, such as hair color, age, etc. Unsupervised models, such as CycleGAN [22] and U-GAT-IT [23], can only translate images to a single domain, which can be considered a special case of multi-domain translation models with a total domain number of 1. This paper focuses on image translation deepfake models and performs the adversarial attack on them to interfere with their normal functions and protect real images from being tampered with.
The adversarial attack was initially applied to classification models [24]. Goodfellow et al. proposed the Fast Gradient Sign Method (FGSM) [25]. The FGSM sets the distance between the model output of the adversarial example and the model output of the original example as the loss function. The gradient of the loss function with respect to the input indicates a direction where the output difference between the adversarial example and the original example ascends fastest. Therefore, the FGSM adds the gradient in the original example to generate an effective adversarial example. Kurakin et al. proposed the iterative FGSM (I-FGSM) [14], which iteratively performs gradient backpropagation to reduce the step size of updating adversarial examples and improving their efficiency. Many studies have since proposed various adversarial attack algorithms to optimize the efficiency of adversarial attacks, such as PGD [26], which uses random noise to initialize the adversarial examples, the MI-FGSM [27], which uses momentum to update the gradient, and APGD [28], which automatically decreases the step size.
Kos et al. [29] first extended adversarial attacks to generation models. Yeh et al. [30] first proposed to attack the deepfake model. They used PGD to generate adversarial examples against CycleGAN, pix2pix [31], etc., which can distort the output of these models. Lv et al. [32] proposed that higher weight should be given to the face part of the images when calculating the loss function so that the output distortion generated by the adversarial examples is concentrated on the face to achieve a better effect of interfering deepfake models. Dong et al. [33] explored the adversarial attacks on encoder–decoder-based deepfake models and proposed to use the loss function with respect to latent variables in encoders to generate the adversarial examples. These studies generate adversarial examples only for certain models and do not take into account that models can output fake images of different domains by setting different condition variables, so the generalization of adversarial attacks is quite limited.
Ruiz et al. [13] considered the generalizability of adversarial attacks across different domains. They verified that the adversarial example generated in a particular domain cannot achieve effective attack in other domains of the model and proposed two methods of iterative traversal and joint summation to generate adversarial examples that are effective for each domain. However, they did not consider the generalization between different models of the adversarial examples. Since the differences between models are much larger than the differences between domains within models, the simple method of iterative traversal or joint summation cannot be equally effective for attacks between different models.
Fang et al. [34] considered the generalizability of adversarial attacks across models. They verified that the adversarial examples against a particular model are ineffective in attacking other models and proposed a method of weighting the loss functions of multiple models to generate adversarial examples against multiple deepfake models, where the weighting factors are found by a line search. However, the tuning experiments are extremely tedious because the weighting coefficients need to be found in J 1 dimensional parameter space, where J denotes the number of models. In addition, the coefficients need to be adjusted when attack models are changed, which is quite inefficient.
Compared with the existing work, this paper focuses on extending the generalization of adversarial examples across various domains and models and proposes a framework of CDMAA. CDMAA can generate adversarial examples that can attack multiple deepfake models under all condition variables with higher efficiency.

3. CDMAA

In this paper, we use the I-FGSM as the basic adversarial attack algorithm to introduce the CDMAA framework. In the model forward propagation phase, we generate the cross-domain loss function of each model by uniformly weighting the loss function corresponding to each conditional variable. In the phase of model backward propagation to calculate the gradient, we use the MGDA to generate a cross-model perturbation vector from the gradient of each cross-domain loss function. The perturbation vector is used to iteratively update the adversarial example to improve its generalizability across multiple models and domains.

3.1. I-FGSM Adversarial Attack Deepfake Model

Given an original image x , its output of the deepfake model is G ( x , c ) , where G denotes the deepfake model and c denotes the conditional variable. The I-FGSM generates the adversarial example x ˜ by the following steps:
x 0 = x
x t + 1 = c l i p ( x t + a s i g n ( x t L ( G ( x t , c ) , G ( x , c ) ) ) )
where x t denotes the adversarial example after t iterations, t does not exceed the number of iteration steps T, a denotes the step size, s i g n is the symbolic function, ε denotes the perturbation range and the c l i p function restricts the size of the adversarial perturbation not to exceed the perturbation range in l p -norm, i.e.,
x t x ε
so that the difference between the adversarial example and the original image is sufficiently small to ensure that the original image is not significantly modified, L denotes the loss function, which uses the mean squared loss (MSE) to measure the distance between the output of the adversarial example G ( x t , c ) and the output of the original image G ( x , c ) [30]:
L ( G ( x t , c ) , G ( x , c ) ) = 1 D G ( x t , c ) , G ( x , c ) 2
where D denotes the dimensionality of the model output, i.e., the l e n g t h w i d t h c h a n n e l s of the output image.
The adversarial example is updated towards the optimization goal:
x ˜ = arg max x ˜ L ( G ( x ˜ , c ) , G ( x , c ) )
The generated adversarial example is considered to have successfully attacked the deepfake model G under the condition variable c when the loss function L keeps increasing and reaches a certain threshold τ , i.e., the output image has a sufficiently noticeable distortion.

3.2. Cross-Domain Adversarial Attack

To extend the generalizability of the adversarial examples across various domains of the model, i.e., the optimization objective (5) is modified as
x ˜ = arg max x ˜ { L ( G ( x ˜ , c i ) , G ( x , c i ) ) | i = 1 K }
where c i denotes the ith conditional variable of model G and K denotes the total number of conditional variables.
The gradient for each of the loss functions in the above optimization objectives is calculated by
g r a d i = x t L ( G ( x t , c i ) , G ( x , c i ) )
where g r a d i indicates the optimization direction for maximizing the distortion of the output of model G with condition variable c i for the current adversarial example x t .
Since the backbone network is fixed in the model, changing only the condition variables has less impact on the model output, resulting in the loss functions L i and their corresponding gradients g r a d i of different condition variables being very similar, i.e., the g r a d i have approximately the same direction, as shown in Figure 2a. Therefore, we integrate a cross-domain gradient g r a d by simply uniformly weighting g r a d i :
g r a d = 1 K i = 1 K g r a d i = 1 K i = 1 K x t L ( G ( x t , c i ) , G ( x , c i ) )
g r a d is obtained by integrating the loss functions corresponding to each conditional variable so that they indicate a common direction of maximizing the loss function of each domain. Using g r a d to update the adversarial example can achieve the optimization objective of (6).
Consider the following equation:
1 K i = 1 K x t L ( G ( x t , c i ) , G ( x , c i ) ) = x t ( 1 K i = 1 K L ( G ( x t , c i ) , G ( x , c i ) ) ) = x t ( 1 K i = 1 K L i )
That is, we can uniformly weight the loss function L i corresponding to each condition variable c i to obtain a cross-domain loss function i = 1 K L i and then calculate the gradient of it with respect to x t , which is the cross-domain gradient g r a d . It ensures that only one backpropagation step is performed for each model so that time consumption is reduced.

3.3. Cross-Model Adversarial Attack

We further extend the generalizability of the adversarial examples across models, i.e., the optimization objective in (6) is modified:
x ˜ = arg max x ˜ { L ( G ( j ) ( x ˜ , c i ) , G ( j ) ( x , c i ) ) | i = 1 K j ,   j = 1 J }
where G ( j ) denotes the jth deepfake model and J denotes the total number of deepfake models.
The group of cross-domain gradients has been obtained from Section 3.2:
g r a d = { g r a d ( 1 ) , , g r a d ( J ) }
where g r a d ( j ) denotes the cross-domain gradient of the jth model. Considering that these gradients come from different models, the large differences between models lead to a low similarity between the gradients, as shown in Figure 2b. Simply iterative traversing or uniform weighting these gradients is prone to a large fluctuation in the optimization process and generates an ineffective adversarial example [35].
In order to derive a cross-model perturbation vector w from the group of gradients g r a d to update the adversarial example, the CDMAA framework draws on the idea of the Multiple Gradient Descent Algorithm (MGDA) to give an idea for finding w :
w = arg min u U ¯ u 2 2
The space U ¯ that the vector u values in satisfies:
U ¯ = { u = j = 1 J a ( j ) g r a d ( j ) | a ( j ) 0 , j ; j = 1 J a ( j ) = 1 }
Theorem 1.
The solution w in (12) is the optimization direction in which the loss function corresponding to each model is increasing for the current adversarial example, i.e., it satisfies:
w g r a d ( j ) > 0 , j
Proof. 
Equation (12) is equivalent to the following optimization problem:
a 0 = a r g m i n a A ¯ j = 1 J a ( j ) g r a d ( j ) 2 2 ,   s . t . A ¯ = { a = ( a ( 1 ) , , a ( J ) ) + J | j = 1 J a ( j ) = 1 }
To solve this extreme value problem of a multivariate function under the linear constraint, construct the Lagrange function:
L ( a , λ ) = u 2 2 + λ ( j = 1 J a ( j ) 1 )
since u = j = 1 J a ( j ) g r a d ( j ) ; hence u a ( j ) = g r a d ( j ) and
L ( a , λ ) a ( j ) = u 2 2 a ( j ) + λ = 2 u u a ( j ) + λ = 2 u g r a d ( j ) + λ
According to the Lagrange multiplier, the equation
L ( a , λ ) a ( j ) = 0 , j
is a necessary condition for w obtaining the minimum of u ; hence,
w g r a d ( j ) = λ 2 , j
considering that
w 2 2 = w w = w j = 1 J ( a ( j ) g r a d ( j ) ) = j = 1 J a ( j ) w g r a d ( j )
In the actual adversarial attack scenario, since the dimension D is much larger than the number J of the gradients in g r a d , it is almost impossible for these gradients to be linearly dependent. Hence, their linear combination w 0 and
w 2 2 > 0
Uniting j = 1 J a ( j ) = 1 , w g r a d ( j ) = λ 2 , there is
λ 2 > 0
Simultaneously, (19), (20) and (14) are proven. □
Since the vector product of w and the gradient of all model loss functions is positive, optimizing the adversarial example with w ensures the whole improvement of loss functions in each model, i.e., the optimization objective (10), which can improve the generalization of the adversarial examples in various models.

3.4. Gradient Regularization

In Section 3.3, if the gradients group g r a d is regularized as
g r a d n o r ( j ) = g r a d ( j ) / S ( j )
and then the MGDA is used on the regularized gradients group g r a d n o r to find a perturbation vector w , the result of (14) holds because
w g r a d ( j ) = w g r a d n o r ( j ) S ( j ) > 0
where S ( j ) > 0 is the regularization factor.
Common regularization methods include L2 regularization: S ( j ) = g r a d ( j ) 2 , which scales the gradients to the unit vector; logarithmic gradients regularization: S ( j ) = L ( j ) , which reduces the gradients by the factor of their corresponding loss function value.
Due to the large difference in norms of each of the cross-domain gradients g r a d ( j ) which are calculated from various models, the resulting vector w is expected to be mostly influenced by the gradients of small norms in g r a d ( j ) . In addition, without some constraints and guidance methods, the generated adversarial example will form an obvious “attack preference” due to the different vulnerability of deepfake models, only achieving high attack effect on the vulnerable models, which eventually leads to a large difference of different models.
To lead the cross-model perturbation vector in the direction of improving the effectiveness of attacks on models that are not vulnerable to adversarial attacks and maximize the success rate of adversarial attacks against all models, we propose a penalty-based gradient regularization method:
S ( j ) = 1 max ( L ( j ) , ς )
where L ( j ) denotes the cross-domain loss function of the jth model and ς is a very small positive number to prevent the zero-denominator error of S ( j ) when L ( j ) = 0 . (The value of the loss function L ( j ) is 0 inevitably in the first iteration of the I-FGSM since the current adversarial example is the same as the original image).
The significance of using this gradient regularization is as follows:
According to (19), the w derived from the regularized gradient g r a d n o r satisfies
w g r a d ( j ) S ( j ) = w g r a d n o r ( j ) = λ 2 , j
Consider the first-order Taylor expansion of the loss function L ( j ) ( x t ) = L ( G ( j ) ( x t , c ) , G ( j ) ( x , c ) ) at the tth iteration:
Δ L ( j ) = L ( j ) ( x t + a s i g n ( w ) ) L ( j ) ( x t ) L ( j ) ( x t + a w ) L ( j ) ( x t ) x t L ( j ) ( x t ) ( a w ) = a w g r a d ( j )
where the first approximately equal sign ignores the effect of taking the sign function for w on the result, and the second approximately equal sign ignores the remainder of the first-order Taylor formula for approximation.
Uniting (26) and (27), there is
Δ L ( j ) a w g r a d ( j ) = a λ 2 S ( j ) = a λ 2 1 L ( j )
The last equal sign of the above equation can be held by taking a sufficiently small ς . Since a λ 2 is a constant, each model’s value of the loss function change Δ L ( j ) is inversely proportional to their current corresponding loss function value L ( j ) , implying that the smaller the value of the loss function, the larger the optimization gain can be obtained. In practical adversarial attacks, the adversarial examples achieve successful attacks on some vulnerable models after a small number of iterative steps, as their corresponding loss functions have reached the threshold. It is meaningless to further improve these loss functions. Using this regularization can make the adversarial example mainly optimized in the direction of improving the loss functions that have not reached the threshold, which can improve the attack effect on their corresponding models and pursue a higher comprehensive attack success rate.

3.5. CDMAA Framework

In summary, this paper proposes a framework of adversarial attacks on multiple domains of multiple models simultaneously. Using the I-FGSM adversarial attack algorithm as an example, the procedure of CDMAA is as follows (Algorithm 1):
Algorithm 1 CDMAA
Input: original image x , iterative steps T, perturbation magnitude ε , step size a, deepfake model group G ( j ) , j = 1 , , J
Output: adversarial example x ˜
Initialization: x 0 = x
1.
For t = 0 to T 1 do
2.
For j = 1 to J do
3.
L ( j ) = 1 K ( j ) i = 1 K ( j ) L ( G ( j ) ( x t , c i ) , G ( j ) ( x , c i ) )
4.
g r a d ( j ) = x t L ( j )
5.
g r a d n o r ( j ) = g r a d ( j ) / 1 max ( L ( j ) , ς )
6.
End for
7.
w = a r g m i n a A ¯ j = 1 J a ( j ) g r a d n o r ( j ) 2 2
8.
x t + 1 = c l i p ( x t + a s i g n ( w ) )
9.
End for
10.
x ˜ = x T
Step 3 is the use of the uniformly weighting method to obtain the cross-domain loss function, which is sufficiently effective due to the similarity of the gradients between domains (Section 3.2). It needs only one backpropagation step to calculate gradient in the following step, while the MGDA needs K ( j ) backpropagation steps to calculate the gradients in each domain, thus ensuring good efficiency.
Step 7 is the use of the MGDA to obtain the cross-model perturbation vector, where a simple uniformly weighting method is less effective due to the low similarity of gradients between models (Section 3.3). Therefore, the MGDA is used to achieve better attacks at the expense of time. We use the Frank–Wolfe method [36] to approximately calculate the minimal norm vector in the convex hull of g r a d n o r , which has a well convergence in such cases that the number of dimensions is much larger than the number of vectors [35,37].
Figure 3 shows the overview of CDMAA. It is noted that CDMAA is not necessarily applied on the I-FGSM, although this paper uses the I-FGSM to introduce CDMAA. The main idea of CDMAA is to obtain a perturbation vector from gradients in multiple domains and models and then update the adversarial examples to ensure their ability of attacking multiple models and domains. Therefore, CDMAA can be applied to any gradient-based adversarial attack algorithms, such as the MI-FGSM and APGD.

4. Experiment and Analysis

To verify the effectiveness of the proposed CDMAA framework, we conduct adversarial attack experiments against deepfake models and analyze the results. In Section 4.1, we introduce the deepfake model, hyper-parameters and evaluation criteria used in the experiments. In Section 4.2, we use CDMAA to attack four models at the same time and show the result of adversarial attacks. In Section 4.3, we conduct ablation experiments to show the impact of CDMAA components on the attack and compare with the methods used in the existing work.

4.1. Deepfake Models, Hyper-Parameters and Evaluation Metrics

We prepared four deepfake models—StarGAN, AttGAN, STGAN and U-GAT-IT—for the adversarial attack experiments, which are chosen in similar existing work [12,13,34]. StarGAN and AttGAN adopt the officially provided pre-training models, which are training respectively in five domains—black hair, blonde hair, brown hair, gender and age—as well as in 13 domains, such as bald head and bangs on the celebA dataset. STGAN uses the model trained on the celebA dataset in five domains—bangs, glasses, beard, slightly opened mouth and pale skin—which are rare attributes in original images. We selected these domains to prevent the possibility that the STGAN output will be the same as the input when the original picture already contains the attributes of the corresponding domains; in which case, the experiment results will be affected since the model is unable to effectively forge the images even without adversarial examples [34]. U-GAT-IT realizes the translation of images from a single domain to another, so it can be regarded as a special case of multi-domain deepfake when the total number of conditional variables K U - GAT - IT = 1 . To unify the dataset in the experiments, we used the U-GAT-IT model to translate from celebA to the anime dataset, which is trained on official codes.
The adversarial attack algorithm uses the I-FGSM, in which the hyper-parameters refer to the settings in the existing work, T = 80 , ε = 0.05   and   a = 0.01 , except where noted. The test data are N = 100 randomly sampled pictures in the celebA dataset (ensure that the pictures used in each contrast experiment are the same).
The value of the loss function
L i ( j ) ( x ˜ ( n ) , x ( n ) ) = L ( G ( j ) ( x ˜ ( n ) , c i ) , G ( j ) ( x ( n ) , c i ) )
is used to quantify the output distortion of the nth adversarial example x ( n ) to model G ( j ) under the condition variable c i . The following evaluation criteria [13] are considered to evaluate the effectiveness of the adversarial attack:
a v g _ L ( j ) = 1 N K j n = 1 N i = 1 K j L i ( j ) ( x ˜ ( n ) , x ( n ) )
a t t a c k _ r a t e ( j ) = 1 N K j n = 1 N i = 1 K j 1 L i ( j ) ( x ˜ ( n ) , x ( n ) ) > τ
where a v g _ L ( j ) represents the average value of the loss function of N adversarial examples in each domain for model G ( j ) and a t t a c k _ r a t e ( j ) represents the proportion of the loss function of N adversarial examples in each domain reaching the threshold τ = 0.05 [13] for Model G ( j ) , i.e., the attack success rate.

4.2. CDMAA Adversarial Attack Experiment

We used CDMAA framework to attack the four deepfake models StarGAN, AttGAN, STGAN and U-GAT-IT at the same time. The results are shown in Table 1:
The results show that the generated adversarial examples achieve certain effects on each domain of the four deepfake models. StarGAN and U-GAT-IT are relatively vulnerable to adversarial attacks because the average L values is much greater than the threshold and the success rate of attack is close to 100%, respectively. The success rates of attacks on AttGAN and StarGAN are relatively lower; AttGAN and StarGAN are relatively less affected by the adversarial attack.
In addition, comparing the three groups of experiments (a), (b) and (c), we see the attack effect can be improved by relaxing the limit of algorithm parameters, such as increasing ε or T (at the expense of a more obvious perturbation or a larger computational cost). Comparing the three groups of experiments (a), (d) and (e), we find that using a better adversarial attack algorithm (the MI-FGSM is an improvement on the I-FGSM and APGD is an improvement on the MI-FGSM) can improve the attack effect. Both the MI-FGSM and APGD perform J gradient backpropagation in each iteration, which is the same as the I-FGSM. Therefore, they have the same algorithm time complexity O(T) and roughly similar computational cost. All this shows that the CDMAA framework is well compatible with adversarial attack algorithms, and the general improvement methods are also applicable to CDMAA.
Figure 4 shows the attack effectiveness of some of the adversarial examples in the above experiment:
Figure 4 shows that the difference between the adversarial example and the original image is so small that the human eye can hardly distinguish it. However, the difference between the output of the deepfake model, i.e., the distortion of the fake image, is large enough to be distinguished. Therefore, using the adversarial example instead of the original image can significantly deform the output of the deepfake model so as to effectively prevent the model from forging pictures.

4.3. Ablation/Contrast Experiments

4.3.1. Cross-Domain Gradient Ablation/Contrast Experiment

To verify that the method of uniformly weighted cross-domain gradients used by CDMAA can effectively expand the generalization of adversarial examples between various domains, we carry out the contrast attack experiment, where we keep other components of CDMAA unchanged and only change the way to handle different gradients in domains: (1) Single gradient: g r a d ( j ) = g r a d 1 ( j ) , i.e., use the gradient of only one domain as the cross-domain gradients, without considering the generalization of the generated adversarial examples in other domains. This problem exists in most current studies [30,32,33]. (2) Iterative gradient:  g r a d ( j ) = g r a d t   mod   K j ( j ) , i.e., iteratively use the gradient of each domain loss function as the cross-domain gradients [13]. (3) The MGDA: g r a d ( j ) = M G D A i = 1 , , K j ( g r a d i ( j ) ) , i.e., use the MGDA to generate cross-domain gradients. The results are shown in Table 2:
Figure 5 shows the visual comparison of the result. Compared with existing research on adversarial attacks against deepfake, which only use single domain gradients or iterative gradients in each domain, the CDMAA framework, using the method of uniform weighting to generate cross-domain gradients, can achieve a higher attack success rate against each model, especially those with more domains, such as AttGAN, and effectively increase the generalization of adversarial attack examples between domains. The average L of some models using the method of single gradient or iterative gradient can exceed CDMAA, which shows that the effectiveness of the adversarial examples generated by these two methods on each domain varies greatly, which is not as well-balanced and stable as CDMAA. In addition, compared with using the MGDA to generate cross-domain gradients, the effectiveness of simply using uniform weighting is not quite different, but it can greatly reduce the time consumption (Section 3.5). Therefore, CDMAA uses the most efficient uniform weighting method to calculate the cross-domain gradients.

4.3.2. Cross-Model Perturbation Ablation/Contrast Experiment

To verify that CDMAA uses the MGDA to calculate the cross-model perturbation vector w, which can effectively expand the generalization of adversarial examples between various models, we carry out the contrast attack experiment, where we keep other components of CDMAA unchanged and only change the way to process each cross-domain gradient: (1) Single gradient: w = g r a d ( 1 ) , i.e., only use the cross-domain gradients of one model to update the adversarial example, which is equivalent to ignoring whether the adversarial example has the generalization ability to attack other models. (2) Iterative gradient: w = g r a d ( t   mod   J ) , i.e., iteratively use the cross-domain gradients of each model to update the adversarial example [12]. (3) Uniform weighting: w = 1 J j = 1 J g r a d ( j ) , i.e., use the mean of the cross-domain gradients of each model to update the adversarial example [34]. The results are shown in Table 3:
Figure 6 shows the visual comparison of the results. Although the methods of single gradient, iterative gradient and uniform weighting used in the current research can reach a high average L value on some models, such as StarGAN, their effectiveness on models that are robust against adversarial attacks (such as AttGAN and STGAN) are very poor. In fact, it is meaningless to reach such a high average L value: as long as the threshold τ = 0.05 is exceeded, the output distortion is obvious enough and a successful adversarial attack is achieved. In contrast, the group “MGDA” can achieve a considerable attack success rate for each model. Therefore, we use the MGDA to calculate cross-model perturbation to improve the generalization of adversarial examples across models.

4.3.3. Gradient Regularization Ablation/Contrast Experiment

To verify the effectiveness of gradient regularization used in CDMAA, we carry out the attack contrast attack experiment, where we keep other CDMAA components unchanged and only change the gradient regularization method used: (1) Without regularization: S ( j ) = 1 , i.e., the regularization factor is always 1, which is equivalent to not using a regularization method. (2) L2 regularization: S ( j ) = g r a d ( j ) 2 . (3) Logarithmic gradient regularization [15]: S ( j ) = L ( j ) . The results are shown in Table 4:
Figure 7 shows the visual comparison of the result. On the metric of attack_rate, the method of penalty-based gradient regularization is superior to other gradient regularizations. It achieves a more uniform attack effect distribution on each model by reducing the effect on the model with large loss function value in exchange for a major attack on the model with small loss function value. In the actual process of adversarial attacks, due to the large gap in the vulnerability of each model to attacks, the use of this gradient regularization method will be more practical.

5. Conclusions and Future Work

In this paper, we propose a framework of an adversarial attack against the deepfake model called CDMAA, which can expand the generalization of the generated adversarial examples in each domain of multiple models. Specifically, using CDMAA to generate adversarial examples can distort fake images, i.e., the output of multiple deepfake models under any condition variables, so as to interfere with the deepfake model and protect the pictures from model tampering. An adversarial attack experiment on four mainstream deepfake models shows that the adversarial examples generated by CDMAA have high attack success rates and can effectively attack multiple deepfake models at the same time. Through ablation experiments, on the one hand, we verify the effectiveness of each CDMAA component; on the other hand, compared with other similar research methods, we verify the superiority of CDMAA.
Since CDMAA needs to use the gradient-based adversarial attack algorithm, future work can focus on how to extend this framework to no-gradients-required adversarial attack algorithms, such as AdvGAN [38] or Boundary Attack [39]. In addition, we will try to extend CDMAA to attack other data types of deepfake, such as video and voice.

Author Contributions

Conceptualization and writing—original draft preparation, H.Q. and Y.D.; methodology, H.Q. and T.L.; software, H.Q.; writing—original draft, H.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the 2021 Fundamental Research Funds for the Central Universities of PPSUC (NO.2021JKF105) and the World-class Discipline University Construction Funds of PPSUC(NO.2021FZB13).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available in a publicly accessible repository.

Acknowledgments

We thank Zhiqiang Song and Cheng Yang, who spent much time and effort on reviewing, revising and translating this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Faceswap: Deepfakes Software for All. 2017. Available online: https://github.com/deepfakes/faceswap (accessed on 10 December 2021).
  2. Chesney, R.; Citron, D. Deepfakes and the new disinformation war: The coming age of post-truth geopolitics. Foreign Aff. 2019, 98, 147. [Google Scholar]
  3. Maras, M.H.; Alexandrou, A. Determining authenticity of video evidence in the age of artificial intelligence and in the wake of deepfake videos. Int. J. Evid. Proof 2019, 23, 255–262. [Google Scholar] [CrossRef]
  4. Rossler, A.; Cozzolino, D.; Verdoliva, L.; Riess, C.; Thies, J.; Nießner, M. Faceforensics++: Learning to detect manipulated facial images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 1–11. [Google Scholar]
  5. Bonettini, N.; Cannas, E.D.; Mandelli, S.; Bondi, L.; Bestagini, P.; Tubaro, S. Video face manipulation detection through ensemble of cnns. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 5012–5019. [Google Scholar]
  6. Tariq, S.; Lee, S.; Woo, S. One detector to rule them all: Towards a general deepfake attack detection framework. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 3625–3637. [Google Scholar]
  7. Afchar, D.; Nozick, V.; Yamagishi, J.; Echizen, I. Mesonet: A compact facial video forgery detection network. In Proceedings of the 2018 IEEE International Workshop on Information Forensics and Security (WIFS), Hong Kong, China, 11–13 December 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–7. [Google Scholar]
  8. Zhao, H.; Zhou, W.; Chen, D.; Wei, T.; Zhang, W.; Yu, N. Multi-attentional deepfake detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online. 19–25 June 2021; pp. 2185–2194. [Google Scholar]
  9. Mittal, T.; Bhattacharya, U.; Chandra, R.; Bera, A.; Manocha, D. Emotions Don’t Lie: An Audio-Visual Deepfake Detection Method using Affective Cues. In Proceedings of the 28th ACM international conference on multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2823–2832. [Google Scholar]
  10. Chugh, K.; Gupta, P.; Dhall, A.; Subramanian, R. Not made for each other-audio-visual dissonance-based deepfake detection and localization. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 439–447. [Google Scholar]
  11. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  12. Huang, H.; Wang, Y.; Chen, Z.; Li, Y.; Tang, Z.; Chu, W.; Chen, J.; Lin, W.; Ma, K.K. CMUA-Watermark: A Cross-Model Universal Adversarial Watermark for Combating Deepfakes. arXiv 2021, arXiv:2105.10872. [Google Scholar]
  13. Ruiz, N.; Bargal, S.A.; Sclaroff, S. Disrupting deepfakes: Adversarial attacks against conditional image translation networks and facial manipulation systems. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2020; pp. 236–251. [Google Scholar]
  14. Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial examples in the physical world. arXiv 2016, arXiv:1607.02533. [Google Scholar]
  15. Désidéri, J.A. Multiple-gradient descent algorithm (MGDA) for multiobjective optimization. Comptes Rendus Math. 2012, 350, 313–318. [Google Scholar] [CrossRef]
  16. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  17. Karras, T.; Aila, T.; Laine, S.; Lehtinen, J. Progressive growing of gans for improved quality, stability, and variation. arXiv 2017, arXiv:1710.10196. [Google Scholar]
  18. Karras, T.; Laine, S.; Aila, T. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4401–4410. [Google Scholar]
  19. Choi, Y.; Choi, M.; Kim, M.; Ha, J.W.; Kim, S.; Choo, J. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8789–8797. [Google Scholar]
  20. He, Z.; Zuo, W.; Kan, M.; Shan, S.; Chen, X. Attgan: Facial attribute editing by only changing what you want. IEEE Trans. Image Process. 2019, 28, 5464–5478. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Liu, M.; Ding, Y.; Xia, M.; Liu, X.; Ding, E.; Zuo, W.; Wen, S. STGAN: A unified selective transfer network for arbitrary image attribute editing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3673–3682. [Google Scholar]
  22. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  23. Kim, J.; Kim, M.; Kang, H.; Lee, K. U-gat-it: Unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation. arXiv 2019, arXiv:1907.10830. [Google Scholar]
  24. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  25. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
  26. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. arXiv 2017, arXiv:1706.06083. [Google Scholar]
  27. Dong, Y.; Liao, F.; Pang, T.; Su, H.; Zhu, J.; Hu, X.; Li, J. Boosting adversarial attacks with momentum. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9185–9193. [Google Scholar]
  28. Croce, F.; Hein, M. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In Proceedings of the International Conference on Machine Learning, PMLR, Online. 13–18 July 2020; pp. 2206–2216. [Google Scholar]
  29. Kos, J.; Fischer, I.; Song, D. Adversarial examples for generative models. In Proceedings of the 2018 IEEE Security and Privacy Workshops (spw), San Francisco, CA, USA, 24–24 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 36–42. [Google Scholar]
  30. Yeh, C.Y.; Chen, H.W.; Tsai, S.L.; Wang, S.D. Disrupting image-translation-based deepfake algorithms with adversarial attacks. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision Workshops, Snowmass Village, CO, USA, 1–5 March 2020; pp. 53–62. [Google Scholar]
  31. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  32. Lv, L. Smart Watermark to Defend against Deepfake Image Manipulation. In Proceedings of the 2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS), Chengdu, China, 23–26 April 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 380–384. [Google Scholar]
  33. Dong, J.; Xie, X. Visually Maintained Image Disturbance Against Deepfake Face Swapping. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar]
  34. Fang, Z.; Yang, Y.; Lin, J.; Zhan, R. Adversarial Attacks For Multi Target Image Translation Networks. In Proceedings of the 2020 IEEE International Conference on Progress in Informatics and Computing (PIC), Shanghai, China, 18–20 December 2020; IEEE: Piscataway, NJ, USA, 2022; pp. 179–184. [Google Scholar]
  35. Jaggi, M. Revisiting Frank-Wolfe: Projection-free sparse convex optimization. In Proceedings of the International Conference on Machine Learning, PMLR, Atlanta, GA, USA, 17–19 June 2013; pp. 427–435. [Google Scholar]
  36. Frank, M.; Wolfe, P. An algorithm for quadratic programming. Nav. Res. Logist. Q. 1956, 3, 95–110. [Google Scholar] [CrossRef]
  37. Sener, O.; Koltun, V. Multi-task learning as multi-objective optimization. arXiv 2018, arXiv:1810.04650. [Google Scholar]
  38. Xiao, C.; Li, B.; Zhu, J.Y.; He, W.; Liu, M.; Song, D. Generating adversarial examples with adversarial networks. arXiv 2018, arXiv:1801.02610. [Google Scholar]
  39. Brendel, W.; Rauber, J.; Bethge, M. Decision-based adversarial attacks: Reliable attacks against black-box machine learning models. arXiv 2017, arXiv:1712.04248. [Google Scholar]
Figure 1. Example of adversarial attack on deepfake. The output image with the input of an adversarial example is significantly distorted, while the perturbation added to the original image is visually undetectable.
Figure 1. Example of adversarial attack on deepfake. The output image with the input of an adversarial example is significantly distorted, while the perturbation added to the original image is visually undetectable.
Futureinternet 14 00046 g001
Figure 2. Visualization of gradients between the domains or models (Case the number of gradients is 5). Gradients between domains (a) tend to have similar directions and norms since they are from the same model. Gradients between models (b) are largely different in their directions and norms.
Figure 2. Visualization of gradients between the domains or models (Case the number of gradients is 5). Gradients between domains (a) tend to have similar directions and norms since they are from the same model. Gradients between models (b) are largely different in their directions and norms.
Futureinternet 14 00046 g002
Figure 3. Overview of CDMAA.
Figure 3. Overview of CDMAA.
Futureinternet 14 00046 g003
Figure 4. Illustrations of adversarial attack. The first column shows the original images and its adversarial examples, and the next eight columns illustrate the output images of models in corresponding domains with the first column as the input.
Figure 4. Illustrations of adversarial attack. The first column shows the original images and its adversarial examples, and the next eight columns illustrate the output images of models in corresponding domains with the first column as the input.
Futureinternet 14 00046 g004
Figure 5. Visual comparison of (a) attack_rate and (b) model output disruption for different cross-model gradient calculation methods. The values of L below 0.05 in (b) have been marked red. Combining the attack rate of four models into consideration, “Uniform weighting” is superior to “Single gradient” and “Iterative gradient” and not far off from “MGDA”.
Figure 5. Visual comparison of (a) attack_rate and (b) model output disruption for different cross-model gradient calculation methods. The values of L below 0.05 in (b) have been marked red. Combining the attack rate of four models into consideration, “Uniform weighting” is superior to “Single gradient” and “Iterative gradient” and not far off from “MGDA”.
Futureinternet 14 00046 g005
Figure 6. Visual comparison of (a) attack_rate and (b) model output disruption for different cross-model perturbation calculation methods. The values of L below 0.05 in (b) have been marked red. The last group ”MGDA” can attack each model well, while the group “Single gradient” has a strong attack effect on only one model and the other two groups have limited attack effects on AttGAN and STGAN.
Figure 6. Visual comparison of (a) attack_rate and (b) model output disruption for different cross-model perturbation calculation methods. The values of L below 0.05 in (b) have been marked red. The last group ”MGDA” can attack each model well, while the group “Single gradient” has a strong attack effect on only one model and the other two groups have limited attack effects on AttGAN and STGAN.
Futureinternet 14 00046 g006
Figure 7. Visual comparison of (a) attack_rate and (b) model output disruption for different gradient regularization. The values of L below 0.05 in (b) have been marked red. The group “Penalty-based gradient regularization” can achieve higher attack_rate than the other three groups over four models.
Figure 7. Visual comparison of (a) attack_rate and (b) model output disruption for different gradient regularization. The values of L below 0.05 in (b) have been marked red. The group “Penalty-based gradient regularization” can achieve higher attack_rate than the other three groups over four models.
Futureinternet 14 00046 g007
Table 1. Adversarial attack deepfake models with CDMAA. The results of five groups of experiments are listed as five columns. (a) Basic group. Set the default attack algorithm (I-FGSM) and parameters ( T = 80 , ε = 0.05 , a = 0.01 ). (b) Increase T to 100 times. (c) Expand ε to 0.06. (d) Use the MI-FGSM as the attack algorithm. (e) Use APGD as the attack algorithm.
Table 1. Adversarial attack deepfake models with CDMAA. The results of five groups of experiments are listed as five columns. (a) Basic group. Set the default attack algorithm (I-FGSM) and parameters ( T = 80 , ε = 0.05 , a = 0.01 ). (b) Increase T to 100 times. (c) Expand ε to 0.06. (d) Use the MI-FGSM as the attack algorithm. (e) Use APGD as the attack algorithm.
(a) I-FGSM ( b )   I - FGSM   T = 100 ( c )   I - FGSM   ε = 0.06 (d) MI-FGSM(e) APGD
Avg_LAttack_RateAvg_LAttack_RateAvg_LAttack_RateAvg_LAttack_RateAvg_LAttack_Rate
StarGAN0.21198.8%0.22299.2%0.25999%0.22399.4%0.24499.8%
AttGAN0.10856.9%0.11460.7%0.10870.5%0.11259.5%0.12364.1%
STGAN0.07358.2%0.08062.2%0.14980.4%0.07461.2%0.08571.8%
U-GAT-IT0.24297%0.26198%0.303100%0.26798%0.27699%
Table 2. The effect of different cross-domain gradient calculation methods on adversarial attack. The last group “Uniform weighting” is used in CDMAA and the other three are contrast groups.
Table 2. The effect of different cross-domain gradient calculation methods on adversarial attack. The last group “Uniform weighting” is used in CDMAA and the other three are contrast groups.
Single GradientIterative GradientMGDAUniform Weighting
Avg_LAttack_RateAvg_LAttack_RateAvg_LAttack_RateAvg_LAttack_Rate
StarGAN0.25399.4%0.20798.8%0.493100%0.200100%
AttGAN0.04929.2%0.06748.4%0.06742.6%0.09455.4%
STGAN0.06447.6%0.06048.2%0.05753.6%0.06348.6%
U-GAT-IT0.36499%0.24297.0%0.742100%0.220100%
Table 3. The effect of different cross-model perturbation calculation methods on adversarial attack. The last group “MGDA” is used in CDMAA and the other three are contrast groups.
Table 3. The effect of different cross-model perturbation calculation methods on adversarial attack. The last group “MGDA” is used in CDMAA and the other three are contrast groups.
Single GradientIterative GradientUniform WeightingMGDA
Avg_LAttack_RateAvg_LAttack_RateAvg_LAttack_RateAvg_LAttack_Rate
StarGAN1.143100%0.851100%1.176100%0.19999.4%
AttGAN0.0000.0%0.03318.3%0.0000.0%0.09959.7%
STGAN0.0000.0%0.0147.0%0.0000.0%0.06857.2%
U-GAT-IT0.0271.0%1.002100%1.629100%0.21699%
Table 4. The effect of different gradient regularization on adversarial attack. The last group “Penalty-based gradient regularization” is used in CDMAA and the other three are contrast groups.
Table 4. The effect of different gradient regularization on adversarial attack. The last group “Penalty-based gradient regularization” is used in CDMAA and the other three are contrast groups.
Without RegularizationL2 RegularizationLogarithmic Gradient RegularizationPenalty-Based Gradient Regularization
Avg_LAttack_RateAvg_LAttack_RateAvg_LAttack_RateAvg_LAttack_Rate
StarGAN0.30499.8%0.967100%0.626100%0.221100%
AttGAN0.11959%0.07846%0.12260.6%0.11562.9%
STGAN0.05444.4%0.07660.8%0.0126.3%0.07362.6%
U-GAT-IT0.38898%1.389100%0.96399%0.253100%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Qiu, H.; Du, Y.; Lu, T. The Framework of Cross-Domain and Model Adversarial Attack against Deepfake. Future Internet 2022, 14, 46. https://doi.org/10.3390/fi14020046

AMA Style

Qiu H, Du Y, Lu T. The Framework of Cross-Domain and Model Adversarial Attack against Deepfake. Future Internet. 2022; 14(2):46. https://doi.org/10.3390/fi14020046

Chicago/Turabian Style

Qiu, Haoxuan, Yanhui Du, and Tianliang Lu. 2022. "The Framework of Cross-Domain and Model Adversarial Attack against Deepfake" Future Internet 14, no. 2: 46. https://doi.org/10.3390/fi14020046

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop