Abstract
In this day and age, fake images can be easily generated using some of the state-of-the-art Generative Adversarial Networks (GANs), including Deepfake. These fake images, along with unfactual content, can pose a substantial threat to our society since they are indistinguishable from real ones by the human eye. Therefore, Deepfake detection has gained immense interest in academia and industry. In practice, most detection methods use simple deep neural networks (DNNs) as the backbone. However, they are vulnerable to adversarial examples. This work presents practical pipelines in both white-box and black-box attack scenarios that can fool DNN-based Deepfake detectors into classifying fake images as real. We show that adversarial attacks can be a real threat to Deepfake detectors, even in a black-box setting. We also analyze the transferability of the white-box attacks from one model to another. Then, defensive perspectives are considered based on the practical context.
N. H. Vo and K. D. Phan—Equal contribution.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
http://cvlab.cse.msu.edu/dffd-diverse-fake-face-dataset.html
I never said that! High-tech deception of ‘deepfake’ videos, July 2018. https://www.cbsnews.com/news/i-never-said-that-high-tech-deception-of-deepfake-videos/
FakeApp 2.2.0 - download for PC free, March 2019. https://www.malavida.com/en/soft/fakeapp/#gref
Athalye, A., Engstrom, L., Ilyas, A., Kwok, K.: Synthesizing robust adversarial examples. In: International Conference on Machine Learning, pp. 284–293. PMLR (2018)
Carlini, N., Wagner, D.: Towards evaluating the robustness of neural networks. In: 2017 IEEE Symposium on Security and Privacy (SP), pp. 39–57. IEEE (2017)
Citron, D., Chesney, R.: Deepfakes and the new disinformation war, June 2020. https://www.foreignaffairs.com/articles/world/2018-12-11/deepfakes-and-new-disinformation-war
Vaccari, C., Chadwick, A.: Deepfakes and disinformation: exploring the impact of synthetic political video on deception, uncertainty, and trust in news (2020). https://journals.sagepub.com/doi/full/10.1177/2056305120903408
Dickson, B.: Adversarial attacks against machine learning systems - everything you need to know, June 2020. https://portswigger.net/daily-swig/adversarial-attacks-against-machine-learning-systems-everything-you-need-to-know
Dziugaite, G.K., Ghahramani, Z., Roy, D.M.: A study of the effect of JPG compression on adversarial images. arXiv preprint arXiv:1608.00853 (2016)
Goodfellow, I., Shlens, J., Szegedy, C.: Explaining and harnessing adversarial examples. CoRR arXiv:1412.6572 (2015)
Guo, C., Gardner, J., You, Y., Wilson, A.G., Weinberger, K.: Simple black-box adversarial attacks. In: International Conference on Machine Learning, pp. 2484–2493. PMLR (2019)
Hussain, S., Neekhara, P., Jere, M., Koushanfar, F., McAuley, J.: Adversarial Deepfakes: evaluating vulnerability of Deepfake detectors to adversarial examples. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 3348–3357 (2021)
Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press (2016). http://www.deeplearningbook. [org. 1, 2]
Ilyas, A., Engstrom, L., Athalye, A., Lin, J.: Black-box adversarial attacks with limited queries and information. In: International Conference on Machine Learning, pp. 2137–2146. PMLR (2018)
Juefei-Xu, F., Wang, R., Huang, Y., Guo, Q., Ma, L., Liu, Y.: Countering malicious Deepfakes: survey, battleground, and horizon. ArXiv arXiv:2103.00218 (2021)
Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4401–4410 (2019)
Kurakin, A., Goodfellow, I., Bengio, S., et al.: Adversarial examples in the physical world (2016)
Li, H., Li, B., Tan, S., Huang, J.: Identification of deep network generated images using disparities in color components. Signal Process. 174, 107616 (2020)
Metz, R.: The fight to stay ahead of Deepfake videos before the 2020 us election, June 2019. https://edition.cnn.com/2019/06/12/tech/deepfake-2020-detection/index.html
Moosavi-Dezfooli, S.M., Fawzi, A., Frossard, P.: DeepFool: a simple and accurate method to fool deep neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2574–2582 (2016)
Neekhara, P., Dolhansky, B., Bitton, J., Ferrer, C.C.: Adversarial threats to DeepFake detection: a practical perspective. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 923–932 (2021)
Neekhara, P., Hussain, S.S., Jere, M., Koushanfar, F., McAuley, J.: Adversarial Deepfakes: evaluating vulnerability of DeepFake detectors to adversarial examples. In: 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 3347–3356 (2021)
NVlabs: Nvlabs/FFHQ-dataset: Flickr-Faces-HQ dataset (FFHQ). https://github.com/NVlabs/ffhq-dataset
Rössler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Nießner, M.: FaceForensics++: learning to detect manipulated facial images. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 1–11 (2019)
Wierstra, D., Schaul, T., Glasmachers, T., Sun, Y., Peters, J., Schmidhuber, J.: Natural evolution strategies. J. Mach. Learn. Res. 15(1), 949–980 (2014)
Xuan, X., Peng, B., Dong, J., Wang, W.: On the generalization of GAN image forensics. ArXiv arXiv:1902.11153 (2019)
Acknowledgement
The assistance provided by Assoc. Prof. Minh-Triet Tran was greatly appreciated. This work was supported by MediaFutures: Research Centre for Responsible Media Technology and Innovation, Norway, through the Centres for Research-based Innovation scheme, project number 309339.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Vo, N.H., Phan, K.D., Tran, AD., Dang-Nguyen, DT. (2022). Adversarial Attacks on Deepfake Detectors: A Practical Analysis. In: Þór Jónsson, B., et al. MultiMedia Modeling. MMM 2022. Lecture Notes in Computer Science, vol 13142. Springer, Cham. https://doi.org/10.1007/978-3-030-98355-0_27
Download citation
DOI: https://doi.org/10.1007/978-3-030-98355-0_27
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-98354-3
Online ISBN: 978-3-030-98355-0
eBook Packages: Computer ScienceComputer Science (R0)