Abstract
Did you see that crazy photo of Chris Hemsworth wearing a gorgeous, blue ballgown? What about the leaked photo of Bernie Sanders dancing with Sarah Palin? If these don’t sound familiar, it’s because these events never happened–but with text-to-image generators and deepfake AI technologies, it is effortless for anyone to produce such images. Over the last decade, there has been an explosive rise in research papers, as well as tool development and usage, dedicated to deepfakes, text-to-image generation, and image synthesis. These tools provide users with great creative power, but with that power comes “great responsibility;” it is just as easy to produce nefarious and misleading content as it is to produce comedic or artistic content. Therefore, given the recent advances in the field, it is important to assess the impact they may have. In this paper, we conduct meta-research on deepfakes to visualize the evolution of these tools and paper publications. We also identify key authors, research institutions, and papers based on bibliometric data. Finally, we conduct a survey that tests the ability of participants to distinguish photos of real people from fake, AI-generated images of people. Based on our meta-research, survey, and background study, we conclude that humans are falling behind in the race to keep up with AI, and we must be conscious of the societal impact.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
- 2.
- 3.
- 4.
- 5.
References
Afchar, D., Nozick, V., Yamagishi, J., Echizen, I.: MesoNet: a compact facial video forgery detection network. In: 2018 IEEE International Workshop on Information Forensics and Security (WIFS). IEEE, December 2018. https://doi.org/10.1109/wifs.2018.8630761
Bao, J., Chen, D., Wen, F., Li, H., Hua, G.: Towards open-set identity preserving face synthesis. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6713–6722 (2018). https://doi.org/10.1109/CVPR.2018.00702
Bregler, C., Covell, M., Slaney, M.: Video rewrite: driving visual speech with audio. In: Proceedings of the 24th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 1997, pp. 353–360. ACM Press/Addison-Wesley Publishing Co., USA (1997). https://doi.org/10.1145/258734.258880
Chen, R., Chen, X., Ni, B., Ge, Y.: SimSwap: an efficient framework for high fidelity face swapping. In: Proceedings of the 28th ACM International Conference on Multimedia, MM 2020, New York, NY, USA, pp. 2003–2011. Association for Computing Machinery (2020). https://doi.org/10.1145/3394171.3413630
Dolhansky, B., et al.: The deepfake detection challenge dataset. arXiv:2006.07397 (2020)
Dolhansky, B., Howes, R., Pflaum, B., Baram, N., Canton-Ferrer, C.: The deepfake detection challenge (DFDC) preview dataset. arXiv:1910.08854 (2019)
Fallis, D.: The epistemic threat of deepfakes. Philos. Technol. 34(4), 623–643 (2021). https://doi.org/10.1007/s13347-020-00419-2
Farid, H.: Creating, using, misusing, and detecting deep fakes. J. Online Trust Saf. 1(4), September 2022. https://doi.org/10.54501/jots.v1i4.56. https://tsjournal.org/index.php/jots/article/view/56
Goodfellow, I.J., et al.: Generative adversarial networks (2014)
Guera, D., Delp, E.J.: Deepfake video detection using recurrent neural networks. In: 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pp. 1–6 (2018)
Hancock, J.T., Bailenson, J.N.: The social impact of deepfakes. Cyberpsychol. Behav. Soc. Netw. 24(3), 149–152 (2021). https://doi.org/10.1089/cyber.2021.29208.jth. pMID: 33760669
Huh, M., Liu, A., Owens, A., Efros, A.A.: Fighting fake news: image splice detection via learned self-consistency. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11215, pp. 106–124. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01252-6_7
Kietzmann, J., Lee, L.W., McCarthy, I.P., Kietzmann, T.C.: Deepfakes: trick or treat? Bus. Horizons 63(2), 135–146 (2020). https://doi.org/10.1016/j.bushor.2019.11.006. https://www.sciencedirect.com/science/article/pii/S0007681319301600. Artificial Intelligence and Machine Learning
Korshunov, P., Marcel, S.: Deepfakes: a new threat to face recognition? Assessment and detection (2018)
Li, J., Li, Z., Cao, J., Song, X., He, R.: FaceInpainter: high fidelity face adaptation to heterogeneous domains. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5089–5098, June 2021
Li, L., Bao, J., Yang, H., Chen, D., Wen, F.: FaceShifter: towards high fidelity and occlusion aware face swapping (2019). https://doi.org/10.48550/ARXIV.1912.13457. https://arxiv.org/abs/1912.13457
Li, Y., Lyu, S.: Exposing deepfake videos by detecting face warping artifacts. arXiv:1811.00656 (2018)
Li, Y., Yang, X., Sun, P., Qi, H., Lyu, S.: Celeb-DF: a large-scale challenging dataset for deepfake forensics. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3204–3213 (2019)
Liu, Y., Xiang, M., Shi, H., Mei, T.: One-stage context and identity hallucination network. In: Proceedings of the 29th ACM International Conference on Multimedia, MM 2021, New York, NY, USA, pp. 835–843. Association for Computing Machinery (2021). https://doi.org/10.1145/3474085.3475257
Mirsky, Y., Lee, W.: The creation and detection of deepfakes: a survey. ACM Comput. Surv. (CSUR) 54(1), 1–41 (2021). https://doi.org/10.1145/3425780
Nguyen, T.T., et al.: Deep learning for deepfakes creation and detection: a survey. Comput. Vis. Image Underst. 223, 103525 (2022). https://doi.org/10.1016/j.cviu.2022.103525. https://www.sciencedirect.com/science/article/pii/S1077314222001114
Nirkin, Y., Keller, Y., Hassner, T.: FSGAN: subject agnostic face swapping and reenactment (2019). https://doi.org/10.48550/ARXIV.1908.05932. https://arxiv.org/abs/1908.05932
Nirkin, Y., Masi, I., Tran, A., Hassner, T., Medioni, G.G.: On face segmentation, face swapping, and face perception. In: 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), pp. 98–105 (2017)
Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., Chen, M.: Hierarchical text-conditional image generation with clip latents. arXiv:2204.06125 (2022)
Rennels, J., Cummings, A.: Sex differences in facial scanning: similarities and dissimilarities between infants and adults. Int. J. Behav. Dev. 37, 111–117 (2013). https://doi.org/10.1177/0165025412472411
Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684–10695 (2022)
Tolosana, R., Vera-Rodriguez, R., Fierrez, J., Morales, A., Ortega-Garcia, J.: Deepfakes and beyond: a survey of face manipulation and fake detection. Inf. Fusion 64, 131–148 (2020). https://doi.org/10.1016/j.inffus.2020.06.014. https://www.sciencedirect.com/science/article/pii/S1566253520303110
Westerlund, M.: The emergence of deepfake technology: a review. Technol. Innov. Manag. Rev. 9, 40–53 (2019). https://doi.org/10.22215/timreview/1282. https://timreview.ca/article/1282
Wu, Y., AbdAlmageed, W., Natarajan, P.: Mantra-Net: manipulation tracing network for detection and localization of image forgeries with anomalous features. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019
Xu, Z., et al.: FaceController: controllable attribute editing for face in the wild (2021). https://doi.org/10.48550/ARXIV.2102.11464. https://arxiv.org/abs/2102.11464
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Pocol, A., Istead, L., Siu, S., Mokhtari, S., Kodeiri, S. (2024). Seeing is No Longer Believing: A Survey on the State of Deepfakes, AI-Generated Humans, and Other Nonveridical Media. In: Sheng, B., Bi, L., Kim, J., Magnenat-Thalmann, N., Thalmann, D. (eds) Advances in Computer Graphics. CGI 2023. Lecture Notes in Computer Science, vol 14496. Springer, Cham. https://doi.org/10.1007/978-3-031-50072-5_34
Download citation
DOI: https://doi.org/10.1007/978-3-031-50072-5_34
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-50071-8
Online ISBN: 978-3-031-50072-5
eBook Packages: Computer ScienceComputer Science (R0)