skip to main content
research-article

The Creation and Detection of Deepfakes: A Survey

Published:02 January 2021Publication History
Skip Abstract Section

Abstract

Generative deep learning algorithms have progressed to a point where it is difficult to tell the difference between what is real and what is fake. In 2018, it was discovered how easy it is to use this technology for unethical and malicious applications, such as the spread of misinformation, impersonation of political leaders, and the defamation of innocent individuals. Since then, these “deepfakes” have advanced significantly.

In this article, we explore the creation and detection of deepfakes and provide an in-depth view as to how these architectures work. The purpose of this survey is to provide the reader with a deeper understanding of (1) how deepfakes are created and detected, (2) the current trends and advancements in this domain, (3) the shortcomings of the current defense solutions, and (4) the areas that require further research and attention.

References

  1. 2017. deepfakes/faceswap: Deepfakes Software For All. Retrieved January 27, 2020 from https://github.com/deepfakes/faceswap.Google ScholarGoogle Scholar
  2. Kfir Aberman, Mingyi Shi, Jing Liao, D. Liscbinski, Baoquan Chen, and Daniel Cohen-Or. 2019. Deep video-based performance cloning. In Computer Graphics Forum. Wiley Online Library.Google ScholarGoogle Scholar
  3. Darius Afchar, Vincent Nozick, Junichi Yamagishi, and Isao Echizen. 2018. Mesonet: A compact facial video forgery detection network. In Proceedings of the 2018 IEEE International Workshop on Information Forensics and Security (WIFS). IEEE, 1--7.Google ScholarGoogle ScholarCross RefCross Ref
  4. Akshay Agarwal, Richa Singh, Mayank Vatsa, and Afzel Noore. 2017. SWAPPED! Digital face presentation attack detection via weighted local magnitude pattern. In Proceedings of the 2017 IEEE International Joint Conference on Biometrics (IJCB). IEEE, 659--665.Google ScholarGoogle ScholarCross RefCross Ref
  5. Shruti Agarwal, Hany Farid, Ohad Fried, and Maneesh Agrawala. 2020. Detecting deep-fake videos from phoneme-viseme mismatches. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. 660--661.Google ScholarGoogle ScholarCross RefCross Ref
  6. Shruti Agarwal, Hany Farid, Yuming Gu, Mingming He, Koki Nagano, and Hao Li. 2019. Protecting world leaders against deep fakes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops. 38--45.Google ScholarGoogle Scholar
  7. Sakshi Agarwal and Lav R. Varshney. 2019. Limits of deepfake detection: A robust estimation viewpoint. arXiv:1905.03493 (2019).Google ScholarGoogle Scholar
  8. Zahid Akhtar and Dipankar Dasgupta. [n.d.]. A comparative evaluation of local feature descriptors for DeepFakes detection. ([n.d.]).Google ScholarGoogle Scholar
  9. Rıza Alp Güler, Natalia Neverova, and Iasonas Kokkinos. 2018. Densepose: Dense human pose estimation in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle ScholarCross RefCross Ref
  10. Irene Amerini and Roberto Caldelli. 2020. Exploiting prediction error inconsistencies through LSTM-based classifiers to detect deepfake videos. In Proceedings of the 2020 ACM Workshop on Information Hiding and Multimedia Security. 97--102.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Irene Amerini, Leonardo Galteri, Roberto Caldelli, and Alberto Del Bimbo. 2019. Deepfake video detection through optical flow based CNN. In Proceedings of the IEEE International Conference on Computer Vision Workshops. 0--0.Google ScholarGoogle ScholarCross RefCross Ref
  12. Arije Antinori. 2019. Terrorism and DeepFake: From hybrid warfare to post-truth warfare in a hybrid world. In ECIAIR 2019 European Conference on the Impact of Artificial Intelligence and Robotics. Academic Conferences and Publishing Limited, 23.Google ScholarGoogle Scholar
  13. Hadar Averbuch-Elor, Daniel Cohen-Or, Johannes Kopf, and Michael F. Cohen. 2017. Bringing portraits to life. ACM Transactions on Graphics (Proceedings of SIGGRAPH Asia 2017) 36, 6 (2017), 196.Google ScholarGoogle Scholar
  14. Guha Balakrishnan, Amy Zhao, Adrian V. Dalca, Fredo Durand, and John Guttag. 2018. Synthesizing images of humans in unseen poses. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 8340--8348.Google ScholarGoogle ScholarCross RefCross Ref
  15. Aayush Bansal, Shugao Ma, Deva Ramanan, and Yaser Sheikh. 2018. Recycle-GAN: Unsupervised video retargeting. In Proceedings of the European Conference on Computer Vision (ECCV).Google ScholarGoogle ScholarCross RefCross Ref
  16. Jianmin Bao, Dong Chen, Fang Wen, Houqiang Li, and Gang Hua. 2017. CVAE-GAN: Fine-grained image generation through asymmetric training. In Proceedings of the IEEE International Conference on Computer Vision. 2745--2754.Google ScholarGoogle ScholarCross RefCross Ref
  17. Jianmin Bao, Dong Chen, Fang Wen, Houqiang Li, and Gang Hua. 2018. Towards open-set identity preserving face synthesis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle ScholarCross RefCross Ref
  18. Dmitri Bitouk, Neeraj Kumar, Samreen Dhillon, Peter Belhumeur, and Shree K. Nayar. 2008. Face swapping: Automatically replacing faces in photographs. ACM Transactions on Graphics (TOG) 27, 3 (2008), 1--8.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Volker Blanz, Curzio Basso, Tomaso Poggio, and Thomas Vetter. 2003. Reanimating faces in images and video. In Computer Graphics Forum, Vol. 22. Wiley Online Library, 641--650.Google ScholarGoogle Scholar
  20. Volker Blanz, Kristina Scherbaum, Thomas Vetter, and Hans-Peter Seidel. 2004. Exchanging faces in images. In Computer Graphics Forum, Vol. 23. Wiley Online Library, 669--676.Google ScholarGoogle Scholar
  21. Volker Blanz and Thomas Vetter. 1999. A morphable model for the synthesis of 3D faces. In Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques. 187--194.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Philip Bontrager, Aditi Roy, Julian Togelius, Nasir Memon, and Arun Ross. 2018. DeepMasterPrints: Generating masterprints for dictionary attacks via latent variable evolution. In Proceedings of the 9th International Conference on Biometrics Theory, Applications and Systems (BTAS). IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  23. Jie Cao, Yibo Hu, Bing Yu, Ran He, and Zhenan Sun. 2019. 3D aided duet GANs for multi-view face image synthesis. IEEE Transactions on Information Forensics and Security 14, 8 (2019), 2028--2042.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Nicholas Carlini and Hany Farid. 2020. Evading deepfake-image detectors with white-and black-box attacks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. 658--659.Google ScholarGoogle ScholarCross RefCross Ref
  25. Caroline Chan, Shiry Ginosar, Tinghui Zhou, and Alexei A. Efros. 2019. Everybody dance now. In Proceedings of the IEEE International Conference on Computer Vision. 5933--5942.Google ScholarGoogle Scholar
  26. Yao-Jen Chang and Tony Ezzat. 2005. Transferable videorealistic speech animation. In Proceedings of the 2005 ACM SIGGRAPH/Eurographics Symposium on Computer Animation. ACM, 143--151.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Lele Chen, Ross K. Maddox, Zhiyao Duan, and Chenliang Xu. 2019. Hierarchical cross-modal talking face generation with dynamic pixel-wise loss. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 7832--7841.Google ScholarGoogle ScholarCross RefCross Ref
  28. Robert Chesney and Danielle Keats Citron. 2018. Deep fakes: A looming challenge for privacy, democracy, and national security. (2018).Google ScholarGoogle Scholar
  29. Yunjey Choi, Minje Choi, Munyoung Kim, Jung-Woo Ha, Sunghun Kim, and Jaegul Choo. 2018. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle ScholarCross RefCross Ref
  30. Umur Aybars Ciftci and Ilke Demir. 2019. FakeCatcher: Detection of synthetic portrait videos using biological signals. arXiv preprint arXiv:1901.02212 (2019).Google ScholarGoogle Scholar
  31. Umur Aybars Ciftci, Ilke Demir, and Lijun Yin. 2020. How do the hearts of deep fakes beat? Deep fake source detection via interpreting residuals with biological signals. arXiv preprint arXiv:2008.11363 (2020).Google ScholarGoogle Scholar
  32. Valentina Conotter, Ecaterina Bodnari, Giulia Boato, and Hany Farid. 2014. Physiologically-based detection of computer generated faces in video. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP). IEEE, 248--252.Google ScholarGoogle ScholarCross RefCross Ref
  33. Davide Cozzolino, Justus Thies, Andreas Rossler, Christian Riess, Matthias Niessner, and Luisa Verdoliva. 2018. Forensictransfer: Weakly-supervised domain adaptation for forgery detection. arXiv preprint arXiv:1812.02510 (2018).Google ScholarGoogle Scholar
  34. Kevin Dale, Kalyan Sunkavalli, Micah K. Johnson, Daniel Vlasic, Wojciech Matusik, and Hanspeter Pfister. 2011. Video face replacement. In Proceedings of the 2011 SIGGRAPH Asia Conference. 1--10.Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. Rodrigo De Bem, Arnab Ghosh, Adnane Boukhayma, Thalaiyasingam Ajanthan, N Siddharth, and Philip Torr. 2019. A conditional deep generative model of people in natural images. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV). IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  36. Oscar de Lima, Sean Franklin, Shreshtha Basu, Blake Karwoski, and Annet George. 2020. Deepfake detection using spatiotemporal convolutional networks. arXiv preprint arXiv:2006.14749 (2020).Google ScholarGoogle Scholar
  37. Jesse Demiani. 2019. A Voice Deepfake Was Used to Scam a CEO Out of $243,000. Forbes. https://bit.ly/38sXb1I.Google ScholarGoogle Scholar
  38. Xinyi Ding, Zohreh Raziei, Eric C. Larson, Eli V. Olinick, Paul Krueger, and Michael Hahsler. 2019. Swapped face detection using deep learning and subjective assessment. arXiv preprint arXiv:1909.04217 (2019).Google ScholarGoogle Scholar
  39. Nhu-Tai Do, In-Seop Na, and Soo-Hyung Kim. 2018. Forensics Face Detection From GANs Using Convolutional Neural Network.Google ScholarGoogle Scholar
  40. Brian Dolhansky, Russ Howes, Ben Pflaum, Nicole Baram, and Cristian Canton Ferrer. 2019. The deepfake detection challenge (DFDC) preview dataset. arXiv preprint arXiv:1910.08854 (2019).Google ScholarGoogle Scholar
  41. Mengnan Du, Shiva Pentyala, Yuening Li, and Xia Hu. 2019. Towards generalizable forgery detection with locality-aware AutoEncoder. arXiv preprint arXiv:1909.05999 (2019).Google ScholarGoogle Scholar
  42. Ricard Durall, Margret Keuper, Franz-Josef Pfreundt, and Janis Keuper. 2019. Unmasking DeepFakes with simple features. arXiv preprint arXiv:1911.00686 (2019).Google ScholarGoogle Scholar
  43. P. Ekman, W. Friesen, and J. Hager. 2002. Facial action coding system: Research Nexus. Network Research Information, Salt Lake City, UT 1 (2002).Google ScholarGoogle Scholar
  44. Chi-Ying Chen et al. 2019. A trusting news ecosystem against fake news from humanity and technology perspectives. In Forbes 2019 19th International Conference on Computational Science and Its Applications (ICCSA). IEEE, 132--137.Google ScholarGoogle Scholar
  45. Daniil Kononenko et al. 2017. Photorealistic monocular gaze redirection using machine learning. IEEE Transactions on Pattern Analysis and Machine Intelligence 40, 11 (2017), 2696--2710.Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. Liqian Ma et al. 2018. Disentangled person image generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 99--108.Google ScholarGoogle Scholar
  47. Pavel Korshunov et al. 2019. Tampered speaker inconsistency detection with phonetically aware audio-visual features. In International Conference on Machine Learning.Google ScholarGoogle Scholar
  48. Shengju Qian et al. 2019. Make a face: Towards arbitrary high fidelity face manipulation. In Proceedings of the IEEE International Conference on Computer Vision.Google ScholarGoogle Scholar
  49. Facebook. 2018. Facing Facts. Retrieved March 2, 2020 from https://about.fb.com/news/2018/05/inside-feed-facing-facts/#watchnow.Google ScholarGoogle Scholar
  50. Tiziano Fagni, Fabrizio Falchi, Margherita Gambini, Antonio Martella, and Maurizio Tesconi. 2020. TweepFake: About detecting deepfake tweets. arXiv preprint arXiv:2008.00036 (2020).Google ScholarGoogle Scholar
  51. Steven Fernandes, Sunny Raj, Rickard Ewetz, Jodh Singh Pannu, Sumit Kumar Jha, Eddy Ortiz, Iustina Vintila, and Margaret Salter. 2020. Detecting deepfake videos using attribution-based confidence metric. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. 308--309.Google ScholarGoogle ScholarCross RefCross Ref
  52. Tharindu Fernando, Clinton Fookes, Simon Denman, and Sridha Sridharan. 2019. Exploiting human social cognition for the detection of fake and fraudulent faces via memory networks. arXiv preprint arXiv:1911.07844 (2019).Google ScholarGoogle Scholar
  53. Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the 34th International Conference on Machine Learning-Volume 70. JMLR. org, 1126--1135.Google ScholarGoogle ScholarDigital LibraryDigital Library
  54. Paula Fraga-Lamas and Tiago M. Fernandez-Carames. 2019. Leveraging distributed ledger technologies and blockchain to combat fake news. arXiv preprint arXiv:1904.05386 (2019).Google ScholarGoogle Scholar
  55. Ohad Fried, Ayush Tewari, Michael Zollhofer, Adam Finkelstein, Eli Shechtman, Dan B. Goldman, Kyle Genova, Zeyu Jin, Christian Theobalt, and Maneesh Agrawala. 2019. Text-based editing of talking-head video. arXiv preprint arXiv:1906.01524 (2019).Google ScholarGoogle Scholar
  56. Chaoyou Fu, Yibo Hu, Xiang Wu, Guoli Wang, Qian Zhang, and Ran He. 2019. High fidelity face manipulation with extreme pose and expression. arXiv preprint arXiv:1903.12003 (2019).Google ScholarGoogle Scholar
  57. Yaroslav Ganin, Daniil Kononenko, Diana Sungatullina, and Victor Lempitsky. 2016. Deepwarp: Photorealistic image resynthesis for gaze manipulation. In European Conference on Computer Vision. Springer.Google ScholarGoogle ScholarCross RefCross Ref
  58. Pablo Garrido, Levi Valgaerts, Ole Rehmsen, Thorsten Thormahlen, Patrick Perez, and Christian Theobalt. 2014. Automatic face reenactment. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 4217--4224.Google ScholarGoogle ScholarDigital LibraryDigital Library
  59. Leon A. Gatys, Alexander S. Ecker, and Matthias Bethge. 2015. A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576 (2015).Google ScholarGoogle Scholar
  60. Jiahao Geng, Tianjia Shao, Youyi Zheng, Yanlin Weng, and Kun Zhou. 2019. Warp-guided GANs for single-photo facial animation. ACM Transactions on Graphics (TOG) 37, 6 (2019), 231.Google ScholarGoogle ScholarDigital LibraryDigital Library
  61. Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. In Advances in Neural Information Processing Systems. 2672--2680.Google ScholarGoogle Scholar
  62. Kuangxiao Gu, Yuqian Zhou, and Thomas S. Huang. 2020. FLNet: Landmark driven fetching and learning network for faithful talking facial animation synthesis. In AAAI. 10861--10868.Google ScholarGoogle Scholar
  63. David Guera and Edward J. Delp. 2018. Deepfake video detection using recurrent neural networks. In IEEE Conference on Advanced Video and Signal Based Surveillance (AVSS). IEEE, 1--6.Google ScholarGoogle Scholar
  64. Zhiqing Guo, Gaobo Yang, Jiyou Chen, and Xingming Sun. 2020. Fake face detection via adaptive residuals extraction network. arXiv preprint arXiv:2005.04945 (2020).Google ScholarGoogle Scholar
  65. Sungjoo Ha, Martin Kersner, Beomsu Kim, Seokjun Seo, and Dongyoung Kim. 2020. MarioNETte: Few-shot face reenactment preserving identity of unseen targets. In Proceedings of the AAAI Conference on Artificial Intelligence.Google ScholarGoogle ScholarCross RefCross Ref
  66. Holly Kathleen Hall. 2018. Deepfake videos: When seeing isn’t believing. Cath. UJL 8 Tech 27 (2018), 51.Google ScholarGoogle Scholar
  67. Karen Hao. 2019. The biggest threat of deepfakes isn’ t the deepfakes themselves. MIT Tech Review. Retrieved from https://www.technologyreview.com/s/614526/the-biggest-threat-of-deepfakes-isnt-the-deepfakes-themselves/.Google ScholarGoogle Scholar
  68. Haya R. Hasan and Khaled Salah. 2019. Combating deepfake videos using blockchain and smart contracts. IEEE Access 7 (2019), 41596--41606.Google ScholarGoogle ScholarCross RefCross Ref
  69. Chih-Chung Hsu, Yi-Xiu Zhuang, and Chia-Yen Lee. 2020. Deep fake image detection based on pairwise learning. Applied Sciences 10, 1 (2020), 370.Google ScholarGoogle ScholarCross RefCross Ref
  70. Yibo Hu, Xiang Wu, Bing Yu, Ran He, and Zhenan Sun. 2018. Pose-guided photorealistic face rotation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 8398--8406.Google ScholarGoogle ScholarCross RefCross Ref
  71. iperov. 2019. DeepFaceLab: DeepFaceLab is a tool that utilizes machine learning to replace faces in videos. Retrieved December 31, 2019 from https://github.com/iperov/DeepFaceLab.Google ScholarGoogle Scholar
  72. Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A. Efros. 2017. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1125--1134.Google ScholarGoogle Scholar
  73. Seyed Ali Jalalifar, Hosein Hasani, and Hamid Aghajan. 2018. Speech-driven facial reenactment using conditional generative adversarial networks. arXiv preprint arXiv:1803.07461 (2018).Google ScholarGoogle Scholar
  74. Amir Jamaludin, Joon Son Chung, and Andrew Zisserman. 2019. You said that?: Synthesising talking faces from audio. International Journal of Computer Vision (2019), 1--13.Google ScholarGoogle ScholarDigital LibraryDigital Library
  75. Ye Jia, Yu Zhang, Ron Weiss, Quan Wang, Jonathan Shen, Fei Ren, Patrick Nguyen, Ruoming Pang, Ignacio Lopez Moreno, Yonghui Wu, et al. 2018. Transfer learning from speaker verification to multispeaker text-to-speech synthesis. In Advances in Neural Information Processing Systems. 4480--4490.Google ScholarGoogle Scholar
  76. Justin Johnson, Alexandre Alahi, and Li Fei-Fei. 2016. Perceptual losses for real-time style transfer and super-resolution. In European Conference on Computer Vision. Springer, 694--711.Google ScholarGoogle ScholarCross RefCross Ref
  77. Angjoo Kanazawa, Michael J. Black, David W. Jacobs, and Jitendra Malik. 2018. End-to-end recovery of human shape and pose. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 7122--7131.Google ScholarGoogle ScholarCross RefCross Ref
  78. Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. 2019. Analyzing and improving the image quality of styleGAN. arXiv preprint arXiv:1912.04958 (2019).Google ScholarGoogle Scholar
  79. Triantafyllos Kefalas, Konstantinos Vougioukas, Yannis Panagakis, Stavros Petridis, Jean Kossaifi, and Maja Pantic. 2019. Speech-driven facial animation using polynomial fusion of features. arXiv preprint arXiv:1912.05833 (2019).Google ScholarGoogle Scholar
  80. Ira Kemelmacher-Shlizerman. 2016. Transfiguring portraits. ACM Transactions on Graphics (TOG) 35, 4 (2016), 94.Google ScholarGoogle ScholarDigital LibraryDigital Library
  81. Hasam Khalid and Simon S. Woo. 2020. OC-FakeDect: Classifying deepfakes using one-class variational autoencoder. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. 656--657.Google ScholarGoogle Scholar
  82. Ali Khodabakhsh, Raghavendra Ramachandra, Kiran Raja, Pankaj Wasnik, and Christoph Busch. 2018. Fake face detection methods: Can they be generalized? In Proceedings of the 2018 International Conference of the Biometrics Special Interest Group (BIOSIG). IEEE, 1--6.Google ScholarGoogle ScholarCross RefCross Ref
  83. Hyeongwoo Kim, Pablo Carrido, Ayush Tewari, Weipeng Xu, Justus Thies, Matthias Niessner, Patrick Perez, Christian Richardt, Michael Zollhofer, and Christian Theobalt. 2018. Deep video portraits. ACM Transactions on Graphics (TOG) 37, 4 (2018), 163.Google ScholarGoogle ScholarDigital LibraryDigital Library
  84. Jiwon Kim, Jung Kwon Lee, and Kyoung Mu Lee. 2016. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1646--1654.Google ScholarGoogle ScholarCross RefCross Ref
  85. Marissa Koopman, Andrea Macarulla Rodriguez, and Zeno Geradts. 2018. Detection of deepfake video manipulation. In Conference: IMVIP.Google ScholarGoogle Scholar
  86. Pavel Korshunov and Sebastien Marcel. 2018. Deepfakes: A new threat to face recognition? Assessment and detection. arXiv preprint arXiv:1812.08685 (2018).Google ScholarGoogle Scholar
  87. Pavel Korshunov and Sebastien Marcel. 2018. Speaker inconsistency detection in tampered video. In Proceedings of the 2018 26th European Signal Processing Conference (EUSIPCO). IEEE, 2375--2379.Google ScholarGoogle ScholarCross RefCross Ref
  88. Iryna Korshunova, Wenzhe Shi, Joni Dambre, and Lucas Theis. 2017. Fast face-swap using convolutional neural networks. In Proceedings of the IEEE International Conference on Computer Vision.Google ScholarGoogle ScholarCross RefCross Ref
  89. Rithesh Kumar, Jose Sotelo, Kundan Kumar, Alexandre de Brebisson, and Yoshua Bengio. 2017. Obamanet: Photo-realistic lip-sync from text. arXiv preprint arXiv:1801.01442 (2017).Google ScholarGoogle Scholar
  90. Dami Lee. 2019. Deepfake Salvador Dalí takes selfies with museum visitors. The Verge. https://bit.ly/3cEim4m.Google ScholarGoogle Scholar
  91. Jessica Lee, Deva Ramanan, and Rohit Girdhar. 2019. MetaPix: Few-shot video retargeting. arXiv preprint arXiv:1910.04742 (2019).Google ScholarGoogle Scholar
  92. Jia Li, Tong Shen, Wei Zhang, Hui Ren, Dan Zeng, and Tao Mei. 2019. Zooming into face forensics: A pixel-level analysis. arXiv preprint arXiv:1912.05790 (2019).Google ScholarGoogle Scholar
  93. Lingzhi Li, Jianmin Bao, Hao Yang, Dong Chen, and Fang Wen. 2019. FaceShifter: Towards high fidelity and occlusion aware face swapping. arXiv preprint arXiv:1912.13457 (2019).Google ScholarGoogle Scholar
  94. Lingzhi Li, Jianmin Bao, Ting Zhang, Hao Yang, Dong Chen, Fang Wen, and Baining Guo. 2020. Face x-ray for more general face forgery detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 5001--5010.Google ScholarGoogle ScholarCross RefCross Ref
  95. Xurong Li, Kun Yu, Shouling Ji, Yan Wang, Chunming Wu, and Hui Xue. 2020. Fighting against deepfake: Patch8pair convolutional neural networks (PPCNN). In Companion Proceedings of the Web Conference 2020. 88--89.Google ScholarGoogle ScholarDigital LibraryDigital Library
  96. Yuezun Li, Ming-Ching Chang, and Siwei Lyu. 2018. In ictu oculi: Exposing AI created fake videos by detecting eye blinking. In 2018 IEEE International Workshop on Information Forensics and Security (WIFS). IEEE, 1--7.Google ScholarGoogle ScholarCross RefCross Ref
  97. Yuezun Li and Siwei Lyu. 2019. DSP-FWA: Dual Spatial Pyramid for Exposing Face Warp Artifacts in DeepFake Videos. Retrieved December 18, 2019 from https://github.com/danmohaha/DSP-FWA.Google ScholarGoogle Scholar
  98. Yuezun Li and Siwei Lyu. 2019. Exposing DeepFake videos by detecting face warping artifacts. In IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).Google ScholarGoogle Scholar
  99. Yuheng Li, Krishna Kumar Singh, Utkarsh Ojha, and Yong Jae Lee. 2019. MixNMatch: Multifactor disentanglement and encodingfor conditional image generation. arXiv preprint arXiv:1911.11758 (2019).Google ScholarGoogle Scholar
  100. Yuezun Li, Xin Yang, Pu Sun, Honggang Qi, and Siwei Lyu. 2019. Celeb-DF: A new dataset for DeepFake forensics. arXiv preprint:1909.12962 (2019).Google ScholarGoogle Scholar
  101. Yuezun Li, Xin Yang, Baoyuan Wu, and Siwei Lyu. 2019. Hiding faces in plain sight: Disrupting AI face synthesis with adversarial perturbations. arXiv preprint arXiv:1906.09288 (2019).Google ScholarGoogle Scholar
  102. Lingjie Liu, Weipeng Xu, Michael Zollhoefer, Hyeongwoo Kim, Florian Bernard, Marc Habermann, Wenping Wang, and Christian Theobalt. 2019. Neural rendering and reenactment of human actor videos. ACM Transactions on Graphics (TOG) 38, 5 (2019), 139.Google ScholarGoogle ScholarDigital LibraryDigital Library
  103. Wen Liu, Zhixin Piao, Jie Min, Wenhan Luo, Lin Ma, and Shenghua Gao. 2019. Liquid warping GAN: A unified framework for human motion imitation, appearance transfer and novel view synthesis. In Proceedings of the IEEE International Conference on Computer Vision.Google ScholarGoogle ScholarCross RefCross Ref
  104. Zhaoxiang Liu, Huan Hu, Zipeng Wang, Kai Wang, Jinqiang Bai, and Shiguo Lian. 2019. Video synthesis of human upper body with realistic face. arXiv preprint arXiv:1908.06607 (2019).Google ScholarGoogle Scholar
  105. Francesco Marra, Diego Gragnaniello, Davide Cozzolino, and Luisa Verdoliva. 2018. Detection of GAN-generated fake images over social networks. In Proceedings of the 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR). IEEE, 384--389.Google ScholarGoogle ScholarCross RefCross Ref
  106. Francesco Marra, Diego Gragnaniello, Luisa Verdoliva, and Giovanni Poggi. 2019. Do GANs leave artificial fingerprints? In Proceedings of the 2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR). IEEE, 506--511.Google ScholarGoogle ScholarCross RefCross Ref
  107. Iacopo Masi, Aditya Killekar, Royston Marian Mascarenhas, Shenoy Pratik Gurudatt, and Wael AbdAlmageed. 2020. Two-branch recurrent network for isolating deepfakes in videos. arXiv preprint arXiv:2008.03412 (2020).Google ScholarGoogle Scholar
  108. Yisroel Mirsky, Tom Mahler, Ilan Shelef, and Yuval Elovici. 2019. CT-GAN: Malicious tampering of 3D medical imagery using deep learning. In USENIX Security Symposium 2019.Google ScholarGoogle Scholar
  109. Trisha Mittal, Uttaran Bhattacharya, Rohan Chandra, Aniket Bera, and Dinesh Manocha. 2020. Emotions don’t lie: A deepfake detection method using audio-visual affective cues. arXiv preprint arXiv:2003.06711 (2020).Google ScholarGoogle Scholar
  110. Huaxiao Mo, Bolin Chen, and Weiqi Luo. 2018. Fake faces identification via convolutional neural network. In Proceedings of the 6th ACM Workshop on Information Hiding and Multimedia Security. ACM.Google ScholarGoogle ScholarDigital LibraryDigital Library
  111. Joel Ruben Antony Moniz, Christopher Beckham, Simon Rajotte, Sina Honari, and Chris Pal. 2018. Unsupervised depth estimation, 3D face rotation, and replacement. In Advances in Neural Information Processing Systems. 9736--9746.Google ScholarGoogle Scholar
  112. Koki Nagano, Jaewoo Seo, Jun Xing, Lingyu Wei, Zimo Li, Shunsuke Saito, Aviral Agarwal, Jens Fursund, Hao Li, Richard Roberts, et al. 2018. paGAN: Real-time avatars using dynamic textures. ACM Trans. Graph. 37, 6 (2018), 1--12.Google ScholarGoogle ScholarDigital LibraryDigital Library
  113. Ryota Natsume, Tatsuya Yatagawa, and Shigeo Morishima. 2018. FSNet: An identity-aware generative model for image-based face swapping. In Asian Conference on Computer Vision. Springer, 117--132.Google ScholarGoogle Scholar
  114. Ryota Natsume, Tatsuya Yatagawa, and Shigeo Morishima. 2018. RSGAN: Face swapping and editing using face and hair representation in latent spaces. arXiv preprint arXiv:1804.03447 (2018).Google ScholarGoogle Scholar
  115. Paarth Neekhara, Shehzeen Hussain, Malhar Jere, Farinaz Koushanfar, and Julian McAuley. 2020. Adversarial deepfakes: Evaluating vulnerability of deepfake detectors to adversarial examples. arXiv preprint arXiv:2002.12749 (2020).Google ScholarGoogle Scholar
  116. Natalia Neverova, Riza Alp Guler, and Iasonas Kokkinos. 2018. Dense pose transfer. In Proceedings of the European Conference on Computer Vision (ECCV). 123--138.Google ScholarGoogle ScholarDigital LibraryDigital Library
  117. Huy H. Nguyen, Fuming Fang, Junichi Yamagishi, and Isao Echizen. 2019. Multi-task learning for detecting and segmenting manipulated facial images and videos. arXiv preprint arXiv:1906.06876 (2019).Google ScholarGoogle Scholar
  118. Huy H. Nguyen, Junichi Yamagishi, and Isao Echizen. 2019. Capsule-forensics: Using capsule networks to detect forged images and videos. In Proceedings of the ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP). IEEE, 2307--2311.Google ScholarGoogle ScholarCross RefCross Ref
  119. Andrew Gully and Nick Dufour. 2019. DFD. Retrieved from https://ai.googleblog.com/2019/09/contributing-data-to-deepfake-detection.html.Google ScholarGoogle Scholar
  120. Yuval Nirkin, Yosi Keller, and Tal Hassner. 2019. FSGAN: Subject agnostic face swapping and reenactment. In Proceedings of the IEEE International Conference on Computer Vision. 7184--7193.Google ScholarGoogle ScholarCross RefCross Ref
  121. Yuval Nirkin, Iacopo Masi, Anh Tran Tuan, Tal Hassner, and Gerard Medioni. 2018. On face segmentation, face swapping, and face perception. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face 8 Gesture Recognition (FG 2018). IEEE, 98--105.Google ScholarGoogle ScholarCross RefCross Ref
  122. Yuval Nirkin, Lior Wolf, Yosi Keller, and Tal Hassner. 2020. DeepFake detection based on the discrepancy between the face and its context. arXiv preprint arXiv:2008.12262 (2020).Google ScholarGoogle Scholar
  123. Kyle Olszewski, Zimo Li, Chao Yang, Yi Zhou, Ronald Yu, Zeng Huang, Sitao Xiang, Shunsuke Saito, Pushmeet Kohli, and Hao Li. 2017. Realistic dynamic facial textures from a single image using GANs. In Proceedings of the IEEE International Conference on Computer Vision. 5429--5438.Google ScholarGoogle ScholarCross RefCross Ref
  124. Naima Otberdout, Mohamed Daoudi, Anis Kacem, Lahoucine Ballihi, and Stefano Berretti. 2019. Dynamic facial expression generation on Hilbert hypersphere with conditional Wasserstein generative adversarial nets. arXiv preprint arXiv:1907.10087 (2019).Google ScholarGoogle Scholar
  125. Nicolas Papernot, Patrick McDaniel, and Ian Goodfellow. 2016. Transferability in machine learning: From phenomena to black-box attacks using adversarial samples. arXiv preprint arXiv:1605.07277 (2016).Google ScholarGoogle Scholar
  126. Hai X. Pham, Yuting Wang, and Vladimir Pavlovic. 2018. Generative adversarial talking head: Bringing portraits to life with a weakly supervised neural network. arXiv preprint arXiv:1803.07716 (2018).Google ScholarGoogle Scholar
  127. Albert Pumarola, Antonio Agudo, Aleix M. Martinez, Alberto Sanfeliu, and Francesc Moreno-Noguer. 2019. GANimation: One-shot anatomically consistent facial animation. International Journal of Computer Vision (2019), 1--16.Google ScholarGoogle Scholar
  128. Md. Shohel Rana and Andrew H. Sung. 2020. DeepfakeStack: A deep ensemble-based learning technique for deepfake detection. In Proceedings of the 2020 7th IEEE International Conference on Cyber Security and Cloud Computing (CSCloud)/2020 6th IEEE International Conference on Edge Computing and Scalable Cloud (EdgeCom). IEEE, 70--75.Google ScholarGoogle Scholar
  129. Andreas Rossler, Davide Cozzolino, Luisa Verdoliva, Christian Riess, Justus Thies, and Matthias Niessner. 2018. Faceforensics: A large-scale video dataset for forgery detection in human faces. arXiv preprint arXiv:1803.09179 (2018).Google ScholarGoogle Scholar
  130. Andreas Rossler, Davide Cozzolino, Luisa Verdoliva, Christian Riess, Justus Thies, and Matthias Niessner. 2019. Faceforensics++: Learning to detect manipulated facial images. arXiv preprint:1901.08971 (2019).Google ScholarGoogle Scholar
  131. Ekraam Sabir, Jiaxin Cheng, Ayush Jaiswal, Wael AbdAlmageed, Iacopo Masi, and Prem Natarajan. 2019. Recurrent-convolution approach to DeepFake detection-state-of-art results on FaceForensics++. arXiv preprint arXiv:1905.00582 (2019).Google ScholarGoogle Scholar
  132. Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen. 2016. Improved techniques for training GANs. In Advances in Neural Information Processing Systems. 2234--2242.Google ScholarGoogle Scholar
  133. Sigal Samuel. 2019. A guy made a deepfake app to turn photos of women into nudes. It didn’t go well. https://www.vox.com/2019/6/27/18761639/ai-deepfake-deepnude-app-nude-women-porn.Google ScholarGoogle Scholar
  134. Enrique Sanchez and Michel Valstar. 2018. Triple consistency loss for pairing distributions in GAN-based face synthesis. arXiv preprint arXiv:1811.03492 (2018).Google ScholarGoogle Scholar
  135. Marco Schreyer, Timur Sattarov, Bernd Reimer, and Damian Borth. 2019. Adversarial learning of deepfakes in accounting. arXiv preprint arXiv:1910.03810 (2019).Google ScholarGoogle Scholar
  136. Oscar Schwartz. 2018. You thought fake news was bad? The Guardian. Retrieved March 2, 2020 from https://www.theguardian.com/technology/2018/nov/12/deep-fakes-fake-news-truth.Google ScholarGoogle Scholar
  137. Shawn Shan, Emily Wenger, Jiayun Zhang, Huiying Li, Haitao Zheng, and Ben Y. Zhao. 2020. Fawkes: Protecting privacy against unauthorized deep learning models. In Proceedings of the 29th {USENIX} Security Symposium ({USENIX} Security 20). 1589--1604.Google ScholarGoogle Scholar
  138. Shaoanlu. 2018. faceswap-GAN: A denoising autoencoder + adversarial losses and attention mechanisms for face swapping. Retrieved December 17, 2020 from https://github.com/shaoanlu/faceswap-GAN.Google ScholarGoogle Scholar
  139. Shaoanlu. 2019. fewshot-face-translation-GAN: Generative adversarial networks integrating modules from FUNIT and SPADE for face-swapping. Retrieved from https://github.com/shaoanlu/fewshot-face-translation-GAN.Google ScholarGoogle Scholar
  140. Yujun Shen, Ping Luo, Junjie Yan, Xiaogang Wang, and Xiaoou Tang. 2018. FaceID-GAN: Learning a symmetry three-player GAN for identity-preserving face synthesis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 821--830.Google ScholarGoogle ScholarCross RefCross Ref
  141. Yujun Shen, Bolei Zhou, Ping Luo, and Xiaoou Tang. 2018. FaceFeat-GAN: A two-stage approach for identity-preserving face synthesis. arXiv preprint arXiv:1812.01288 (2018).Google ScholarGoogle Scholar
  142. Taiki Shimba, Ryuhei Sakurai, Hirotake Yamazoe, and Joo-Ho Lee. 2015. Talking heads synthesis from audio with deep neural networks. In Proceedings of the 2015 IEEE/SICE International Symposium on System Integration (SII). IEEE, 100--105.Google ScholarGoogle ScholarCross RefCross Ref
  143. Aliaksandr Siarohin, Stephane Lathuiliere, Sergey Tulyakov, Elisa Ricci, and Nicu Sebe. 2019. Animating arbitrary objects via deep motion transfer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2377--2386.Google ScholarGoogle ScholarCross RefCross Ref
  144. Aliaksandr Siarohin, Stéphane Lathuilière, Sergey Tulyakov, Elisa Ricci, and Nicu Sebe. 2019. First order motion model for image animation. In Advances in Neural Information Processing Systems 32. Curran Associates, Inc., 7135--7145. http://papers.nips.cc/paper/8935-first-order-motion-model-for-image-animation.pdf.Google ScholarGoogle Scholar
  145. Aliaksandr Siarohin, Enver Sangineto, Stephane Lathuiliere, and Nicu Sebe. 2018. Deformable GANs for pose-based human image generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 3408--3416.Google ScholarGoogle ScholarCross RefCross Ref
  146. Yang Song, Jingwen Zhu, Xiaolong Wang, and Hairong Qi. 2018. Talking face generation by conditional recurrent adversarial network. arXiv preprint arXiv:1804.04786 (2018).Google ScholarGoogle Scholar
  147. Jose Sotelo, Soroush Mehri, Kundan Kumar, Joao Felipe Santos, Kyle Kastner, Aaron Courville, and Yoshua Bengio. 2017. Char2wav: End-to-end speech synthesis. Openreview.net (2017).Google ScholarGoogle Scholar
  148. Joel Stehouwer, Hao Dang, Feng Liu, Xiaoming Liu, and Anil Jain. 2019. On the detection of digital face manipulation. arXiv preprint arXiv:1910.01717 (2019).Google ScholarGoogle Scholar
  149. Jeremy Straub. 2019. Using subject face brightness assessment to detect “deep fakes” (Conference Presentation). In Real-Time Image Processing and Deep Learning 2019, Vol. 10996. International Society for Optics and Photonics, 109960H.Google ScholarGoogle ScholarCross RefCross Ref
  150. Qianru Sun, Ayush Tewari, Weipeng Xu, Mario Fritz, Christian Theobalt, and Bernt Schiele. 2018. A hybrid model for identity obfuscation by face replacement. In Proceedings of the European Conference on Computer Vision (ECCV). 553--569.Google ScholarGoogle ScholarDigital LibraryDigital Library
  151. Supasorn Suwajanakorn, Steven M Seitz, and Ira Kemelmacher-Shlizerman. 2017. Synthesizing Obama: Learning lip sync from audio. ACM Transactions on Graphics (TOG) 36, 4 (2017), 1--13.Google ScholarGoogle ScholarDigital LibraryDigital Library
  152. Shahroz Tariq, Sangyup Lee, Hoyoung Kim, Youjin Shin, and Simon S. Woo. 2018. Detecting both machine and human created fake face images in the wild. In Proceedings of the 2nd International Workshop on Multimedia Privacy and Security. ACM, 81--87.Google ScholarGoogle Scholar
  153. Justus Thies, Mohamed Elgharib, Ayush Tewari, Christian Theobalt, and Matthias Niessner. 2019. Neural voice puppetry: Audio-driven facial reenactment. arXiv preprint arXiv:1912.05566 (2019).Google ScholarGoogle Scholar
  154. Justus Thies, Michael Zollhofer, and Matthias Niessner. 2019. Deferred neural rendering: Image synthesis using neural textures. arXiv preprint arXiv:1904.12356 (2019).Google ScholarGoogle Scholar
  155. Justus Thies, Michael Zollhofer, Matthias Niessner, Levi Valgaerts, Marc Stamminger, and Christian Theobalt. 2015. Real-time expression transfer for facial reenactment. ACM Trans. Graph. 34, 6 (2015), 183--1.Google ScholarGoogle ScholarDigital LibraryDigital Library
  156. Justus Thies, Michael Zollhofer, Marc Stamminger, Christian Theobalt, and Matthias Niessner. 2016. Face2face: Real-time face capture and reenactment of rgb videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2387--2395.Google ScholarGoogle ScholarDigital LibraryDigital Library
  157. Justus Thies, Michael Zollhofer, Christian Theobalt, Marc Stamminger, and Matthias Niessner. 2018. Headon: Real-time reenactment of human portrait videos. ACM Transactions on Graphics (TOG) 37, 4 (2018), 1--13.Google ScholarGoogle ScholarDigital LibraryDigital Library
  158. Luan Tran, Xi Yin, and Xiaoming Liu. 2018. Representation learning by rotating your faces. IEEE Transactions on Pattern Analysis and Machine Intelligence 41, 12 (2018), 3007--3021.Google ScholarGoogle ScholarCross RefCross Ref
  159. Soumya Tripathy, Juho Kannala, and Esa Rahtu. 2019. ICface: Interpretable and controllable face reenactment using GANs. arXiv preprint arXiv:1904.01909 (2019).Google ScholarGoogle Scholar
  160. Xiaoguang Tu, Hengsheng Zhang, Mei Xie, Yao Luo, Yuefei Zhang, and Zheng Ma. 2019. Deep transfer across domains for face anti-spoofing. arXiv preprint arXiv:1901.05633 (2019).Google ScholarGoogle Scholar
  161. Sergey Tulyakov, Ming-Yu Liu, Xiaodong Yang, and Jan Kautz. 2018. MocoGAN: Decomposing motion and content for video generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1526--1535.Google ScholarGoogle ScholarCross RefCross Ref
  162. Daniel Vlasic, Matthew Brand, Hanspeter Pfister, and Jovan Popovic. 2006. Face transfer with multilinear models. In ACM SIGGRAPH 2006 Courses. 24–es.Google ScholarGoogle ScholarDigital LibraryDigital Library
  163. Konstantinos Vougioukas, Stavros Petridis, and Maja Pantic. 2019. End-to-end speech-driven realistic facial animation with temporal GANs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops. 37--40.Google ScholarGoogle Scholar
  164. Konstantinos Vougioukas, Stavros Petridis, and Maja Pantic. 2019. Realistic speech-driven facial animation with GANs. arXiv preprint arXiv:1906.06337 (2019).Google ScholarGoogle Scholar
  165. Run Wang, Lei Ma, Felix Juefei-Xu, Xiaofei Xie, Jian Wang, and Yang Liu. 2019. Fakespotter: A simple baseline for spotting AI-synthesized fake faces. arXiv preprint arXiv:1909.06122 (2019).Google ScholarGoogle Scholar
  166. Sheng-Yu Wang, Oliver Wang, Richard Zhang, Andrew Owens, and Alexei A. Efros. 2020. CNN-generated images are surprisingly easy to spot... for now. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vol. 7.Google ScholarGoogle Scholar
  167. Ting-Chun Wang, Ming-Yu Liu, Andrew Tao, Guilin Liu, Jan Kautz, and Bryan Catanzaro. 2019. Few-shot video-to-video synthesis. In Advances in Neural Information Processing Systems (NeurIPS).Google ScholarGoogle Scholar
  168. Ting-Chun Wang, Ming-Yu Liu, Jun-Yan Zhu, Guilin Liu, Andrew Tao, Jan Kautz, and Bryan Catanzaro. 2018. Video-to-video synthesis. In Advances in Neural Information Processing Systems (NeurIPS).Google ScholarGoogle Scholar
  169. Ting-Chun Wang, Ming-Yu Liu, Jun-Yan Zhu, Andrew Tao, Jan Kautz, and Bryan Catanzaro. 2018. High-resolution image synthesis and semantic manipulation with conditional GANs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle ScholarCross RefCross Ref
  170. Yaohui Wang, Piotr Bilinski, Francois Bremond, and Antitza Dantcheva. 2020. ImaGINator: Conditional spatio-temporal GAN for video generation.Google ScholarGoogle Scholar
  171. Olivia Wiles, A. Sophia Koepke, and Andrew Zisserman. 2018. X2face: A network for controlling face generation using images, audio, and pose codes. In Proceedings of the European Conference on Computer Vision (ECCV). 670--686.Google ScholarGoogle ScholarDigital LibraryDigital Library
  172. Michael Workman. 2008. Wisecrackers: A theory-grounded investigation of phishing and pretext social engineering threats to information security. Journal of the American Society for Information Science and Technology 59, 4 (2008), 662--674.Google ScholarGoogle ScholarDigital LibraryDigital Library
  173. Wayne Wu, Yunxuan Zhang, Cheng Li, Chen Qian, and Chen Change Loy. 2018. ReenactGAN: Learning to reenact faces via boundary transfer. In Proceedings of the European Conference on Computer Vision (ECCV). 603--619.Google ScholarGoogle ScholarDigital LibraryDigital Library
  174. Fanyi Xiao, Haotian Liu, and Yong Jae Lee. 2019. Identity from here, pose from there: Self-supervised disentanglement and generation of objects using unlabeled videos. In Proceedings of the IEEE International Conference on Computer Vision. 7013--7022.Google ScholarGoogle ScholarCross RefCross Ref
  175. Runze Xu, Zhiming Zhou, Weinan Zhang, and Yong Yu. 2017. Face transfer with generative adversarial network. arXiv preprint:1710.06090 (2017).Google ScholarGoogle Scholar
  176. Xinsheng Xuan, Bo Peng, Wei Wang, and Jing Dong. 2019. On the generalization of GAN image forensics. In Proceedings of the Chinese Conference on Biometric Recognition. Springer, 134--141.Google ScholarGoogle ScholarCross RefCross Ref
  177. Xin Yang, Yuezun Li, and Siwei Lyu. 2019. Exposing deep fakes using inconsistent head poses. In Proceedings of the ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 8261--8265.Google ScholarGoogle ScholarCross RefCross Ref
  178. Lingyun Yu, Jun Yu, and Qiang Ling. 2019. Mining audio, text and visual information for talking face generation. In Proceedings of the 2019 IEEE International Conference on Data Mining (ICDM). IEEE, 787--795.Google ScholarGoogle ScholarCross RefCross Ref
  179. Ning Yu, Larry S. Davis, and Mario Fritz. 2019. Attributing fake images to GANs: Learning and analyzing GAN fingerprints. In Proceedings of the IEEE International Conference on Computer Vision.Google ScholarGoogle ScholarCross RefCross Ref
  180. Yu Yu, Gang Liu, and Jean-Marc Odobez. 2019. Improving few-shot user-specific gaze adaptation via gaze redirection synthesis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 11937--11946.Google ScholarGoogle ScholarCross RefCross Ref
  181. Polina Zablotskaia, Aliaksandr Siarohin, Bo Zhao, and Leonid Sigal. 2019. DwNet: Dense warp-based network for pose-guided human video generation. arXiv preprint arXiv:1910.09139 (2019).Google ScholarGoogle Scholar
  182. Egor Zakharov, Aliaksandra Shysheya, Egor Burkov, and Victor Lempitsky. 2019. Few-shot adversarial learning of realistic neural talking head models. arXiv preprint arXiv:1905.08233 (2019).Google ScholarGoogle Scholar
  183. Rowan Zellers, Ari Holtzman, Hannah Rashkin, Yonatan Bisk, Ali Farhadi, Franziska Roesner, and Yejin Choi. 2019. Defending against neural fake news. In Advances in Neural Information Processing Systems 32. Curran Associates, Inc., 9054--9065. Retrieved from http://papers.nips.cc/paper/9106-defending-against-neural-fake-news.pdf.Google ScholarGoogle Scholar
  184. Jiangning Zhang, Xianfang Zeng, Yusu Pan, Yong Liu, Yu Ding, and Changjie Fan. 2019. FaceSwapNet: Landmark guided many-to-many face reenactment. arXiv preprint arXiv:1905.11805 (2019).Google ScholarGoogle Scholar
  185. Yunxuan Zhang, Siwei Zhang, Yue He, Cheng Li, Chen Change Loy, and Ziwei Liu. 2019. One-shot face reenactment. arXiv preprint arXiv:1908.03251 (2019).Google ScholarGoogle Scholar
  186. Ying Zhang, Lilei Zheng, and Vrizlynn L. L. Thing. 2017. Automated face swapping and its detection. In Proceedings of the 2017 IEEE 2nd International Conference on Signal and Image Processing (ICSIP). IEEE, 15--19.Google ScholarGoogle Scholar
  187. Lilei Zheng, Ying Zhang, and Vrizlynn L. L. Thing. 2019. A survey on image tampering and its detection in real-world photos. Journal of Visual Communication and Image Representation 58 (2019), 380--399.Google ScholarGoogle ScholarCross RefCross Ref
  188. Hang Zhou, Yu Liu, Ziwei Liu, Ping Luo, and Xiaogang Wang. 2019. Talking face generation by adversarially disentangled audio-visual representation. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33. 9299--9306.Google ScholarGoogle ScholarDigital LibraryDigital Library
  189. Yuqian Zhou and Bertram Emil Shi. 2017. Photorealistic facial expression synthesis by the conditional difference adversarial autoencoder. In Proceedings of the 2017 7th International Conference on Affective Computing and Intelligent Interaction (ACII). IEEE, 370--376.Google ScholarGoogle ScholarCross RefCross Ref
  190. Yipin Zhou, Zhaowen Wang, Chen Fang, Trung Bui, and Tamara L. Berg. 2019. Dance dance generation: Motion transfer for Internet videos. arXiv preprint arXiv:1904.00129 (2019).Google ScholarGoogle Scholar
  191. Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A. Efros. 2017. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision. 2223--2232.Google ScholarGoogle Scholar
  192. Zhen Zhu, Tengteng Huang, Baoguang Shi, Miao Yu, Bofei Wang, and Xiang Bai. 2019. Progressive pose attention transfer for person image generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. The Creation and Detection of Deepfakes: A Survey

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in

        Full Access

        • Published in

          cover image ACM Computing Surveys
          ACM Computing Surveys  Volume 54, Issue 1
          January 2022
          844 pages
          ISSN:0360-0300
          EISSN:1557-7341
          DOI:10.1145/3446641
          Issue’s Table of Contents

          Copyright © 2021 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 2 January 2021
          • Revised: 1 September 2020
          • Accepted: 1 September 2020
          • Received: 1 May 2020
          Published in csur Volume 54, Issue 1

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article
          • Research
          • Refereed

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        HTML Format

        View this article in HTML Format .

        View HTML Format