Ijraset Journal For Research in Applied Science and Engineering Technology
Authors: Aditi Indre, Jayesh Shinde, Dr. Srivaramangai Ramanujam
DOI Link: https://doi.org/10.22214/ijraset.2025.67796
Certificate: View Certificate
With digital manipulation blurring the lines between reality and fabrication, deepfakes have become one of the most shocking threats to media authenticity and public trust. Fed by advanced artificial intelligence techniques like Generative Adversarial Networks (GANs) and deep learning, deepfakes can create hyper-realistic videos, images, and audio that convincingly imitate real people. These forgeries have far-reaching implications, from spreading political misinformation and financial fraud to non-consensual explicit content and identity theft. As deepfake technology becomes more sophisticated and accessible, traditional detection methods struggle to keep pace, sparking an urgent technological arms race between creators and defenders. This survey discusses the rapidly changing landscape of deepfake detection, focusing on multimodal approaches that combine visual, audio, and physiological cues to improve detection accuracy. Combining facial landmark analysis, voice consistency checks, and even heart rate variability extraction, multimodal detection methods provide a strong defense against increasingly complex forgeries. We discuss state-of-the-art models, including Convolutional Neural Networks (CNNs), Vision Transformers, and XGBoost classifiers, along with their evaluation against challenging datasets like Celeb-DF and FaceForensics++. Lastly, the paper emphasizes the problems in current detection techniques, especially in real- time processing and the generalization problems of deepfakes across a wide variety of generation methods. This paper helps researchers and practitioners navigate the battlefield of digital deception by identifying the critical gaps of existing research and proposing future directions.
Summary:
Deepfakes, created using AI techniques like Generative Adversarial Networks (GANs), enable realistic manipulation of faces, voices, and expressions in videos and images. While innovative, deepfakes pose serious threats to media integrity, privacy, and public trust by enabling misinformation, non-consensual content, and fraud. The creation involves two neural networks—a generator and a discriminator—that together produce highly realistic fake media. Tools like FaceSwap have made deepfake creation accessible to non-experts, increasing concerns about misinformation and identity theft.
Despite risks, deepfakes also have positive uses in entertainment, education, and healthcare, such as de-aging actors or recreating historical figures. However, as generation techniques improve, detection methods must evolve to keep pace.
The paper surveys advances in deepfake detection, covering traditional machine learning (e.g., XGBoost), deep learning models (CNNs, Vision Transformers), and multimodal approaches combining visual and audio cues. Key datasets like Celeb-DF and FaceForensics++ support model training and benchmarking. Challenges include real-time detection, model generalization, adversarial attacks, and ethical issues.
Recent methods focus on subtle artifacts, physiological signals (like eye blinking), audio verification, and temporal inconsistencies. Defense strategies include proactive perturbations to disrupt fake generation and watermarking techniques. Hybrid and attention-based models improve detection accuracy, while new datasets reflect real-world complexities.
The ongoing “arms race” between deepfake creation and detection demands continuous innovation. The paper underscores the societal impact of deepfakes on politics, business, and privacy, and calls for combined efforts in detection technology, legal frameworks, and public awareness to mitigate risks while harnessing benefits.
In summary, although deepfake detection research has achieved great milestones, more gaps still need to be filled in order to really respond to the ever-evolving complexity of deepfake technologies. The state-of-the-art methods, which integrate heart rate analysis, facial feature recognition, attention mechanisms, etc., have promising roles in improving detection accuracy and robustness; these are some good generalization frameworks with a capability for real-world adaptability. Other quality-oriented approaches, such as Forgery Quality Scores and synthetic data generation, have shown promise in boosting model performance to ensure that the detection methods are more robust on diverse datasets. However, several key advancements need to be achieved to bridge the remaining gaps. First, more sophisticated and novel deepfake techniques are challenging to detect. Most of the models fail to generalize across new manipulation techniques, especially when it is unseen data or more sophisticated forgeries. Hence, adaptive methods such as adversarial training and dynamic curriculum learning are required in order to keep updating the model and improving the resilience against emerging threats. Better integration of multi-modal detection approaches that combine facial, audio, and behavioral cues will likely be essential for tackling a wide range of deepfakecontent.Although tremendous progress has been achieved in the accuracy of detection, the next steps would involve ensuring that such models are interpretable and transparent. Most current methods are very complex and not clear enough to be used in real-world applications, especially those in legal or forensic applications. Detection systems should be interpretable, explainable, and adaptive to new manipulation techniques for them to be useful in the long run. Finally, there is an urgent need to increase the number of real-world datasets, particularly those that capture diverse and natural scenarios, as most of the existing datasets tend to focus on specific types of forgeries or controlled environments. Improving the diversity of the dataset and the use of real-time data will make models more adaptable and accurate in practical situations. In summary, deepfake detection has made tremendous progress. However, the technology itself continues to evolve, and there is still much to be discovered on developing adaptable, interpretable, and real-time solutions. Coordinative work across research, technology, and policy spheres will be crucial towards bridging the identified gaps so that detection systems stay up with the evolving nature of deepfake technology.
[1] Muhammad Umar Farooq, Ali Javed, Khalid Mahmood Malik, Muhammad Anas Raza “A Lightweight and Interpretable Deepfakes Detection Framework”, arXiv:2501.11927, 2025 [2] Wentang Song, Zhiyuan Yan, Yuzhen Lin, Taiping Yao, Changsheng Chen, Shen Chen, Yandan Zhao, Shouhong Ding, Bin Li, “A Quality-Centric Framework for Generic Deepfake Detection”, arXiv:2411.05335, 202425 [3] Yuezun Li, Xin Yang, Pu Sun, Honggang Qi, SiweiLyu “Celeb-DF- A Large-scale Challenging Dataset for DeepFake Forensics”, , arXiv:1909.12962, 2019. [4] Dragos-AlexandruBoldisor, Stefan Smeu, Dan Oneata, ElisabetaOneata “Circumventing shortcuts in audio-visual deepfake detection datasets with unsupervised learning”, arXiv:2412.00175, 2024. [5] Bahar Uddin Mahmud, AfsanaSharmin, “Deep Insights of DeepfakeTechnology : A Review”, arXiv:2105.00192, 2021 [6] Amal Naitali, Mohammed Ridouani , Fatima Salahdineand Naima Kaabouch , “Deepfake Attacks: Generation, Detection, Datasets, Challenges, and Research Directions”, In MDPI, Computers 2023. [7] Falguni Sharma, Priyanka Gupta “Deepfake Detection of Singing Voices With Whisper Encodings”, arXiv:2501.18919v1, 2025 [8] Md ShoRana, Mohammad Nur Nobi, Beddhu Murali, Andrew H. Sung , “Deepfake Detection: A Systematic Literature Review” , In IEEE, Jan-2022 [9] Yogesh Patel, Sudeep Tanwar, Rajesh Gupta, Pronaya Bhattacharya, Innocent Ewean Davidson, RoyiNyameko “Deepfake Generation and Detection Case Study and Challenges ”, In IEEE December 2023 , Digital Object Identifier 10.1109/ACCESS.2023.3342107 [10] Florinel-Alin Croitoru, Andrei-Iulian Hiji, Vlad Hondru, Nicolae CatalinRistea, Paul Irofti, Marius Popescu, Cristian Rusu, Radu Tudor Ionescu, Fahad Shahbaz Khan, Mubarak Shah “Deepfake Media Generation and Detection in the Generative AI Era: A Survey and Outlook”, arXiv:2411.19537, Nov 2024. [11] Pavel Korshunov, Sebastien Marcel , “DeepFakes: a New Threat to Face Recognition? Assessment and Detection”, arXiv:1812.08685, Dec 2018 [12] Jan Kietzmann, Linda W. Lee , Ian P. McCarthy, Tim C. Kietzmann , “Deepfakes: Trick or Treat?”, Business Horizons, Vol. 63, 2020 [13] Jiazhi Guan, Hang Zhou, Zhibin Hong, Errui Ding, Jingdong Wang, Chengbin Quan, Youjian Zhao , “Delving into Sequential Patches for Deepfake Detection”, arXiv:2207.02803 , 2022 [14] Kaede Shiohara Toshihiko Yamasaki, “Detecting Deepfakes with Self-Blended Images”, In IEEE, June 2022 , DOI: 10.1109/CVPR52688.2022.01816 [15] MD Sadik Hossain Shanto, Mahir Labib Dihan, Souvik Ghosh, Riad Ahmed Anonto, Hafijul Hoque Chowdhury, AbirMuhtasim, Rakib Ahsan, MD Tanvir Hassan, MD RoqunuzzamanSojib, Sheikh Azizul Hakim, M. Saifur Rahman , “DFCon: Attention-Driven Supervised Contrastive Learning for Robust Deepfake Detection”, arXiv:2501.16704 , Jan 2025 [16] Peipeng Yu, Hui Gao, Zhitao Huang, Zhihua Xia, Chip-Hong Chang “DFREC: DeepFake Identity Recovery Based on Identity-aware Masked Autoencoder”, arXiv:2412.07260, Dec 2024 [17] Xinghe Fu, Zhiyuan Yan, Taiping Yao, Shen Chen, Xi Li, “Exploring Unbiased Deepfake Detection via Token-Level Shuffling and Mixing”, arXiv:2501.04376v1, Jan 2025 [18] Xin Yang, Yuezun Li, SiweiLyu, “Exposing Deep Fakes Using Inconsistent Head Poses”, arXiv:1811.00661, 2018 [19] Veronika Solopova, Lucas Schmidt, Dorothea Kolossa, “Extending Information Bottleneck Attribution to Video Sequences”, arXiv:2501.16889, Jan 2025 [20] Lingzhi Li, Jianmin Bao, Ting Zhang, Hao Yang, Dong Chen, Fang Wen, Baining Guo, “Face X-ray for More General Face Forgery Detection”, arXiv:1912.13458, Apr 2020. [21] Justus Thies, Michael Zollhöfer, Marc Stamminger, Christian Theobalt, Matthias Nießner“Face2Face: Real-time Face Capture and Reenactment of RGB Videos”, arXiv:2007.14808, Jul 2020 [22] Andreas Rössler, Davide Cozzolino, Luisa Verdoliva, Christian Riess, Justus Thies, Matthias Nießner “FaceForensics++: Learning to Detect Manipulated Facial Images”, arXiv:1901.08971, Jan 2019 [23] Andreas Rössler, Davide Cozzolino, Luisa Verdoliva, Christian Riess, Justus Thies, Matthias Nießner , “FaceForensics: A Large-scale Video Dataset for Forgery Detection in Human Faces”, arXiv:1803.09179 , Mar 2018 [24] JaehwanJeong, Sumin In, Sieun Kim, Hannie Shin, JongheonJeong, Sang Ho Yoon, Jaewook Chung, Sangpil Kim, “FaceShield: Defending Facial Image against Deepfake Threats” rXiv:2412.09921, Dec 2024 [25] Lingzhi Li, Jianmin Bao, Hao Yang, Dong Chen, Fang Wen, “FaceShifter: Towards High Fidelity And Occlusion Aware Face Swapping”, arXiv:1912.13457, Dec 2019 [26] Shulin Lan, Kanlin Liu, Yazhou Zhao, Chen Yang, Yingchao Wang, Xingshan Yao, Liehuang Zhu, “Facial Features Matter: a Dynamic Watermark based Proactive Deepfake Detection Approach”, arXiv:2411.14798, Nov 2024 [27] Younhun Kim, Myung-Joon Kwon, Wonjun Lee, Changick Kim, “FRIDAY: Mitigating Unintentional Facial Identity in Deepfake Detectors Guided by Facial Recognizers”, In IEEE Xplore January 2025, DOI: 10.1109/VCIP63160.2024.10849915 [28] Gaojian Wang, Feng Lin, Tong Wu, Zhenguang Liu, Zhongjie Ba, Kui Ren, “FSFM: A Generalizable Face Security Foundation Model via Self-Supervised Facial Representation Learning” , arXiv:2412.12032 , Dec 2024 [29] Wen Huang, Yanmei Gu, Zhiming Wang, Huijia Zhu, Yanmin Qian, “Generalizable Audio Deepfake Detection via Latent Space Refinement and Augmentation”, arXiv:2501.14240 , Jan 2025 [30] Jiazhen Yan, Ziqiang Li, Ziwen He, Zhangjie Fu, “Generalizable Deepfake Detection via Effective Local-Global Feature Extraction”, arXiv:2501.15253, Jan 2025 [31] Huikai Wu, Shuai Zheng, Junge Zhang, Kaiqi Huang, “GP-GAN: Towards Realistic High-Resolution Image Blending” ,arXiv:1703.07195 , Mar 2017 [32] Yuezun Li, Xin Yang, Baoyuan Wu, SiweiLyu, “Hiding Faces in Plain Sight: Disrupting AI Face Synthesis with Adversarial Perturbations”, arXiv:2412.01101, Dec 2024 [33] Zirui Zhang, Wei Hao, AroonSankoh, William Lin, Emanuel Mendiola-Ortiz, Junfeng Yang, Chengzhi Mao, “I Can Hear You: Selective Robust Training for Deepfake Audio Detection”, arXiv:2411.00121v1, Oct 2024 [34] Yuezun Li, Ming-Ching Chang, SiweiLyu, “In Ictu Oculi: Exposing AI Created Fake Videos by Detecting Eye Blinking”, Published in 2018 IEEE International Workshop on Information Forensics and Security (WIFS), January 2019 [35] Joon Son Chung, Andrew Senior, Oriol Vinyals, Andrew Zisserman, “Lip Reading Sentences in the Wild”, Published in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017 [36] Jyoti Bhadana, Michael A. Kouritzin, Seoyeon Park, Ian Zhang, “Markov Processes for Enhanced Deepfake Generation and Detection”, arXiv:2411.07993, Nov 2024 [37] Hanqing Zhao, Tianyi Wei, Wenbo Zhou, Weiming Zhang, Dongdong Chen, Nenghai Yu, “Multi-attentional Deepfake Detection”, Published in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), November 2021, DOI: 10.1109/CVPR46437.2021.00222 [38] Zhiyuan Yan, Jiangming Wang, Peng Jin, Ke-Yue Zhang, Chengchun Liu, Shen Chen,Taiping Yao, Shouhong Ding, Baoyuan Wu, Li Yuan, “Orthogonal Subspace Decomposition for Generalizable AI-Generated Image Detection”, arXiv:2411.15633v2, Jan 2025 [39] Hong-Hanh Nguyen-Le, Van-Tuan Tran, Dinh-Thuc Nguyen, Nhien-An Le-Khac, “Passive Deepfake Detection Across Multi-modalities: A Comprehensive Survey”, arXiv:2411.17911, Nov 2024 [40] Kuiyuan Zhang, Zhongyun Hua, Rushi Lan, Yushu Zhang, Yifang Guo, “Phoneme-Level Feature Discrepancies: A Key to Detecting Sophisticated Speech Deepfakes”, arXiv:2412.12619, 17 Dec 2024 [41] Taesung Park, Ming-Yu Liu, Ting-Chun Wang, Jun-Yan Zhu, “Semantic Image Synthesis with Spatially-Adaptive Normalization”,arXiv:1903.07291, Mar 2019 [42] Yongyi Zang, You Zhang, MojtabaHeydari, ZhiyaoDuan,” SingFake: Singing Voice Deepfake Detection”, arXiv:2309.07525, Sep 2023 [43] Honggu Liu, Xiaodan Li, Wenbo Zhou, Yuefeng Chen, Yuan He, Hui Xue, Weiming Zhang, Nenghai Yu , “Spatial-Phase Shallow Learning: Rethinking Face Forgery Detection in Frequency Domain”, arXiv:2103.01856 , Mar 2021 [44] Mika Westerlund, “The Emergence of DeepfakeTechnology:A Review”, Technology Innovation Management Review, November 2019 [45] Yuyang Qian, Guojun Yin, Lu Sheng, Zixuan Chen, Jing Shao, “Thinking in Frequency: Face Forgery Detection by Mining Frequency-aware Clues”, arXiv:2007.09355 , Jul 2020 [46] Wentang Song, Yuzhen Lin, Bin Li, “ Towards General Deepfake Detection with Dynamic Curriculum ”, arXiv:2410.11162 , Oct 2024 [47] Zhiyuan Yan, Yuhao Luo, SiweiLyu, Qingshan Liu, Baoyuan Wu , “Transcending Forgery Specificity with Latent Space Augmentation for Generalizable Deepfake Detection”, arXiv:2311.11278, Nov 2023 \\ [48] Hannah Lee, Changyeon Lee, Kevin Farhat, Lin Qiu, Steve Geluso, Aerin Kim, Oren Etzioni, “The Tug-of-War Between Deepfake Generation and Detection”, arXiv:2407.06174 , Jul 2024 [49] Dat Nguyen, Marcella Astrid, Anis Kacem, EnjieGhorbel, DjamilaAouada, “Vulnerability-Aware Spatio-Temporal Learning for Generalizable and Interpretable Deepfake Video Detection”, arXiv:2501.01184, Jan 2025 [50] Bojia Zi, Minghao Chang, Jingjing Chen, Xingjun Ma, Yu-Gang Jiang, “WildDeepfake: A Challenging Real-World Dataset for Deepfake Detection”, arXiv:2101.01456 , 5 Jan 2021
Copyright © 2025 Aditi Indre, Jayesh Shinde, Dr. Srivaramangai Ramanujam . This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Paper Id : IJRASET67796
Publish Date : 2025-03-23
ISSN : 2321-9653
Publisher Name : IJRASET
DOI Link : Click Here