MIMS Final Project 2025

MADDÉ: Model-Agnostic Deepfake Detection and Explainability

Motivation

In recent years, deepfakes have emerged as a prominent and concerning development in the fields of media and artificial intelligence [1]. With advancements in generative models, the creation of synthetic content is becoming more sophisticated, making it increasingly challenging to distinguish real from fake [2], [3]. The implications of this technology are vast, impacting areas like misinformation, security, and personal privacy [4], [5]. As deepfake techniques evolve, so too must our detection methods [6]. Addressing this issue is essential to maintain trust in digital media and protect against malicious misuse of synthetic content [7].

What We Are Doing

In the current deepfake research landscape, we identify two crucial areas for the future development of deepfake detection systems:

  1. Pattern Detection
    The ability to detect patterns across various types of generative models is critical in deepfake detection [8], [9]. Generative techniques are rapidly advancing, with models like GANs, autoencoders, and more recently, diffusion models, each introducing unique artifacts and synthesis characteristics [10]. Identifying common patterns across these diverse generative methods is essential for building robust and generalizable detection systems that are not limited to a specific model type [11]. This approach is necessary to future-proof deepfake detection as new generative techniques continue to emerge [12].
  2. Explainability
    Explainability in deepfake detection models is vital, as it enhances trust in the detection process and provides clear insights into how the model arrives at its decisions [13], [14]. Given the potential societal impact of deepfakes, it is crucial that detection models not only achieve high accuracy but also offer interpretable outputs that allow users to understand why a particular image or video has been classified as fake [15]. This explainability is particularly important in applications involving law enforcement, media verification, and public safety, where transparency is key to ethical AI use [16].

References

  1. Z. Wang et al., "DIRE for Diffusion-Generated Image Detection," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2023.
  2. X. Guo et al., "Hierarchical Fine-Grained Image Forgery Detection and Localization," in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2022.
  3. H. Farid, "Creating, Using, Misusing, and Detecting Deep Fakes," Communications of the ACM, vol. 64, no. 11, pp. 56-64, 2021.
  4. S. Mundra et al., "Exposing GAN-Generated Profile Photos from Compact Embeddings," in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2021.
  5. W. Xie and A. Zisserman, "ForensicTransfer: Weakly-supervised Domain Adaptation for Forgery Detection," in Proceedings of the British Machine Vision Conference (BMVC), 2020.
  6. P. Korshunov and T. Ebrahimi, "Deepfakes: A New Threat to Face Recognition? Assessment and Detection," IEEE Transactions on Information Forensics and Security, vol. 15, pp. 1193-1204, 2020.
  7. Y. Zhang et al., "Detecting GAN-Generated Fake Images using Co-occurrence Matrices," in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2019.
  8. R. Wang et al., "Revisiting the GAN Training Procedure: Perspectives on Stability, Fidelity, and Diversity," International Journal of Computer Vision, vol. 128, no. 9, pp. 2455-2475, 2020.
  9. A. Tolstikhin et al., "MLP-Mixer: An All-MLP Architecture for Vision," in Advances in Neural Information Processing Systems (NeurIPS), 2021.
  10. Y. Kong et al., "Ensemble of Deep Convolutional Neural Networks for Image Steganalysis," in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2021.
  11. L. Chen et al., "DRUNet: Denoising Residual U-Net for Image Restoration," IEEE Transactions on Image Processing, vol. 29, pp. 607-620, 2020.
  12. K. Lee et al., "Detecting Deepfake Videos from the Frequency Domain," IEEE Signal Processing Letters, vol. 27, pp. 1360-1364, 2020.
  13. D. D. Tan et al., "A Survey on Deepfake Detection and Analysis Techniques," IEEE Access, vol. 9, pp. 149957-149972, 2021.
  14. A. Rozsa and T. Boult, "Explainable Deepfake Detection: Analyzing the Interpretability of Neural Networks," in Proceedings of the IEEE International Conference on Pattern Recognition (ICPR), 2022.
  15. X. Hu et al., "A Comprehensive Review of Explainable Artificial Intelligence (XAI) for Deepfake Detection," IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 12, pp. 5609-5623, 2021.
  16. M. S. Yoon et al., "Saliency-Guided Deepfake Detection: Improving Explainability in Detection Models," IEEE Transactions on Multimedia, vol. 23, pp. 1826-1839, 2021.
Last updated: January 27, 2025