Discovery of Deepfakes in Art

Authors

  • Alexander I. Iliev SRH University of Applied Sciences, Ernst-Reuter-Platz 10, 10587 Berlin, Germany; Institute of Mathematics and Informatics, Bulgarian Academy of Sciences, Acad. Georgi Bonchev Str., Block 8, 1113 Sofia, Bulgaria

DOI:

https://doi.org/10.55630/dipp.2025.15.5

Keywords:

Deepfakes, Generative Models, Art Forensics, GANs, Transformers, Multimedia AI

Abstract

The proliferation of deepfakes—AI-generated or manipulated media—has transformed the landscape of contemporary art. Deep generative models, including GANs, VAEs, diffusion models, and Transformers, have enabled artists to explore new creative realms while simultaneously raising critical questions around authenticity, ethics, and detection. This paper presents a comprehensive analysis of deepfake technologies across five key media modalities: image, video, text, speech, and music. We examine the architectures that enable content creation, and the state-of-the-art techniques used for detection. Further, we evaluate detection accuracy, robustness, and practical implementation, incorporating diagrams, comparative tables, and performance formulas. This work aims to provide a balanced perspective on the opportunities and challenges posed by synthetic media in the artistic domain.

References

Alrashoud, M. (2025). Ethical boundaries of AI-generated content in digital culture. Journal of Emerging Media and Society, 14 (2), 133–150.

Babael, F., Amini, S., & Chen, Y. (2025). Multimodal deepfake detection in crossmedia environments. In Proceedings of the International Conference on Multimedia Security (pp. 45–59).

Chakraborty, A., Nair, V., Singh, R., & Joshi, P. (2024). Hybrid attention GAN for consistent face swapping in artistic performances. In Proceedings of the International Conference on Creative AI (pp. 22–34).

Dolhansky, B., Bitton, J., Pflaum, B., Lu, J., Howes, R., Wang, M., & Ferrer, C. C. (2020). The DeepFake Detection Challenge (DFDC) dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2020. https://ieeexplore.ieee.org/document/9156360

Gao, Y., Lin, Q., & Wu, T. (2024). StyleMusic: A cross-modal generative framework for stylized AI music composition. Journal of Neural Audio Processing, 17 (2), 110– 125.

Ignatova, D. (2025). Application of wellness program for the prevention of aggressive behavior. Scientific Journal "Kindergarten School", 6 (1), 60–116.

Jain, M., & Aggarwal, R. (2024). Understanding and detecting deepfakes across visual and audio modalities. International Journal of Artificial Intelligence Research, (1), 78–99.

Mlynar, V., Polák, M., & Kruliš, M. (2023). DeepTextMark: A deep learning-driven text watermarking approach. Applied Sciences, 13 (17), 9875. https://www.mdpi.com/2076-3417/13/17/9875

Nguyen, T. T., Nguyen, Q. V. H., Nguyen, D. T., Nguyen, D. T., Huynh-The, T., Nahavandi, S., Nguyen, T. T., Pham, Q.-V., & Nguyen C. M. (2019) ‘ Deep Learning for Deepfakes Creation and Detection: A Survey ’. Computer Vision and Image Understanding, 223, Article 103525. https://doi.org/10.1016/j.cviu.2022.103525.

Ranjan, S., Li, H., & Das, T. (2023). DetectGPT and the future of synthetic text forensics. In Proceedings of the Computational Linguistics and AI Ethics Symposium (pp. 87–101).

Sabatelli, M., Sutherland, M., Cakmakci, O., & Reinecke, K. (2023). Organic or diffused: Can we distinguish human art from AI-generated images? In Proceedings of the 28th International Conference on Intelligent User Interfaces (IUI '23). https://dl.acm.org/doi/abs/10.1145/3576915.3623189

Solaiman, I., Clark, J., & Zhang, J. (2024). Human vs. machine: A comparative study on the detection of AI-generated content. arXiv Preprint, arXiv:2402.03214. https://arxiv.org/abs/2402.03214

Vasudeva, N., Kapoor, A., & Iyer, S. (2024). DeepTextMark: Watermarking-based authorship verification for AI-generated text. In Advances in AI for Digital Humanities (pp. 54–69). Springer.

Wolter, M., Blanke, F., Heese, R. & Garcke, J. (2022) Wavelet-packets for deepfake image analysis and detection. Machine Learning, 111, 4295–4327. https://doi.org/10.1007/s10994-022-06225-5

Zhou, L., He, K., & Yuan, M. (2025). UC-VITS: Cross-lingual voice cloning and emotional speech synthesis via variational inference. Transactions on Speech Technologies, 29 (1), 15–37.

Downloads

Published

2025-09-05

How to Cite

I. Iliev, A. (2025). Discovery of Deepfakes in Art. Digital Presentation and Preservation of Cultural and Scientific Heritage, 15, 55–64. https://doi.org/10.55630/dipp.2025.15.5

Most read articles by the same author(s)