CLaRE ๐ต๏ธโโ๏ธ
CLaRE ๐ต๏ธโโ๏ธ
This project introduces CLaRE, a deepfake detection method that builds on CLIP by integrating Latent Reconstruction Error (LaRE) [1] with Context Optimization (CoOp) [2] or Conditional Context Optimization (CoCoOp) [3] for adaptive prompt learning. While CLIP with learnable prompts offers strong generalization, it struggles with realistic forgeries from modern generative tools. CLaRE addresses this gap and is evaluated on the DF40 dataset across multiple training regimes. It shows competitive performance on GAN-generated images and surpasses prompt-tuned CLIP on diffusion-based fakes.
The code for our study is available at: link
Note: The poster does not represent the final camera-ready version.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
@inproceedings{10.1145/3733813.3764366,
author = {Thakur, Udit and Khan, Mohammad Hafeez and Changlani, Meher and Dotsinski, Asen and Mahadevan, Aswin Krishna and Kechagias, Ioannis and Najdenkoska, Ivona},
title = {CLaRE: CLIP with Latent Reconstruction Errors for Generated Face Detection},
year = {2025},
isbn = {9798400719028},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3733813.3764366},
doi = {10.1145/3733813.3764366},
booktitle = {Proceedings of the 1st ACM Workshop on Deepfake, Deception, and Disinformation Security},
pages = {5โ15},
numpages = {11},
keywords = {Deepfake detection, Diffusion models, CLIP, Reconstruction error, Prompt tuning},
series = {3D-Sec '25}
}
References
- [1] Y. Luo, J. Du, K. Yan, and S. Ding, โLaRE^2: Latent Reconstruction Error Based Method for Diffusion-Generated Image Detection,โ in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 17006โ17015.
- [2] K. Zhou, J. Yang, C. C. Loy, and Z. Liu, โLearning to Prompt for Vision-Language Models,โ International Journal of Computer Vision, vol. 130, no. 9, pp. 2337โ2348, Jul. 2022, doi: 10.1007/s11263-022-01653-1.
- [3] K. Zhou, J. Yang, C. C. Loy, and Z. Liu, โConditional prompt learning for vision-language models,โ in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 16816โ16825.
This post is licensed under CC BY 4.0 by the author.