TY - GEN
T1 - DDRE
T2 - 8th Chinese Conference on Pattern Recognition and Computer Vision, PRCV 2025
AU - Li, Mengcheng
AU - Chao, Fei
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2026.
PY - 2026/1/24
Y1 - 2026/1/24
N2 - The growing diversity of generative models has made distinguishing AI-generated images increasingly challenging, especially under unknown generator scenarios. While diffusion reconstruction-based methods have shown promise in detecting generated content, a key limitation persists: the entanglement between artifact and image content distributions during reconstruction. In this paper, we propose DDRE (Decoupled Diffusion Reconstruction Error), a novel, training-free detection framework that explicitly disentangles these two distributions via a two-stage reconstruction process. By comparing reconstruction error maps from successive denoising stages and applying a simple norm-based ratio metric, DDRE effectively isolates generative artifacts. Furthermore, we introduce a self-conditioning strategy that leverages CLIP embeddings to guide reconstruction in the absence of prompts, enhancing the fidelity of inversion. To provide a more comprehensive evaluation, we propose GenNEXT, a benchmark that extends GenImage by reducing JPEG bias and incorporating more recent generative models. Extensive experiments on GenNEXT demonstrate that DDRE achieves state-of-the-art performance and exhibits strong generalization to previously unseen generators, all without requiring additional training.
AB - The growing diversity of generative models has made distinguishing AI-generated images increasingly challenging, especially under unknown generator scenarios. While diffusion reconstruction-based methods have shown promise in detecting generated content, a key limitation persists: the entanglement between artifact and image content distributions during reconstruction. In this paper, we propose DDRE (Decoupled Diffusion Reconstruction Error), a novel, training-free detection framework that explicitly disentangles these two distributions via a two-stage reconstruction process. By comparing reconstruction error maps from successive denoising stages and applying a simple norm-based ratio metric, DDRE effectively isolates generative artifacts. Furthermore, we introduce a self-conditioning strategy that leverages CLIP embeddings to guide reconstruction in the absence of prompts, enhancing the fidelity of inversion. To provide a more comprehensive evaluation, we propose GenNEXT, a benchmark that extends GenImage by reducing JPEG bias and incorporating more recent generative models. Extensive experiments on GenNEXT demonstrate that DDRE achieves state-of-the-art performance and exhibits strong generalization to previously unseen generators, all without requiring additional training.
KW - Artifact Disentanglement
KW - Diffusion Inversion
KW - Image Forgery Detection
UR - https://www.scopus.com/pages/publications/105028984680
U2 - 10.1007/978-981-95-5764-6_19
DO - 10.1007/978-981-95-5764-6_19
M3 - Conference Proceeding (ISBN)
AN - SCOPUS:105028984680
SN - 9789819557639
T3 - Lecture Notes in Computer Science
SP - 273
EP - 287
BT - Pattern Recognition and Computer Vision - 8th Chinese Conference, PRCV 2025, Proceedings
A2 - Kittler, Josef
A2 - Xiong, Hongkai
A2 - Lin, Weiyao
A2 - Yang, Jian
A2 - Chen, Xilin
A2 - Lu, Jiwen
A2 - Yu, Jingyi
A2 - Zheng, Weishi
PB - Springer Nature
Y2 - 15 October 2025 through 18 October 2025
ER -