Poster
Origin Identification for Text-Guided Image-to-Image Diffusion Models
Wenhao Wang · Yifan Sun · Zongxin Yang · Zhentao Tan · Zhengdong Hu · Yi Yang
West Exhibition Hall B2-B3 #W-217
Modern AI tools can now rewrite any picture just by following a short text instruction—turning a daytime street into a rainy night scene, or adding new objects that never existed. While fun and useful, this power makes it easy to spread fake images, dodge copyright rules, and hide the true source of a picture. We tackle this problem by asking a simple but crucial question: given an edited image, can we reliably find the original photo it came from?Our new task, called ID² (Origin IDentification for text-guiding Image-to-image Diffusion), shows why earlier “look-for-similar-parts” tricks break down: different AI editors leave very different fingerprints, so a system trained on one often fails on another. To fix this, we built OriPID, the first large benchmark that pairs thousands of originals with their AI-altered versions from many diffusion models. We then prove that a single linear tweak to the images’ hidden VAE features can pull each edited picture back toward its source—and that this tweak works across editors. In tests, our lightweight method beats previous similarity-based approaches by over 31 percentage points in mean average precision, bringing practical image provenance a big step closer. Code and data: id2icml.github.io.