Micro-Defects Expose Macro-Fakes: Detecting AI-Generated Images via Local Distributional Shifts
Abstract
A local distribution-aware detection framework that amplifies micro-scale statistical irregularities to identify AI-generated images with improved accuracy.
Recent generative models can produce images that appear highly realistic, raising challenges in distinguishing real and AI-generated images. Yet existing detectors based on pre-trained feature extractors tend to over-rely on global semantics, limiting sensitivity to the critical micro-defects. In this work, we propose Micro-Defects expose Macro-Fakes (MDMF), a local distribution-aware detection framework that amplifies micro-scale statistical irregularities into macro-level distributional discrepancies. To avoid localized forensic cues being diluted by plain aggregation, we introduce a learnable Patch Forensic Signature that projects semantic patch embeddings into a compact forensic latent space. We then use Maximum Mean Discrepancy (MMD) to quantify distributional discrepancies between generated and real images. Our theory-grounded analysis shows that patch-wise modeling yields provably larger discrepancies when localized forensic signals are present in generated images, enabling more reliable separation from real images. Extensive experiments demonstrate that MDMF consistently outperforms baseline detectors across multiple benchmarks, validating its general effectiveness. Project page: https://zbox1005.github.io/MDMF-project/
Community
Overview
MDMF (Micro-Defects expose Macro-Fakes) reframes AI-generated image detection as a local distributional problem rather than an image-level classification one. Instead of compressing each image into a single representation that tends to over-rely on global semantics, MDMF treats every image as a collection of patches, projects each patch into a learnable forensic latent space — the Patch Forensic Signature (PFS) — and measures the distributional discrepancy between the test image and a small reference bank of clean real images via a Maximum Mean Discrepancy (MMD) score.
Key Highlights
- Patch Forensic Signature (PFS) — A learnable forensic reparameterization of frozen DINOv2 patch tokens that suppresses semantic invariances and amplifies generation-induced statistical irregularities.
- MDMF detector — A distribution-aware detection framework that aggregates patch-level evidence into a stable image-level score via MMD between PFS distributions, avoiding the per-patch decision boundary that destabilizes hard-voting baselines.
- Theory-grounded — We prove that patch-wise PFS modeling yields a provably larger MMD signal than global pooling whenever localized forensic cues are present, and we derive a finite-sample separation guarantee with a finite optimal patch count $K^\star$.
- Strong cross-generator generalization — Trained on a single 4-class ProGAN split, MDMF reaches 95.65 average AUROC on the ImageNet benchmark (9 generators spanning diffusion, GAN, and AR families) and remains state-of-the-art on LSUN-Bedroom, GenImage, WildRF, LDMFakeDetect, and an OpenSora video-frame stress test, with markedly gentler degradation under JPEG, blur, and noise post-processing than the strongest training-based baseline.
Get this paper in your agent:
hf papers read 2605.09296 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper