GReX-Bench: Benchmarking Generalization, Robustness, and Explainability in AI-Generated Image Detection

Read the full article

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Generative AI has significantly transformed media creation and accessibility, enabling the rapid generation of fake content, particularly images. However, AI-generated images pose growing challenges for misinformation and biometric security, contributing to declining public trust, rising fraud, and increasing social engineering attacks.Despite progress in forensics, existing methods still face major limitations: (i) reliance on non-standardized benchmarks, (ii) inconsistent training protocols, (iii) limited evaluation metrics, (iv) weak interpretability, (v) lack of human-readable explanations, and (vi) insufficient attention to deployment and usability in real-world settings. These issues hinder fair comparison and obscure true reliability in security-critical applications.To address this, we introduce \textbf{GReX-Bench}, the first unified benchmarking framework for reproducible evaluation of forensic, anti-forensic (AF), and explainability. We benchmark sixteen prior detectors across eight public datasets (e.g., GAN, diffusion, and low-level vision) under six AF attacks. We analyze model behavior through confidence, ROC, and explainability techniques, including model-specific, model-agnostic, and generative LLM. Our findings reveal significant generalization gaps, with many methods performing well in-distribution but degrading across datasets, particularly under AF attacks. We also examine deployment factors such as efficiency, latency, and scalability to guide practical adoption.

Article activity feed