Reliability-Oriented Evaluation of Explanation Signals under Real and Synthetic Stress Tests

Read the full article See related articles

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

This paper presents a reliability-oriented empirical study of global importance signals in machine learning models, focusing on stability, sensitivity, and cross-model consistency rather than proposing a new explanation algorithm. (i) Across two public real-world datasets and controlled synthetic stress tests, we show that explanation reliability can degrade even when predictive performance remains strong (often high); (ii) we provide empirical evidence that stability, sensitivity, and cross-model agreement capture complementary failure modes overlooked by accuracy-centric evaluation; and (iii) we discuss implications for reliability-aware XAI benchmarking under spurious correlations and retraining variability. These findings treat explanation reliability as a primary empirical property in our evaluated setting and offer practical guidance under the considered regimes for building and using future XAI benchmarks.

Article activity feed