The Paradox of Explainability vs. Performance in High-Stakes Autonomous AI Systems: A Systematic Review of Trade-offs, Regulatory Gaps, and Emerging Solutions
Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
This study systematically examines the inherent trade-offs between explainability and performance in high-stakes autonomous AI systems operating within critical domains such as healthcare, criminal justice, finance, and aviation. Employing a PRISMA protocol-led systematic literature review, 45 peer-reviewed articles were selected through rigorous inclusion and exclusion criteria. The findings reveal a persistent tension: while complex models, especially deep learning architectures, achieve superior predictive accuracy, they often lack interpretability, undermining trust, accountability, and regulatory compliance in sensitive decision-making environments. The study emphasizes that explainability is not a peripheral concern but a foundational requirement in applications with profound ethical, legal, and societal implications. It advocates for the development and adoption of hybrid and context-sensitive AI frameworks that balance predictive performance with interpretability. By addressing the governance, user trust, and technical dimensions of this trade-off, the research offers critical insights for both AI developers and policymakers working toward responsible, transparent, and human-aligned AI deployment