A Comprehensive Review of Explainable AI (XAI) Methods in Deep Learning

Authors

  • Yogesh Thakur Department of Electrical and Computer Engineering, Bharati Vidyapeeth (Deemed to be University) College of Engineering, Pune, 411043, Maharashtra, India Author
  • Anudeep Dash Department of Electrical and Computer Engineering, Bharati Vidyapeeth (Deemed to be University) College of Engineering, Pune, 411043, Maharashtra, India Author
  • Ayush Kumar Patel Department of Electrical and Computer Engineering, Bharati Vidyapeeth (Deemed to be University) College of Engineering, Pune, 411043, Maharashtra, India Author
  • Prof. Ayesha Sayyad Department of Electrical and Computer Engineering, Bharati Vidyapeeth (Deemed to be University) College of Engineering, Pune, 411043, Maharashtra, India Author

DOI:

https://doi.org/10.32628/IJSRST25126244

Keywords:

Deep learning, Explainable AI (XAI), interpretability, model transparency, post-hoc explanation techniques

Abstract

The proliferation of deep learning models across critical domains has been met with a growing demand for transparency and accountability. The inherent complexity of these models, often characterized as "black boxes," poses significant challenges to their adoption in high-stakes environments such as healthcare and finance, where understanding the rationale behind a decision is as crucial as the decision itself. This review provides a comprehensive and systematic survey of the field of Explainable AI (XAI) as it pertains to deep learning. It begins by establishing the imperative for explainability, driven by technical, ethical, and regulatory pressures. A detailed taxonomy is proposed to structure the diverse landscape of XAI methods, categorizing them along the dimensions of intrinsic versus post-hoc, local versus global, and model-agnostic versus model-specific. The core of this review is a deep dive into the foundational post-hoc explanation techniques, including the local surrogate modelling of LIME, the game-theoretic framework of SHAP, and the gradient-based approaches of Grad-CAM and Integrated Gradients. Intrinsically interpretable methods, most notably the attention mechanism in Transformer models, are also analysed. Furthermore, the paper critically examines the methodologies for evaluating XAI techniques, contrasting functionally grounded metrics like fidelity and robustness with human-grounded, in-the-loop evaluations. Through case studies in healthcare and finance, the practical application and stakeholder-specific utility of these methods are illustrated. The review concludes by identifying key open challenges and charting future research directions, including the critical need for standardized evaluation benchmarks, the pursuit of causal explanations over correlational ones, and the emerging interplay between XAI and generative AI. This work aims to serve as a definitive reference for researchers and practitioners, providing a structured understanding of the principles, techniques, and future trajectory of explainability in deep learning.

Downloads

Download data is not yet available.

References

"Black Box Problem in AI" -GeeksforGeeks.

"What Is Black Box AI and How Does It Work?". By Matthew Kosinski. IBM.

"What is Explainable AI (XAI)?"- IBM.

"What is Explainable AI, or XAI? | Glossary" - HPE.

"What is Explainable AI? Benefits & Best Practices" - Qlik.

"Explainable Artificial Intelligence (XAI): General Taxonomy" – SogetiLabs (Capgemini).

"A Unified Approach to Interpreting Model Predictions" By Scott M. Lundberg and Su-In Lee. NIPS.

"Axiomatic Attribution for Deep Networks". By Mukund Sundararajan, Ankur Taly, and Qiqi Yan. (Semantic Scholar)

"Attention is All You Need". By Ashish Vaswani, et al. NIPS.

"'Why Should I Trust You?': Explaining the Predictions of Any Classifier" (The LIME Paper). By Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin - arXiv.

"Local Interpretable Model-Agnostic Explanations (LIME): An Introduction". By Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. - O'Reilly Media.

"An Introduction to SHAP Values and Machine Learning Interpretability". By Abid Ali Awan. - DataCamp.

"Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization". By Ramprasaath R. Selvaraju, et al. - arXiv.

"Should AI models be explainable to clinicians?". By Gwénolé Abgrall, Andre L. Holder, Zaineb Chelly Dagdia, Karine Zeitouni, and Xavier Monnet. - PubMed Central, National Library of Medicine.

"Interpretable AI for Bio-medical Applications". By Anoop Sathyan, Abraham Itzhak Weinberg, and Kelly Cohen - PubMed Central, National Library of Medicine.

"Survey of Explainable AI Techniques in Healthcare". By Dimitrios M. V. K. Moustakis - MDPI.

"The Usefulness of Gradient-Weighted CAM in Assisting Medical Diagnoses". By Piotr Bator, Maciej J. Ogorzałek, and Kamil Książek. MDPI.

"Explainable AI in Finance: Addressing the Needs of Diverse Stakeholders" By Cheryll-Ann Wilson - CFA Institute Research and Policy.

"Explainable AI for Regulatory Compliance in Financial and Healthcare Sectors: A comprehensive review" By Nikhil Gupta - International Journal of Advances in Engineering and Management (IJAEM).

"A Survey on Explainable AI: Techniques And Challenges" By Sai Teja Boppiniti - International Journal of Innovative Research in Engineering and Technology (IJIERT).

"Benchmarking eXplainable AI - A Survey on Available Toolkits and Open Challenges" By Jasper van der Donk, Daan de Klerk, and Cor J. Veenman - International Joint Conference on Artificial Intelligence (IJCAI).

"A Framework for Systematic Assessment and Reporting of Explainable AI Features". By Anis Kchaou, Daniel Omeiza, and Oana-Maria Camburu - arXiv.

"Human in the AI loop via xAI and Active Learning for Visual Inspection" By Jože Rožanec - University of Groningen.

"Explainable AI Evaluation: A Top-Down Approach for Selecting Optimal Explanations for Black Box Models By João Saleiro, Rui Abreu, and Miguel Couceiro - MDPI.

"F-Fidelity: A Robust Framework for Faithfulness Evaluation of Explainable AI". By Yuyang Liu, et al. - OpenReview.

"Evaluating Fidelity of Explainable Methods for Predictive Process Analytics". By Alireza Tavakoli, Michael T. Lash, and Marcos R. S. Borges - ResearchGate.

"Evaluation Metrics Research for Explainable Artificial Intelligence Global Methods Using Synthetic Data". By Jin-Woo Jeong, et al - MDPI.

Downloads

Published

05-10-2025

Issue

Section

Research Articles

How to Cite

[1]
Yogesh Thakur, Anudeep Dash, Ayush Kumar Patel, and Prof. Ayesha Sayyad, Trans., “A Comprehensive Review of Explainable AI (XAI) Methods in Deep Learning”, Int J Sci Res Sci & Technol, vol. 12, no. 5, pp. 333–352, Oct. 2025, doi: 10.32628/IJSRST25126244.