Explainable Artificial Intelligence in Medical Image based Diagnosis: A Comprehensive Study

Authors

  • Adeela Hayat Department of Computer Sciences, Faculty of Computer Sciences, Lahore Garrison University, Lahore, 54000, Pakistan Author
  • Zainab Azhar Department of Software Engineering, Faculty of Computer Sciences, Lahore Garrison University, Lahore, 54000, Pakistan Author
  • Amna Kosar Department of Computer Sciences, Faculty of Computer Sciences, Lahore Garrison University, Lahore, 54000, Pakistan Author
  • Hafiz Burhan Ul Haq Department of Information Technology, Faculty of Computer Sciences, Lahore Garrison University, Lahore, 54000, Pakistan Author
  • Sabir Abbas Department of Information Technology, Faculty of Computer Sciences, Lahore Garrison University, Lahore, 54000, Pakistan Author
  • Rabia Younas Department of Information Technology, Faculty of Computer Sciences, Lahore Garrison University, Lahore, 54000, Pakistan Author

DOI:

https://doi.org/10.57041/dygg4143

Keywords:

Explainable AI, Deep Learning, Medical Imaging, Grad-CAM, SHAP, LIME, Clinical Diagnosis

Abstract

Artificial intelligence (AI) has made remarkable contributions to medical imaging, enhancing the accuracy of the diagnosis and its efficiency; the challenge to the adoption in clinical settings has been the lack of clarity of the deep learning models. Explainable Artificial Intelligence (XAI) is a solution to this dilemma, as it provides easy-to-understand and understandable information about the way the model arrived at its decisions. The review summarizes the latest advances in XAI in the radiology, ophthalmology, pathology, neurology, and oncology domains as well as the conventional machine learning attribution algorithms; saliency-based deep learning algorithms, including Grad-CAM, SHAP, and LIME; and multimodal models. Its main clinical uses are cancer detection, classification of Alzheimer's disease, COVID-19 diagnosis, and evaluation of retinal disease. There are both quantitative (fidelity, stability, and localization) and qualitative (interpretability and radiologist trust) evaluation strategies. Although there is significant improvement, there exist dataset quality issues, interpretability-accuracy trade-offs, and generalizability issues across clinical settings. Furthermore, a comparative study with the available literature is done in a systematically organized way the evaluation parameters are clearly developed to evaluate objectively the strengths of our approach. This comparative parameter-based parameter shows the advances of our framework with respect to generalizability, explainability quality, and clinical relevance. Our study makes a more effective justification of its usefulness than the previous methods since the comparison is based on quantifiable measures.

Downloads

Published

2025-12-30

How to Cite

Explainable Artificial Intelligence in Medical Image based Diagnosis: A Comprehensive Study. (2025). International Journal of Emerging Engineering and Technology, 4(2), 28-35. https://doi.org/10.57041/dygg4143

Most read articles by the same author(s)

Similar Articles

1-10 of 16

You may also start an advanced similarity search for this article.