From Black Box to Glass Box: A Survey of Explainable AI in Mammographic Screening
Abstract
Breast cancer remains a leading global health concern, with early, accurate diagnosis through mammography being critical for effective treatment. The emergence of artificial intelligence (AI) has revolutionized breast cancer screening, yet the opacity of “black box” models in clinical applications has sparked pressing calls for greater transparency. Explainable AI (XAI) offers essential solutions by making model decisions interpretable, enabling clinicians to trust and adopt advanced algorithms more confidently. This review synthesizes the current landscape of XAI methods applied to mammographic imaging, examining cutting-edge techniques such as Grad-CAM, LIME, SHAP, attention mechanisms, and prototype-based models. We analyze how these approaches provide meaningful visual and textual explanations that bridge the gap between technical innovation and clinical utility. Unique to this survey is its focus on practical case studies, integration pathways, and challenges in real-world implementation, from balancing interpretability and diagnostic accuracy to the urgent need for robust, diverse datasets. As demand grows for ethical, transparent AI in medicine, our review highlights actionable strategies, future directions, and the collaborative role of radiologists, AI specialists, and patients. By connecting technical advances to clinical trust and patient-centered care, this work sets the foundation for safe, transparent breast cancer diagnosis and aims to inspire further progress throughout the field.
Keywords
References
- Y. Feng et al., “Breast cancer development and progression: Risk factors, cancer stem cells, signaling pathways, genomics, and molecular pathogenesis,” Genes Dis., vol. 5, no. 2, pp. 77–106, 2018. doi: 10.1016/j.gendis.2018.05.001
- Y. Shen et al., “Artificial intelligence system reduces false-positive findings in the interpretation of breast ultrasound exams,” Nat. Commun., vol. 12, no. 1, p. 5645, Dec. 2021. doi: 10.1038/s41467-021-26023-2
- H. Sung et al., “Global cancer statistics 2020: Globocan estimates of incidence and mortality worldwide for 36 cancers in 185 countries,” CA Cancer J. Clin., vol. 71, no. 3, pp. 209–249, May. 2021. doi: 10.3322/caac.21660
- A. Gerbasi et al., “DeepMiCa: Automatic segmentation and classification of breast MIcroCAlcifications from mammograms,” Comput. Methods Programs Biomed., vol. 235, Jun. 2023. doi: 10.1016/j.cmpb.2023.107483
- S. Bai, S. Nasir, R. A. Khan, S. Arif, A. Meyer, and H. Konik, “Breast cancer diagnosis: A comprehensive exploration of explainable artificial intelligence (XAI) techniques,” arXiv preprint, arXiv:2406.00532, Jun. 2024.
- E. Abu Abeelh, and Z. AbuAbeileh, “Impact of mammography screening frequency on breast cancer mortality rates,” Cureus, Nov. 2023. doi: 10.7759/cureus.49066
- N. Shifa, M. Saleh, Y. Akbari, and S. Al Maadeed, “A review of explainable AI techniques and their evaluation in mammography for breast cancer screening,” Clin. Imaging, p. 110492, Jul. 2025. doi: 10.1016/j.clinimag.2025.110492
- B. Abhisheka, S. K. Biswas, and B. Purkayastha, “A Comprehensive review on breast cancer detection, classification and segmentation using deep learning,” Arch. Comput. Methods Eng., Nov. 2023. doi:10.1007/s11831-023-09968-z
Details
Primary Language
English
Subjects
Artificial Intelligence (Other)
Journal Section
Review
Early Pub Date
March 30, 2026
Publication Date
March 30, 2026
Submission Date
August 16, 2025
Acceptance Date
October 3, 2025
Published in Issue
Year 2026 Volume: 9 Number: 1
