Quantitative evaluation of Saliency-Based Explainable artificial intelligence (XAI) methods in Deep Learning-Based mammogram analysis

Esma Cerekci, Deniz Alis*, Nurper Denizoglu, Ozden Camurdan, Mustafa Ege Seker, Caner Ozer, Muhammed Yusuf Hansu, Toygar Tanyel, Ilkay Oksuz, Ercan Karaarslan

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

12 Citations (Scopus)

Abstract

Background: Explainable Artificial Intelligence (XAI) is prominent in the diagnostics of opaque deep learning (DL) models, especially in medical imaging. Saliency methods are commonly used, yet there's a lack of quantitative evidence regarding their performance. Objectives: To quantitatively evaluate the performance of widely utilized saliency XAI methods in the task of breast cancer detection on mammograms. Methods: Three radiologists drew ground-truth boxes on a balanced mammogram dataset of women (n = 1496 cancer-positive and negative scans) from three centers. A modified, pre-trained DL model was employed for breast cancer detection, using MLO and CC images. Saliency XAI methods, including Gradient-weighted Class Activation Mapping (Grad-CAM), Grad-CAM++, and Eigen-CAM, were evaluated. We utilized the Pointing Game to assess these methods, determining if the maximum value of a saliency map aligned with the bounding boxes, representing the ratio of correctly identified lesions among all cancer patients, with a value ranging from 0 to 1. Results: The development sample included 2,244 women (75%), with the remaining 748 women (25%) in the testing set for unbiased XAI evaluation. The model's recall, precision, accuracy, and F1-Score in identifying cancer in the testing set were 69%, 88%, 80%, and 0.77, respectively. The Pointing Game Scores for Grad-CAM, Grad-CAM++, and Eigen-CAM were 0.41, 0.30, and 0.35 in women with cancer and marginally increased to 0.41, 0.31, and 0.36 when considering only true-positive samples. Conclusions: While saliency-based methods provide some degree of explainability, they frequently fall short in delineating how DL models arrive at decisions in a considerable number of instances.

Original languageEnglish
Article number111356
JournalEuropean Journal of Radiology
Volume173
DOIs
Publication statusPublished - Apr 2024

Bibliographical note

Publisher Copyright:
© 2024 Elsevier B.V.

Funding

This paper has been published benefiting from the 1001 Science and Technology National Grant Program of TUBITAK (Project no: 122E022). However, the entire responsibility of the paper belongs to the owner of the paper. The financial support received from TUBITAK does not mean that the content of the publciation is approved in a scientific sense by TUBITAK.

FundersFunder number
Türkiye Bilimsel ve Teknolojik Araştırma Kurumu122E022

    Keywords

    • Breast Cancer
    • Deep Learning
    • Mammogram
    • XAI

    Fingerprint

    Dive into the research topics of 'Quantitative evaluation of Saliency-Based Explainable artificial intelligence (XAI) methods in Deep Learning-Based mammogram analysis'. Together they form a unique fingerprint.

    Cite this