A comparative study of explainability methods for whole slide classification of lymph node metastases using vision transformers
Jens Rahnfeld,
Mehdi Naouar,
Gabriel Kalweit,
Joschka Boedecker,
Estelle Dubruc and
Maria Kalweit
PLOS Digital Health, 2025, vol. 4, issue 4, 1-21
Abstract:
Recent advancements in deep learning have shown promise in enhancing the performance of medical image analysis. In pathology, automated whole slide imaging has transformed clinical workflows by streamlining routine tasks and diagnostic and prognostic support. However, the lack of transparency of deep learning models, often described as black boxes, poses a significant barrier to their clinical adoption. This study evaluates various explainability methods for Vision Transformers, assessing their effectiveness in explaining the rationale behind their classification predictions on histopathological images. Using a Vision Transformer trained on the publicly available CAMELYON16 dataset comprising of 399 whole slide images of lymph node metastases of patients with breast cancer, we conducted a comparative analysis of a diverse range of state-of-the-art techniques for generating explanations through heatmaps, including Attention Rollout, Integrated Gradients, RISE, and ViT-Shapley. Our findings reveal that Attention Rollout and Integrated Gradients are prone to artifacts, while RISE and particularly ViT-Shapley generate more reliable and interpretable heatmaps. ViT-Shapley also demonstrated faster runtime and superior performance in insertion and deletion metrics. These results suggest that integrating ViT-Shapley-based heatmaps into pathology reports could enhance trust and scalability in clinical workflows, facilitating the adoption of explainable artificial intelligence in pathology.Author summary: The objective of our research was to investigate methods for enhancing the explainability of state-of-the-art Vision Transformer models for medical image analysis in histopathology. These models are often perceived as opaque, lacking transparency, which can limit their deployment in clinical settings. In our study, we evaluated various approaches for generating visual explanations that highlight critical areas in histopathology images that influence the Vision Transformers’ decisions and serve as heatmaps for pathologists. By comparing different techniques, we identified that the ViT-Shapley method generated the most reliable and effective heatmaps. In evaluating the practical impact, we considered both clinical usability and computational efficiency. These heatmaps can help clinicians understand and trust the model’s decisions, potentially integrating these advanced AI tools into routine clinical workflows to improve diagnostic support. The findings of our study indicate that the integration of Transformer-based models with effective explainability methods can enhance the accuracy and transparency of automated whole slide image analysis, ultimately benefiting patient care.
Date: 2025
References: View references in EconPapers View complete reference list from CitEc
Citations:
Downloads: (external link)
https://journals.plos.org/digitalhealth/article?id=10.1371/journal.pdig.0000792 (text/html)
https://journals.plos.org/digitalhealth/article/fi ... 00792&type=printable (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:plo:pdig00:0000792
DOI: 10.1371/journal.pdig.0000792
Access Statistics for this article
More articles in PLOS Digital Health from Public Library of Science
Bibliographic data for series maintained by digitalhealth ().