» Articles » PMID: 39507460

Heatmap Analysis for Artificial Intelligence Explainability in Diabetic Retinopathy Detection: Illuminating the Rationale of Deep Learning Decisions

Overview
Journal Ann Transl Med
Date 2024 Nov 7
PMID 39507460
Authors
Affiliations
Soon will be listed here.
Abstract

Background: The opaqueness of artificial intelligence (AI) algorithms decision processes limit their application in healthcare. Our objective was to explore discrepancies in heatmaps originated from slightly different retinal images from the same eyes of individuals with diabetes, to gain insights into the deep learning (DL) decision process.

Methods: Pairs of retinal images from the same eyes of individuals with diabetes, composed of images obtained before and after pupil dilation, underwent automatic analysis by a convolutional neural network for the presence of diabetic retinopathy (DR), output being a score ranging from 0 to 1. Gradient-based Class Activation Maps (GradCam) allowed visualization of activated areas. Pairs of images with discordant DL scores or outputs within the pair were objectively compared to the concordant pairs, regarding the sum of activations of Class Activation Mapping (CAM), the number of activated areas, and DL score differences. Heatmaps of discordant pairs were also qualitatively assessed.

Results: Algorithmic performance for the detection of DR attained 89.8% sensitivity, 96.3% specificity and area under the receiver operating characteristic (ROC) curve of 0.95. Out of 210 comparable pairs of images, 20 eyes and 10 eyes were considered discordant according to DL score difference and regarding DL output, respectively. Comparison of concordant versus discordant groups showed statistically significant differences for all objective variables. Qualitative analysis pointed to subtle differences in image quality within discordant pairs.

Conclusions: The successfully established relationship among objective parameters extracted from heatmaps and DL output discrepancies reinforces the role of heatmaps for DL explainability, fostering acceptance of DL systems for clinical use.

References
1.
Chang J, Lee J, Ha A, Han Y, Bak E, Choi S . Explaining the Rationale of Deep Learning Glaucoma Decisions with Adversarial Examples. Ophthalmology. 2020; 128(1):78-88. DOI: 10.1016/j.ophtha.2020.06.036. View

2.
Korot E, Pontikos N, Liu X, Wagner S, Faes L, Huemer J . Predicting sex from retinal fundus photographs using automated deep learning. Sci Rep. 2021; 11(1):10286. PMC: 8119673. DOI: 10.1038/s41598-021-89743-x. View

3.
Yousefi S, Pasquale L, Boland M . Artificial Intelligence and Glaucoma: Illuminating the Black Box. Ophthalmol Glaucoma. 2020; 3(5):311-313. DOI: 10.1016/j.ogla.2020.04.008. View

4.
Amann J, Blasimme A, Vayena E, Frey D, Madai V . Explainability for artificial intelligence in healthcare: a multidisciplinary perspective. BMC Med Inform Decis Mak. 2020; 20(1):310. PMC: 7706019. DOI: 10.1186/s12911-020-01332-6. View

5.
Quellec G, Al Hajj H, Lamard M, Conze P, Massin P, Cochener B . ExplAIn: Explanatory artificial intelligence for diabetic retinopathy diagnosis. Med Image Anal. 2021; 72:102118. DOI: 10.1016/j.media.2021.102118. View