I. Introduction
Fundus images are indispensable in diagnosing retinal diseases, such as diabetic retinopathy (DR), age-related macular degeneration (AMD), and pathological myopia (PM). With the advancement of deep learning, deep neural networks (DNNs) have been widely applied to assist ophthalmologists in diagnosing these retinal diseases automatically by learning the discriminative features from fundus images through large-scale datasets and sophisticated architectures, such as Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs). These models have demonstrated remarkable accuracy and performance in various retinal disease diagnosis tasks [1], [2], [3], [4], [5], [6], [7], [8]. However, these DNN-based approaches are rarely applied in clinical practice, mainly due to the black-box nature of deep learning. Given that medical diagnosis demands a profound understanding of the domain knowledge, having decisions that are not comprehensible to human experts is unacceptable. This lack of transparency would undermine the trust of doctors [9]. To address this issue, various strategies [10], [11] have been proposed to help people understand the decision-making process of deep neural networks. Currently, the explanation of medical image analysis models mainly relies on saliency maps, which highlight the regions of an image that the model deems important for making predictions [12], [13]. Despite being widely used, some studies [14], [15], [16] have revealed that the saliency maps generated by these methods could be inconsistent for misclassified samples.