1. Introduction
Machine learning models have enormous potential in solving sophisticated problems in variety of domains be it object detection [9], autonomous driving [2], DNA sequence generation [17], speech recognition [24] and language processing [6]. These models are capable of learning complex representations. However, human interpretability of these models has been very challenging due to the Black Box nature of it, thus making these models untrustworthy to be used in critical scenarios like healthcare applications as in disease discovery and diagnosis [11], drug discovery [10], autonomous driving [2] etc. Trusting these models in critical application requires us to be cognizant about the pertinent features and their effectiveness, that the model has learnt. These models would have been validated only on the perceived scenario and lacks to accommodate the unseen scenarios - drifting distribution. Hence there is a high need to strengthen the modelling methods for meaningful predictions from such models. The emerging field in machine learning, Explainable AI [7], aims to address this problem of discovering the Black Box decisions in deep neural networks.