Description

Saliency maps are visual explanations for image classification models that highlight which pixels in an image most strongly influence the model's prediction. Computed by calculating gradients of the model's output with respect to input pixels, saliency maps produce heatmaps where brighter regions indicate pixels that, when changed, would most significantly affect the prediction. This technique helps users understand which parts of an image the model is 'looking at' when making decisions.

Example Use Cases

Explainability

Analysing X-ray images in a pneumonia detection model to verify that the algorithm focuses on lung regions showing inflammatory patterns rather than irrelevant areas like medical equipment or patient positioning markers.

Examining skin lesion classification models to ensure the algorithm identifies diagnostic features (irregular borders, colour variation) rather than artifacts like rulers, hair, or skin markings that shouldn't influence medical decisions.

Fairness

Auditing a dermatology AI system to verify it focuses on medical symptoms rather than skin colour when diagnosing conditions, ensuring equitable treatment across racial groups by revealing inappropriate attention to demographic features.

Limitations

  • Saliency maps are often noisy and can change dramatically with small input perturbations, making them unstable.
  • Highlighted regions may not correspond to semantically meaningful or human-understandable features.
  • Only indicates local gradient information, not causal importance or actual decision-making logic.
  • May highlight irrelevant pixels that happen to have high gradients due to model artifacts rather than meaningful patterns.

Resources

Research Papers

Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps
Karen Simonyan, Andrea Vedaldi, and Andrew ZissermanDec 20, 2013

This paper addresses the visualisation of image classification models, learnt using deep Convolutional Networks (ConvNets). We consider two visualisation techniques, based on computing the gradient of the class score with respect to the input image. The first one generates an image, which maximises the class score [Erhan et al., 2009], thus visualising the notion of the class, captured by a ConvNet. The second technique computes a class saliency map, specific to a given image and class. We show that such maps can be employed for weakly supervised object segmentation using classification ConvNets. Finally, we establish the connection between the gradient-based ConvNet visualisation methods and deconvolutional networks [Zeiler et al., 2013].

Software Packages

pytorch-cnn-visualizations
Oct 21, 2017

Pytorch implementation of convolutional neural network visualization techniques

Documentations

Concepts of Saliency and Explainability in AI
Xaitk-saliency Developers
Occlusion Saliency Example
Xaitk-saliency Developers

Tags

Explainability Dimensions

Attribution Methods:
Visualization Methods:
Explanation Target:
Explanatory Scope:

Other Categories

Data Requirements:
Data Type:
Evidence Type:
Expertise Needed:
Lifecycle Stage:
Technique Type: