Explainable AI (XAI)#

Explainable AI (XAI) is a field of research that aims to make machine learning models more transparent and interpretable to humans. The goal is to help users understand how and why AI systems make decisions and provide insight into their inner workings. It allows us to detect, analyze, and prevent common mistakes, for example, when the model uses irrelevant features to make a prediction. XAI can help to build trust in AI, make sure that the model is safe for development and increase its adoption in various domains.

Most XAI methods generate saliency maps as a result. Saliency map is a visual representation, suitable for human comprehension, that highlights the most important parts of the image from the model point of view. It looks like a heatmap, where warm-colored areas represent the areas with main focus.

this image shows the result of XAI algorithm

These images are taken from D-RISE paper.#

We can generate saliency maps for a certain model that was trained in OpenVINO™ Training Extensions, using otx explain command line. Learn more about its usage in XAI Tutorial tutorial.

XAI algorithms for classification#

this image shows the comparison of XAI classification algorithms

For classification networks these algorithms are used to generate saliency maps:

  • Activation Map​ - this is the most basic and naive approach. It takes the outputs of the model’s feature extractor (backbone) and averages it in channel dimension. The results highly rely on the backbone and ignore neck and head computations. Basically, it gives a relatively good and fast result.

  • Recipro-CAM​ uses Class Activation Mapping (CAM) to weigh the activation map for each class, so it can generate different saliency per class. Recipro-CAM is a fast gradient-free Reciprocal CAM method. The method involves spatially masking the extracted feature maps to exploit the correlation between activation maps and network predictions for target classes. Default option for CNN models.

  • ViT Recipro-CAM​ is a modification of Recipro-CAM for Transformers-based models. It uses the same approach as Recipro-CAM but with a different way to extract feature maps from the Vision Transformer. Default option for transformer-based models.

Below we show the comparison of described algorithms. Access to the model internal state means the necessity to modify the model’s outputs and dump inner features. Per-class explanation support means generation different saliency maps for different classes.

Classification algorithm

Activation Map

Recipro-CAM

ViT Recipro-CAM

Need access to model internal state

Yes

Yes

Yes

Gradient-free

Yes

Yes

Yes

Single-shot

Yes

No (re-infer neck + head H*W times, where HxW – feature map size)

No (re-infer last transformer encoder block)

Per-class explanation support

No

Yes

Yes

Execution speed

Fast

Medium

Medium

XAI algorithms for detection#

For detection networks these algorithms are used to generate saliency maps:

  • Activation Map​ - the same approach as for classification networks, which uses the outputs from feature extractor. This algorithm is used to generate saliency maps for two-stage detectors.

  • DetClassProbabilityMap - this approach takes the raw classification head output and uses class probability maps to calculate regions of interest for each class. So, it creates different salience maps for each class. This algorithm is implemented for single-stage detectors only. Default option for detection models.

this image shows the detailed description of XAI detection algorithm

The main limitation of this method is that, due to the training loss design of most single-stage detectors, activation values drift towards the center of the object while propagating through the network. This prevents clear explanation in the input image space using intermediate activations.

Below is a comparison of the described algorithms:

  • “Access to the model internal state” refers to the necessity to modify the model’s outputs and dump inner features.

  • “Per-class explanation support” means the generation of different saliency maps for different classes.

  • “Per-box explanation support” means the generation of standalone saliency maps for each detected prediction.

Detection algorithm

Activation Map

DetClassProbabilityMap

Need access to model internal state

Yes

Yes

Gradient-free

Yes

Yes

Single-shot

Yes

Yes

Per-class explanation support

No

Yes

Per-box explanation support

No

No

Execution speed

Fast

Fast

XAI algorithms for instance segmentation#

For instance segmentation networks the following algorithm is used to generate saliency maps:

  • InstSegExplainAlgo​ - in this approach the predicted object masks are combined and aggregated per class to generate the saliency maps for each class.

engine.explain(
  checkpoint="<checkpoint-path>", # .pth or .xml weights of the model
  datamodule=OTXDataModule(), # The data module to use for predictions
  dump=True # Wherether to save saliency map images or not
  )
(otx) ...$ otx explain ... --checkpoint <checkpoint-path> # .pth or .xml weights of the model
                           --data_root <dataset_path> # Path to data folder or single image
                           --dump True # Wherether to save saliency map images or not