Home

Awesome

This Looks Like That, Because ... Explaining Prototypes for Interpretable Image Recognition

This repository contains Python code and Jupyter notebooks to explain prototypes learned by the Prototypical Part Network (ProtoPNet). Only visualizing prototypes can be insufficient for understanding what a prototype exactly represents, and why a prototype and an image are considered similar. We improve interpretability by automatically quantifying the influence of color hue, shape, texture, contrast and saturation in a prototype.

This is a fork of ProtoPNet with a new folder preprocessing and an extra file prototype_activation.py.

Corresponding paper on ArXiv: "This Looks Like That, Because ... Explaining Prototypes for Interpretable Image Recognition"

Overview of explaining prototypical learning for image recognition. Overview of explaining prototypical learning for image recognition. 1) Prototype-based image classification with ProtoPNet 2) Our contribution: Quantifying the importance of visual characteristics to explain why the classification model deemed an image patch and prototype similar. Left: Logical explanation for the clear similarity between the patches. Right: a 'misleading' prototype: humans might expect these patches to be dissimilar, but our method explains that the classification model considers these patches alike because of similar texture.

Usage

  1. Download and unpack the dataset CUB_200_2011.tgz
  2. Run preprocessing/crop_images.py to crop images according to the provided bounding box annotations
  3. Create images with modified hue, shape, texture, contrast and saturation by running preprocessing/create_modified_datasets.py.
  4. Train a ProtoPNet as presented by the original authors: https://github.com/cfchen-duke/ProtoPNet with main.py and the appropriate parameters in settings.py.
  5. Run prototype_activation.py to get global and local importance scores which explain prototypes.