Home

Awesome

<div align="center"> <img src="docs/assets/banner2.png" width="75%" alt="Influenciae" align="center" /> </div> <br> <div align="center"> <a href="#"> <img src="https://img.shields.io/badge/Python-3.7, 3.8, 3.9, 3.10-efefef"> </a> <a href="#tf"> <img src="https://img.shields.io/badge/TensorFlow-2.7, 2.8, 2.9-00458A"> </a> <a href="https://github.com/deel-ai/influenciae/actions/workflows/linter.yml"> <img alt="PyLint" src="https://github.com/deel-ai/influenciae/actions/workflows/linter.yml/badge.svg"> </a> <a href="https://github.com/deel-ai/influenciae/actions/workflows/tests.yml"> <img alt="Tox" src="https://github.com/deel-ai/influenciae/actions/workflows/tests.yml/badge.svg"> </a> <a href="https://github.com/deel-ai/influenciae/actions/workflows/publish.yml"> <img alt="Pypi" src="https://github.com/deel-ai/influenciae/actions/workflows/publish.yml/badge.svg"> </a> <a href="#"> <img src="https://img.shields.io/badge/License-MIT-efefef"> </a> <br> <a href="https://deel-ai.github.io/influenciae/"><strong>Explore Influenciae docs Β»</strong></a> </div> <br>

Influenciae is a Python toolkit dedicated to computing influence values for the discovery of potentially problematic samples in a dataset and the generation of data-centric explanations for deep learning models. In this library based on Tensorflow, we gather state-of-the-art methods for estimating the importance of training samples and their influence on test data-points for validating the quality of datasets and of the models trained on them.

πŸ”₯ Tutorials

We propose some hands-on tutorials to get familiar with the library and it's API:

πŸš€ Quick Start

Influenciae requires a version of python 3.7 or higher and several libraries, including Tensorflow and Numpy. Installation can be done using Pypi:

pip install influenciae

Once Influenciae is installed, there are two major applications for the different modules (that all follow the same API). So, except for group-specific functions that are only available on the influence module, all the classes are able to compute self-influence values, the influence with one point wrt another, as well as find the top-k samples for both of these situations.

Discovering influential examples

Particularly useful when validating datasets, influence functions (and related notions) allow for gaining an insight into what samples the models thinks to be "important". For this, the training dataset and a trained model are needed.

from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import FirstOrderInfluenceCalculator
from deel.influenciae.utils import ORDER

# load the model, the training loss (without reduction) and the training data (with the labels and in a batched TF dataset)

influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = FirstOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator)
data_and_influence_dataset = influence_calculator.compute_influence_values(train_dataset)
# or influence_calculator.compute_top_k_from_training_dataset(train_dataset, k_samples, ORDER.DESCENDING) when the
# dataset is too large

This is also explained more in depth in the Getting Started tutotial <sub> Open In Colab </sub>

Explaining neural networks through their training data

Another application is to explain some model's predictions by looking on which training samples they are based on. Again, the training dataset, the model and the samples we wish to explain are needed.

from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import FirstOrderInfluenceCalculator
from deel.influenciae.utils import ORDER

# load the model, the training loss (without reduction), the training data and
# the data to explain (with the labels and in batched a TF dataset)

influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = FirstOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator)
data_and_influence_dataset = influence_calculator.estimate_influence_values_in_batches(samples_to_explain, train_dataset)
# or influence_calculator.top_k(samples_to_explain, train_dataset, k_samples, ORDER.DESCENDING) when the
# dataset is too large

This is also explained more in depth in the Getting Started tutorial <sub> Open In Colab </sub>

Determining the influence of groups of samples

The previous examples use notions of influence that are applied individually to each data-point, but it is possible to extend this to groups. That is, answer the question of what would a model look like if it hadn't seen a whole group of data-points during training, for example. This can be computed namely using the FirstOrderInfluenceCalculator and SecondOrderInfluenceCalculator, for implementations where pairwise interactions between each of the data-points are not taken into account and do, respectively.

For obtaining the groups' influence:

from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import SecondOrderInfluenceCalculator

# load the model, the training loss (without reduction), the training data and
# the data to explain (with the labels and in a batched TF dataset)

influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = SecondOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator)  # or FirstOrderInfluenceCalculator
data_and_influence_dataset = influence_calculator.estimate_influence_values_group(groups_train, groups_to_explain)

For the data-centric explanations:

from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import SecondOrderInfluenceCalculator

# load the model, the training loss (without reduction), the training data and
# the data to explain (with the labels and in a batched TF dataset)

influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = SecondOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator)  # or FirstOrderInfluenceCalculator
data_and_influence_dataset = influence_calculator.estimate_influence_values_group(groups_train)

πŸ“¦ What's Included

All the influence calculation methods work on Tensorflow models trained for any sort of task and on any type of data. Visualization functionality is implemented for image datasets only (for the moment).

Influence MethodSourceTutorial
Influence FunctionsPaperOpen In Colab
RelatIFPaperOpen In Colab
Influence Functions (first order, groups)PaperOpen In Colab
Influence Functions (second order, groups)PaperOpen In Colab
Arnoldi iteration (Scaling Up Influence Functions)PaperOpen In Colab
Trac-InPaperOpen In Colab
Representer Point Selection (L2)PaperOpen In Colab
Representer Point Selection (Local Jacobian Expansion)PaperOpen In Colab
Boundary-based influence--Open In Colab

πŸ‘€ See Also

This library proposes implementations of some of the different popular ways of calculating the influence of data-points on TF, but there are also other ones using other frameworks.

Some other tools for efficiently computing influence functions.

More from the DEEL project:

πŸ™ Acknowledgments

<img align="right" src="https://www.deel.ai/wp-content/uploads/2021/05/logo-DEEL.png" width="25%"> This project received funding from the French ”Investing for the Future – PIA3” program within the Artificial and Natural Intelligence Toulouse Institute (ANITI). The authors gratefully acknowledge the support of the <a href="https://www.deel.ai/"> DEEL </a> project.

πŸ‘¨β€πŸŽ“ Creators

This library was first created as a research tool by Agustin Martin PICARD in the context of the DEEL project with the help of David Vigouroux and Thomas FEL. Later on, Lucas Hervier joined the team to transform the code base as a practical user-(almost)-friendly and efficient tool.

πŸ—žοΈ Citation

If you use Influenciae as part of your workflow in a scientific publication, please consider citing the πŸ—žοΈ official paper:

@unpublished{picard:hal-04284178,
  TITLE = {Influenci{\ae}: A library for tracing the influence back to the data-points},
  AUTHOR = {Picard, Agustin Martin and Hervier, Lucas and Fel, Thomas and Vigouroux, David},
  URL = {https://hal.science/hal-04284178},
  NOTE = {working paper or preprint},
  YEAR = {2023},
  MONTH = Nov,
  KEYWORDS = {Data-centric ai ; XAI ; Explainability ; Influence Functions ; Open-source toolbox},
  PDF = {https://hal.science/hal-04284178/file/ms.pdf},
  HAL_ID = {hal-04284178},
  HAL_VERSION = {v1},
}

πŸ“ License

The package is released under <a href="https://choosealicense.com/licenses/mit"> MIT license</a>.