Awesome
<div align="center"> <img src="docs/assets/banner2.png" width="75%" alt="Influenciae" align="center" /> </div> <br> <div align="center"> <a href="#"> <img src="https://img.shields.io/badge/Python-3.7, 3.8, 3.9, 3.10-efefef"> </a> <a href="#tf"> <img src="https://img.shields.io/badge/TensorFlow-2.7, 2.8, 2.9-00458A"> </a> <a href="https://github.com/deel-ai/influenciae/actions/workflows/linter.yml"> <img alt="PyLint" src="https://github.com/deel-ai/influenciae/actions/workflows/linter.yml/badge.svg"> </a> <a href="https://github.com/deel-ai/influenciae/actions/workflows/tests.yml"> <img alt="Tox" src="https://github.com/deel-ai/influenciae/actions/workflows/tests.yml/badge.svg"> </a> <a href="https://github.com/deel-ai/influenciae/actions/workflows/publish.yml"> <img alt="Pypi" src="https://github.com/deel-ai/influenciae/actions/workflows/publish.yml/badge.svg"> </a> <a href="#"> <img src="https://img.shields.io/badge/License-MIT-efefef"> </a> <br> <a href="https://deel-ai.github.io/influenciae/"><strong>Explore Influenciae docs Β»</strong></a> </div> <br>Influenciae is a Python toolkit dedicated to computing influence values for the discovery of potentially problematic samples in a dataset and the generation of data-centric explanations for deep learning models. In this library based on Tensorflow, we gather state-of-the-art methods for estimating the importance of training samples and their influence on test data-points for validating the quality of datasets and of the models trained on them.
π₯ Tutorials
We propose some hands-on tutorials to get familiar with the library and it's API:
- Getting Started <sub> </sub>
- Benchmarking with Mislabeled sample detection <sub> </sub>
- Using the first order influence calculator <sub> </sub>
- Using the second order influence calculator <sub> </sub>
- Using Arnoldi Influence Calculator <sub> </sub>
- Using TracIn <sub> </sub>
- Using Representer Point Selection - L2 (RPS_L2) <sub> </sub>
- Using Representer Point Selection - Local Jacobian Expansion (RPS_LJE) <sub> </sub>
- Using Boundary-based Influence <sub> </sub>
π Quick Start
Influenciae requires a version of python 3.7 or higher and several libraries, including Tensorflow and Numpy. Installation can be done using Pypi:
pip install influenciae
Once Influenciae is installed, there are two major applications for the different modules (that all follow the same API).
So, except for group-specific functions that are only available on the influence
module, all the classes are able to compute self-influence values, the influence with one point wrt another, as well as find the top-k samples for both of these situations.
Discovering influential examples
Particularly useful when validating datasets, influence functions (and related notions) allow for gaining an insight into what samples the models thinks to be "important". For this, the training dataset and a trained model are needed.
from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import FirstOrderInfluenceCalculator
from deel.influenciae.utils import ORDER
# load the model, the training loss (without reduction) and the training data (with the labels and in a batched TF dataset)
influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = FirstOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator)
data_and_influence_dataset = influence_calculator.compute_influence_values(train_dataset)
# or influence_calculator.compute_top_k_from_training_dataset(train_dataset, k_samples, ORDER.DESCENDING) when the
# dataset is too large
This is also explained more in depth in the Getting Started tutotial <sub> </sub>
Explaining neural networks through their training data
Another application is to explain some model's predictions by looking on which training samples they are based on. Again, the training dataset, the model and the samples we wish to explain are needed.
from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import FirstOrderInfluenceCalculator
from deel.influenciae.utils import ORDER
# load the model, the training loss (without reduction), the training data and
# the data to explain (with the labels and in batched a TF dataset)
influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = FirstOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator)
data_and_influence_dataset = influence_calculator.estimate_influence_values_in_batches(samples_to_explain, train_dataset)
# or influence_calculator.top_k(samples_to_explain, train_dataset, k_samples, ORDER.DESCENDING) when the
# dataset is too large
This is also explained more in depth in the Getting Started tutorial <sub> </sub>
Determining the influence of groups of samples
The previous examples use notions of influence that are applied individually to each data-point, but it is possible to extend this to groups. That is, answer the question of what would a model look like if it hadn't seen a whole group of data-points during training, for example. This can be computed namely using the FirstOrderInfluenceCalculator
and SecondOrderInfluenceCalculator
, for implementations where pairwise interactions between each of the data-points are not taken into account and do, respectively.
For obtaining the groups' influence:
from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import SecondOrderInfluenceCalculator
# load the model, the training loss (without reduction), the training data and
# the data to explain (with the labels and in a batched TF dataset)
influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = SecondOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator) # or FirstOrderInfluenceCalculator
data_and_influence_dataset = influence_calculator.estimate_influence_values_group(groups_train, groups_to_explain)
For the data-centric explanations:
from deel.influenciae.common import InfluenceModel, ExactIHVP
from deel.influenciae.influence import SecondOrderInfluenceCalculator
# load the model, the training loss (without reduction), the training data and
# the data to explain (with the labels and in a batched TF dataset)
influence_model = InfluenceModel(model, start_layer=target_layer, loss_function=loss_function)
ihvp_calculator = ExactIHVP(influence_model, train_dataset)
influence_calculator = SecondOrderInfluenceCalculator(influence_model, train_dataset, ihvp_calculator) # or FirstOrderInfluenceCalculator
data_and_influence_dataset = influence_calculator.estimate_influence_values_group(groups_train)
π¦ What's Included
All the influence calculation methods work on Tensorflow models trained for any sort of task and on any type of data. Visualization functionality is implemented for image datasets only (for the moment).
Influence Method | Source | Tutorial |
---|---|---|
Influence Functions | Paper | |
RelatIF | Paper | |
Influence Functions (first order, groups) | Paper | |
Influence Functions (second order, groups) | Paper | |
Arnoldi iteration (Scaling Up Influence Functions) | Paper | |
Trac-In | Paper | |
Representer Point Selection (L2) | Paper | |
Representer Point Selection (Local Jacobian Expansion) | Paper | |
Boundary-based influence | -- |
π See Also
This library proposes implementations of some of the different popular ways of calculating the influence of data-points on TF, but there are also other ones using other frameworks.
Some other tools for efficiently computing influence functions.
- Scaling Up Influence Functions a Python library using JAX implementing a scalable algorithm for computing influence functions.
- FastIF: Scalable Influence Functions for Efficient Model Interpretation and Debugging a Python library using PyTorch implementing another scalable algorithm for computing influence functions.
More from the DEEL project:
- Xplique a Python library exclusively dedicated to explaining neural networks.
- deel-lip a Python library for training k-Lipschitz neural networks on TF.
- deel-torchlip a Python library for training k-Lipschitz neural networks on PyTorch.
- DEEL White paper a summary of the DEEL team on the challenges of certifiable AI and the role of data quality, representativity and explainability for this purpose.
π Acknowledgments
<img align="right" src="https://www.deel.ai/wp-content/uploads/2021/05/logo-DEEL.png" width="25%"> This project received funding from the French βInvesting for the Future β PIA3β program within the Artificial and Natural Intelligence Toulouse Institute (ANITI). The authors gratefully acknowledge the support of the <a href="https://www.deel.ai/"> DEEL </a> project.π¨βπ Creators
This library was first created as a research tool by Agustin Martin PICARD in the context of the DEEL project with the help of David Vigouroux and Thomas FEL. Later on, Lucas Hervier joined the team to transform the code base as a practical user-(almost)-friendly and efficient tool.
ποΈ Citation
If you use Influenciae as part of your workflow in a scientific publication, please consider citing the ποΈ official paper:
@unpublished{picard:hal-04284178,
TITLE = {Influenci{\ae}: A library for tracing the influence back to the data-points},
AUTHOR = {Picard, Agustin Martin and Hervier, Lucas and Fel, Thomas and Vigouroux, David},
URL = {https://hal.science/hal-04284178},
NOTE = {working paper or preprint},
YEAR = {2023},
MONTH = Nov,
KEYWORDS = {Data-centric ai ; XAI ; Explainability ; Influence Functions ; Open-source toolbox},
PDF = {https://hal.science/hal-04284178/file/ms.pdf},
HAL_ID = {hal-04284178},
HAL_VERSION = {v1},
}
π License
The package is released under <a href="https://choosealicense.com/licenses/mit"> MIT license</a>.