Interpreto: An Explainability Library for Transformers
By: Antonin Poché , Thomas Mullor , Gabriele Sarti and more
Interpreto is a Python library for post-hoc explainability of text HuggingFace models, from early BERT variants to LLMs. It provides two complementary families of methods: attributions and concept-based explanations. The library connects recent research to practical tooling for data scientists, aiming to make explanations accessible to end users. It includes documentation, examples, and tutorials. Interpreto supports both classification and generation models through a unified API. A key differentiator is its concept-based functionality, which goes beyond feature-level attributions and is uncommon in existing libraries. The library is open source; install via pip install interpreto. Code and documentation are available at https://github.com/FOR-sight-ai/interpreto.
Similar Papers
From Black Box to Transparency: Enhancing Automated Interpreting Assessment with Explainable AI in College Classrooms
Computation and Language
Helps computers judge translation quality better.
nnterp: A Standardized Interface for Mechanistic Interpretability of Transformers
Machine Learning (CS)
Lets scientists understand how AI brains work.
Towards the Formalization of a Trustworthy AI for Mining Interpretable Models explOiting Sophisticated Algorithms
Artificial Intelligence
Makes AI fair, private, and understandable.