@inproceedings{f719d93eca374965b93569bc53a48a4f,
title = "Democratizing Advanced Attribution Analyses of Generative Language Models with the Inseq Toolkit",
abstract = "Inseq1 is a recent toolkit providing an intuitive and optimized interface to conduct feature attribution analyses of generative language models. In this work, we present the latest improvements to the library, including efforts to simplify the attribution of large language models on consumer hardware, additional attribution approaches, and a new client command to detect and attribute context usage in language model generations. We showcase an online demo using Inseq as an attribution backbone for context reliance analysis, and we highlight interesting contextual patterns in language model generations. Ultimately, this release furthers Inseq{\textquoteright}s mission of centralizing good interpretability practices and enabling fair and reproducible model evaluations.",
keywords = "Feature Attribution, Generative Language Models, Natural Language Processing, Python Toolkit",
author = "Gabriele Sarti and Nils Feldhus and Jirui Qi and Malvina Nissim and Arianna Bisazza",
note = "Publisher Copyright: {\textcopyright} 2024 Copyright for this paper by its authors.; Joint of the 2nd World Conference on eXplainable Artificial Intelligence Late-Breaking Work, Demos and Doctoral Consortium, xAI-2024:LB/D/DC ; Conference date: 17-07-2024 Through 19-07-2024",
year = "2024",
language = "English",
series = "CEUR Workshop Proceedings",
publisher = "CEUR Workshop Proceedings (CEUR-WS.org)",
pages = "289--296",
editor = "Luca Longo and Weiru Liu and Gr{\'e}goire Montavon",
booktitle = "xAI-2024 Late-breaking Work, Demos and Doctoral Consortium Joint Proceedings",
}