Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 3: System Demonstrations) 2023
DOI: 10.18653/v1/2023.acl-demo.40
|View full text |Cite
|
Sign up to set email alerts
|

Inseq: An Interpretability Toolkit for Sequence Generation Models

Abstract: Past work in natural language processing interpretability focused mainly on popular classification tasks while largely overlooking generation settings, partly due to a lack of dedicated tools. In this work, we introduce Inseq 1 , a Python library to democratize access to interpretability analyses of sequence generation models. Inseq enables intuitive and optimized extraction of models' internal information and feature importance scores for popular decoderonly and encoder-decoder Transformers architectures. We … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 13 publications
(10 citation statements)
references
References 28 publications
0
1
0
Order By: Relevance
“…However, the Translation stage is still opaque, meaning it is not self-interpretable how the LM generates the reasoning chain from the question. It is still an under-explored question whether it is possible to improve the interpretability of the LM generation process in general, and a few recent studies have made promising early progress (Yin and Neubig, 2022;Sarti et al, 2023) that might be used to improve the faithfulness of the Translation stage.…”
Section: Limitationsmentioning
confidence: 99%
“…However, the Translation stage is still opaque, meaning it is not self-interpretable how the LM generates the reasoning chain from the question. It is still an under-explored question whether it is possible to improve the interpretability of the LM generation process in general, and a few recent studies have made promising early progress (Yin and Neubig, 2022;Sarti et al, 2023) that might be used to improve the faithfulness of the Translation stage.…”
Section: Limitationsmentioning
confidence: 99%
“…More recent solutions enable the study of the impact of source and target tokens (Ferrando et al, 2022) or discover the causes of hallucinations (Dale et al, 2023). Concurrent work by Sarti et al (2023b) uses post-hoc XAI methods to uncover gender bias in Turkish-English neural MT models. We expand their setup to more complex sentences and the notional-togrammatical gender MT in two more languages.…”
Section: Related Workmentioning
confidence: 99%
“…We aggregate first over f and then g because we expect token-level per-unit scores to represent token attribution more expressively, and we do not want to lose such information with an initial pooling along the hidden size. We use Inseq (Sarti et al, 2023b) to compute and aggregate the scores.…”
Section: B1 Interpretabilitymentioning
confidence: 99%
“…These logits vary vastly among different languages. To focus on the relation between the original and edited fact, we normalize the logits following the previous work (Sarti et al, 2023) as…”
Section: Subword Vocabulary Overlapmentioning
confidence: 99%