Proceedings of the First Workshop on Interactive Learning for Natural Language Processing 2021
DOI: 10.18653/v1/2021.internlp-1.1
|View full text |Cite
|
Sign up to set email alerts
|

HILDIF: Interactive Debugging of NLI Models Using Influence Functions

Abstract: Biases and artifacts in training data can cause unwelcome behavior in text classifiers (such as shallow pattern matching), leading to lack of generalizability. One solution to this problem is to include users in the loop and leverage their feedback to improve models. We propose a novel explanatory debugging pipeline called HILDIF, enabling humans to improve deep text classifiers using influence functions as an explanation method. We experiment on the Natural Language Inference (NLI) task, showing that HILDIF c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 14 publications
(20 citation statements)
references
References 15 publications
0
18
0
Order By: Relevance
“…Most papers in Table 1 focus on text classification with single input (TC) for a variety of specific problems such as email categorization , topic classification (Kulesza et al, 2015;Teso and Kersting, 2019), spam classification (Koh and Liang, 2017), sentiment analysis (Ribeiro et al, 2018b), and auto-coding of transcripts (Kulesza et al, 2010). By contrast, Zylberajch et al (2021) Ghai et al (2021) suggested that most researchers work on TC because, for this task, it is much easier for lay participants to understand explanations and give feedback (e.g., which keywords should be added or removed from the list of top features). 4 Meanwhile, some other NLP tasks require the feedback providers to have linguistic knowledge such as part-of-speech tagging, parsing, and machine translation.…”
Section: Tasksmentioning
confidence: 99%
See 4 more Smart Citations
“…Most papers in Table 1 focus on text classification with single input (TC) for a variety of specific problems such as email categorization , topic classification (Kulesza et al, 2015;Teso and Kersting, 2019), spam classification (Koh and Liang, 2017), sentiment analysis (Ribeiro et al, 2018b), and auto-coding of transcripts (Kulesza et al, 2010). By contrast, Zylberajch et al (2021) Ghai et al (2021) suggested that most researchers work on TC because, for this task, it is much easier for lay participants to understand explanations and give feedback (e.g., which keywords should be added or removed from the list of top features). 4 Meanwhile, some other NLP tasks require the feedback providers to have linguistic knowledge such as part-of-speech tagging, parsing, and machine translation.…”
Section: Tasksmentioning
confidence: 99%
“…Some papers even allowed humans to adjust the word importance scores (WS) (Kulesza et al, 2009(Kulesza et al, , 2015. This is analogous to specifying relevancy scores for example-based explanations (ES) in Zylberajch et al (2021). Meanwhile, feedback at the level of learned features (FE) (i.e., the internal neurons in the model) and learned rules (RU) rather than individual words, was asked in Lertvittayakumjorn et al (2020) andRibeiro et al (2018b), respectively.…”
Section: Collecting Feedbackmentioning
confidence: 99%
See 3 more Smart Citations