2021
DOI: 10.1162/tacl_a_00440
|View full text |Cite
|
Sign up to set email alerts
|

Explanation-Based Human Debugging of NLP Models: A Survey

Abstract: Debugging a machine learning model is hard since the bug usually involves the training data and the learning process. This becomes even harder for an opaque deep learning model if we have no clue about how the model actually works. In this survey, we review papers that exploit explanations to enable humans to give feedback and debug NLP models. We call this problem explanation-based human debugging (EBHD). In particular, we categorize and discuss existing work along three dimensions of EBHD (the bug context, t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 25 publications
(16 citation statements)
references
References 84 publications
0
13
0
Order By: Relevance
“…Another context where AXPLR could be useful is explanation-based human debugging of the model [49]. The individual model weight w i for the pattern feature p i may not make sense to humans when p i is in fact related to other pattern features (as we can see in Experiment 2, where τ(α i ) does not quite correlate with human reasoning).…”
Section: General Considerations On Axplrmentioning
confidence: 99%
“…Another context where AXPLR could be useful is explanation-based human debugging of the model [49]. The individual model weight w i for the pattern feature p i may not make sense to humans when p i is in fact related to other pattern features (as we can see in Experiment 2, where τ(α i ) does not quite correlate with human reasoning).…”
Section: General Considerations On Axplrmentioning
confidence: 99%
“…Besides, Lertvittayakumjorn and Toni (2021) recognize that there are explanations which stay between the local and the global scopes. These amount to explanations for groups of examples such as a group of false positives of a certain class and a cluster of examples with some fixed features.…”
Section: Scopes Of Explanationsmentioning
confidence: 99%
“…For instance, one may develop an annotation tool which shows explanations for MT metrics as supporting information and measure human annotators' efficiency, compared to the case where they use the system with no explanation. Also, developing a new framework for incorporating human feedback on different types of explanations to improve the metric is another way to evaluate the explanations with respect to a downstream task (i.e., metric improvement) (Lertvittayakumjorn and Toni 2021). Lastly, it is also possible to measure user trust in the metrics with and without the explanations so as to assess whether the explanations can boost the user trust and promote adoption of complex model-based metrics (Hoffman et al 2018;).…”
Section: Extrinsic Evaluation Of Explanations For Mt Metricsmentioning
confidence: 99%
See 1 more Smart Citation
“…Different from computer vision, the basic input units in NLP for neural models are discrete language tokens rather than continuous pixels in images [20,38,15]. This discrete nature of language poses a challenge for interpreting neural NLP models, making the interpreting methods in CV hard to be directly applied to NLP domain [22,114]. To accommodate the discrete nature of texts, a great variety of works have rapidly emerged over the past a few years for neural model interpretability.…”
Section: Introductionmentioning
confidence: 99%