2020
DOI: 10.1609/icwsm.v14i1.7282
|View full text |Cite
|
Sign up to set email alerts
|

Feature-Based Explanations Don't Help People Detect Misclassifications of Online Toxicity

Abstract: We present an experimental assessment of the impact of feature attribution-style explanations on human performance in predicting the consensus toxicity of social media posts with advice from an unreliable machine learning model. By doing so we add to a small but growing body of literature inspecting the utility of interpretable machine learning in terms of human outcomes. We also evaluate interpretable machine learning for the first time in the important domain of online toxicity, where fully-automated methods… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 31 publications
(9 citation statements)
references
References 32 publications
1
4
0
Order By: Relevance
“…Withholding predictions of an uncalibrated model may improve decision quality. Consistent with prior work on AI-advised decision-making (e.g., [7,10,21,38]), our results suggest that when a model is well-calibrated and more accurate than humans alone, users with access to its predictions can perform better than without the model, but not as well as the model alone for easier instances. When the model is poorly calibrated, the type of prediction display affects whether people can perform better by accessing the model predictions.…”
Section: Discussionsupporting
confidence: 86%
“…Withholding predictions of an uncalibrated model may improve decision quality. Consistent with prior work on AI-advised decision-making (e.g., [7,10,21,38]), our results suggest that when a model is well-calibrated and more accurate than humans alone, users with access to its predictions can perform better than without the model, but not as well as the model alone for easier instances. When the model is poorly calibrated, the type of prediction display affects whether people can perform better by accessing the model predictions.…”
Section: Discussionsupporting
confidence: 86%
“…In recent years, there has been a surge of research in human-AI decision-making, with a growing number of studies conducting behavioral experiments to gain a better understanding of how humans form decisions in the presence of AI (Alufaisan et al n.d.;Buçinca et al 2020;Carton et al 2020;Lai et al 2020;Lai and Tan 2019;Liu et al 2021;Zhang et al 2020). This research has focused on improving human-AI decisionmaking to optimize team performance (Buçinca et al 2020;Zhang et al 2020).…”
Section: Human-ai Decision-making and Appropriate Reliancementioning
confidence: 99%
“…Thus, the AI model often provides the confidence level of the decision [51,75] or an additional explanation for its decision [1,42]. Several works have evaluated whether different types of explanations can support humans' understanding of the AI model so that they identify the right cases to rely on the recommendations [4,13,15,69]. Explanations can lead people to rely too much on the decision of the AI model, particularly when its suggestion is incorrect [6].…”
Section: Related Workmentioning
confidence: 99%