The Eighth International Workshop of Chinese CHI 2020
DOI: 10.1145/3403676.3403691
|View full text |Cite
|
Sign up to set email alerts
|

RECAST: Interactive Auditing of Automatic Toxicity Detection Models

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
10
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(10 citation statements)
references
References 2 publications
0
10
0
Order By: Relevance
“…Besides, they found no evidence that context actually improves the performance of toxicity classifiers. In another work [38] the authors presented an interactive tool for auditing toxicity detection models by visualizing explanations for predictions and providing alternative wordings for detected toxic speech. In particular, they displayed the attention of toxicity detection models on user input, providing suggestions on how to replace sensitive text with less toxic words.…”
Section: Related Workmentioning
confidence: 99%
“…Besides, they found no evidence that context actually improves the performance of toxicity classifiers. In another work [38] the authors presented an interactive tool for auditing toxicity detection models by visualizing explanations for predictions and providing alternative wordings for detected toxic speech. In particular, they displayed the attention of toxicity detection models on user input, providing suggestions on how to replace sensitive text with less toxic words.…”
Section: Related Workmentioning
confidence: 99%
“…These can be used to show how changing the input features affects the distribution of predictions and counterfactuals, especially for systems that support users in verifying and refining the user hypotheses. Conversely, when the focus is on individual instances, it is important to highlight both the differences between the input and the counterfactuals (Figure 10a), and the difference in terms of predictions [WSP*21, SGPR18], for example, by using tables [WPB*19, CMQ21], enhanced representation of the input [CBN*20] (e.g. heatmaps for images) or colour coding.…”
Section: Papers Categorizationmentioning
confidence: 99%
“…Combined with interactions, they allow users to investigate and analyse the models. Examples of added information are the attribution scores’ magnitude, which is encoded using colours, size [JCM20], opacity [WSP*21] or just its value, or bounding boxes [HSL*21, JKV*22, CBN*20], which highlight the most important region over images. While sorting [CHS20, MXC*20, KCK*19, Vig19, PCN*19] and filtering [WONM18, DWSZ20, JKV*22, JTH*21] by attribution scores capabilities are quite common to ease the data exploration and reduce the visual clutters, some works provide additional tools for a deeper understanding.…”
Section: Papers Categorizationmentioning
confidence: 99%
See 2 more Smart Citations