2021
DOI: 10.1145/3449163
|View full text |Cite
|
Sign up to set email alerts
|

AI-Assisted Human Labeling

Abstract: Human labeling of training data is often a time-consuming, expensive part of machine learning. In this paper, we study "batch labeling", an AI-assisted UX paradigm, that aids data labelers by allowing a single labeling action to apply to multiple records. We ran a large scale study on Mechanical Turk with 156 participants to investigate labeler-AI-batching system interaction. We investigate the efficacy of the system when compared to a single-item labeling interface (i.e., labeling one record at-a-time), and e… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 23 publications
(7 citation statements)
references
References 37 publications
0
6
0
Order By: Relevance
“…For instance, several domain experts described scenarios where a review may entail working with thousands of documents and/or tight timelines, making it impossible to manually verify all document details. The strategies we observed with Marco for cross-checking and verification could also become ineffective in these scenarios and potentially encourage overreliance [5]. Future work can investigate additional interactions that provide guardrails and mitigate risk in these high-stakes scenarios, for example by providing explorable uncertainty visualizations or further scaffolding results through clustering [5,18,26,46].…”
Section: Discussionmentioning
confidence: 98%
“…For instance, several domain experts described scenarios where a review may entail working with thousands of documents and/or tight timelines, making it impossible to manually verify all document details. The strategies we observed with Marco for cross-checking and verification could also become ineffective in these scenarios and potentially encourage overreliance [5]. Future work can investigate additional interactions that provide guardrails and mitigate risk in these high-stakes scenarios, for example by providing explorable uncertainty visualizations or further scaffolding results through clustering [5,18,26,46].…”
Section: Discussionmentioning
confidence: 98%
“…Past studies have developed intuitive interfaces that simplify the task of labeling, incorporating features like drag-and-drop [55] or batch labeling [5], highlighting [19], and leveraging languagebased models [7]. Stureborg et al grouped more similar contents together and different kinds of pass logic to coordinate between the crowdsourced annotators in multi-labeling tasks [56].…”
Section: Related Workmentioning
confidence: 99%
“…uncritically to avoid over-reliance (e.g. as observed in Moroz et al's study of Copilot [51], and discussed more generally in Ashktorab et al [9]) as well as automation bias [45,46,65]. We present the full text of the prompt used for the assistant in Appendix D.…”
Section: Supporting Conversational Interactionmentioning
confidence: 99%