2018
DOI: 10.1109/tkde.2017.2750683
|View full text |Cite
|
Sign up to set email alerts
|

Computing Crowd Consensus with Partial Agreement

Abstract: Abstract-Crowdsourcing has been widely established as a means to enable human computation at large-scale, in particular for tasks that require manual labelling of large sets of data items. Answers obtained from heterogeneous crowd workers are aggregated to obtain a robust result. However, existing methods for answer aggregation are designed for discrete tasks, where answers are given as a single label per item. In this paper, we consider partial-agreement tasks that are common in many applications such as imag… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
10
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
6
1
1

Relationship

3
5

Authors

Journals

citations
Cited by 31 publications
(10 citation statements)
references
References 52 publications
(77 reference statements)
0
10
0
Order By: Relevance
“…This assumption matches with real-life settings, e.g. training data, prior knowledge, crowdsourcing, human expert [11,20,21,22,23]. Note that the output size is also known as recall since the size of ground truth is fixed.…”
mentioning
confidence: 71%
“…This assumption matches with real-life settings, e.g. training data, prior knowledge, crowdsourcing, human expert [11,20,21,22,23]. Note that the output size is also known as recall since the size of ground truth is fixed.…”
mentioning
confidence: 71%
“…Also, the baseline performs worse for small compression ratios (e.g. 10%), highlighting the practicality of our approach: Acknowledging cognitive load limits of users (b ≤ 20 according to [20,21,42,43,[47][48][49]), our approach helps to identify important data regularities, outperforming a (naive) uniform partitioning. Even with a compression ratio of 90%, the distortion of uniform partitioning is two times higher than our approach.…”
Section: Evaluating the Partition Qualitymentioning
confidence: 96%
“…The sample-label information collected via crowdsourcing is generally erroneous, due to the fact that online workers may lack expertise and proper incentives [30], [32]. This heterogeneous nature leads to the diverse submission quality of the completed tasks, pressing an urgent need for quality control [26], [27], [29], [33]- [39].…”
Section: Related Workmentioning
confidence: 99%
“…As a result, they may perform poorly when dealing with the more general multi-label data setting, where each object may have a set of non-exclusive labels, and labels may exhibit semantic correlations. Several multi-label crowd consensus algorithms have been recently proposed [26], [27], [29], [38], [39]. Nowak et al [38] studied the inter-annotator agreement for multilabel image annotation and focused on the annotation quality differences between expert and non-expert workers.…”
Section: Related Workmentioning
confidence: 99%