2012
DOI: 10.1007/s10791-011-9181-9
|View full text |Cite
|
Sign up to set email alerts
|

Increasing cheat robustness of crowdsourcing tasks

Abstract: Crowdsourcing successfully strives to become a widely used means of collecting large-scale scientific corpora. Many research fields, including Information Retrieval, rely on this novel way of data acquisition. However, it seems to be undermined by a significant share of workers that are primarily interested in producing quick generic answers rather than correct ones in order to optimise their time-efficiency and, in turn, earn more money. Recently, we have seen numerous sophisticated schemes of identifying suc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
114
2
1

Year Published

2012
2012
2020
2020

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 135 publications
(117 citation statements)
references
References 18 publications
(18 reference statements)
0
114
2
1
Order By: Relevance
“…However crowd-sourcing is not a perfect system, issues such as quality control and reliability of results need further investigation. Automatic quality control methods, as the gold units used in this study, can be a good option since data cleaning is a time-consuming and tedious task [14]. However, they do not always ensure high quality data; since answers can only be rejected at runtime, they may attract more spammers, malicious, and sloppy workers [12] [14].…”
Section: Validation Of Crowd-sourcing Analysismentioning
confidence: 99%
“…However crowd-sourcing is not a perfect system, issues such as quality control and reliability of results need further investigation. Automatic quality control methods, as the gold units used in this study, can be a good option since data cleaning is a time-consuming and tedious task [14]. However, they do not always ensure high quality data; since answers can only be rejected at runtime, they may attract more spammers, malicious, and sloppy workers [12] [14].…”
Section: Validation Of Crowd-sourcing Analysismentioning
confidence: 99%
“…В работах [68,69] делается попытка выделения причин получения некачественных результатов в крауд-вычислениях. Среди основных факторов, влияющих на качество, названы компетентность, заинтересованность в выполнении заданий (мотивация), ясность представления заданий, наличие или отсутствие «злого умысла» (целенаправленного подрыва работы системы).…”
Section: теоретико-игровые методыunclassified
“…A commonly used solution is to employ a worker reputation system with assigning tasks to workers with approval ratings above a certain pre-set level [33]. Another set of methods of identification and expulsion of the unethical workers is based on a set of indices measuring (1) agreement with the expert "golden standard" data; (2) agreement with the other workers; (3) agreement with the attention check questions and (4) an amount of effort estimated from the task completion time [34]. The "golden standard" is a subset of data that is processed by experts in the field; an important condition is that a lay person should be able to process this data easily and unambitiously.…”
mentioning
confidence: 99%
“…The attention check and language comprehension questions are verifiable questions [29] that do not require factual knowledge [36]; the results obtained from the workers failing to answer the attention questions correctly should be discarded. Finally, the average time to complete a single task is used to identify low-quality workers presumably spending a lesser amount of time per task [34].…”
mentioning
confidence: 99%
See 1 more Smart Citation