2014
DOI: 10.1007/978-3-319-11915-1_31
|View full text |Cite
|
Sign up to set email alerts
|

CrowdTruth: Machine-Human Computation Framework for Harnessing Disagreement in Gathering Annotated Data

Abstract: Abstract. In this paper, we introduce the CrowdTruth open-source software framework for machine-human computation, that implements a novel approach to gathering human annotation data in a wide range of annotation tasks and on a variety of media (e.g. text, images, videos). The CrowdTruth approach captures human semantics through a pipeline of three processes: a) combining various machine processing of text, image and video in order to understand better the input content and optimise its suitability for micro-t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
36
0

Year Published

2015
2015
2016
2016

Publication Types

Select...
5
2

Relationship

2
5

Authors

Journals

citations
Cited by 63 publications
(36 citation statements)
references
References 10 publications
0
36
0
Order By: Relevance
“…We have published work on CrowdTruth [14], our framework for crowdsourcing ground truth data. CrowdTruth connects with both Amazon Mechanical Turk and CrowdFlower for launching and monitoring tasks, and implements the disagreement metrics for a live analysis of the results from the crowd.…”
Section: Preliminary Resultsmentioning
confidence: 99%
“…We have published work on CrowdTruth [14], our framework for crowdsourcing ground truth data. CrowdTruth connects with both Amazon Mechanical Turk and CrowdFlower for launching and monitoring tasks, and implements the disagreement metrics for a live analysis of the results from the crowd.…”
Section: Preliminary Resultsmentioning
confidence: 99%
“…This is an interesting phenomena, also observed in other knowledge-intensive content annotation use cases (e.g. medical [33]). The result suggests the need for more articulated annotations campaigns, possibly organised in workflows [34] that interleave automatic and human operations.…”
Section: Accepted Manuscriptmentioning
confidence: 89%
“…Novel frameworks for quality modeling and assessment are needed to reflect the richness of domain knowledge and insight that collective intelligence could bring into traditional computational processes, going beyond the rigid, Boolean view of ground truth used in many computer science areas. The work of Inel et al (2014) around the CrowdTruth framework proposes a useful starting point. Methods that match ontologies to microtask design could also offer a helpful way to prune the space of possible solutions.…”
Section: Quality Of Contributionsmentioning
confidence: 99%
“…Sarasua and Thimm (2014) propose Crowd Work CV, an ontology to capture crowd workers' and requesters' information across different crowdsourcing platforms. The CrowdTruth framework proposes a useful schema for annotating the provenance of crowdsourced data (Inel et al, 2014). A second challenge will be to identify the most appropriate level of granularity to record activities and their results to meet the requirements of various use cases (data aggregation, quality control, tasks assignment, personalization etc.…”
Section: Defining Vocabularies or Ontologiesmentioning
confidence: 99%
See 1 more Smart Citation