2018
DOI: 10.3233/sw-160239
|View full text |Cite
|
Sign up to set email alerts
|

Detecting Linked Data quality issues via crowdsourcing: A DBpedia study

Abstract: In this paper we examine the use of crowdsourcing as a means to detect Linked Data quality problems that are difficult to uncover automatically. We base our approach on the analysis of the most common errors encountered in the DBpedia dataset, and a classification of these errors according to the extent to which they are likely to be amenable to crowdsourcing. We then propose and study different crowdsourcing approaches to identify these Linked Data quality issues, employing DBpedia as our use case: (i) a cont… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 30 publications
(25 citation statements)
references
References 51 publications
0
24
0
Order By: Relevance
“…MOVIE, based on IMDb 6 [2] and WiKiData [4], is a knowledge base with entertainment-related facts mostly pertaining to actors, directors, movies, TV series, musicals etc. It contains more than 2 million factual triples.…”
Section: Dataset Descriptionmentioning
confidence: 99%
“…MOVIE, based on IMDb 6 [2] and WiKiData [4], is a knowledge base with entertainment-related facts mostly pertaining to actors, directors, movies, TV series, musicals etc. It contains more than 2 million factual triples.…”
Section: Dataset Descriptionmentioning
confidence: 99%
“…This special issue attracted a total of 10 submissions from which three papers [1,9,14] were accepted for publication as summarized in the next sections. A fourth paper is under review during the writing of this editorial.…”
Section: Special Issue Papersmentioning
confidence: 99%
“…This paper focuses on the problem of verifying the quality of Linked Data, in particular data from DBpedia [1]. As such it is illustrative of the scenario, where HC&C is used for knowledge validation and enhancement (HC4SW-Kn.Validation).…”
Section: Detecting Linked Data Quality Issues Via Crowdsourcing: a Dbmentioning
confidence: 99%
“…An example of RDF triples is (dbr:Birmingham dbo:populationTotal "1123000"ˆˆxsd:integer ), which represents the fact that the city of Birmingham has a total population of 1123000 (dbr and dbo are the namespace prefixes of DBpedia repositories). 1 In recent years, several large-scale knowledge graphs have been constructed such as DBpedia 2 , YAGO 3 , Freebase 4 , Wikidata 5 , and others. Many of these knowledge graphs were created by extracting Web contents or through crowdsourcing.…”
Section: Introductionmentioning
confidence: 99%
“…These processes could be very noisy, and the created knowledge graphs are unlikely to be fully correct. There is an increasing interest in quality assessment for knowledge graphs [1] [4] [17] [20] [8] [6]. Some approaches focus on completing or correcting entity type information, while others target towards relations between entities, or interlinks between di↵erent knowledge graphs.…”
Section: Introductionmentioning
confidence: 99%