Proceedings of the 7th ACM International Conference on Web Search and Data Mining 2014
DOI: 10.1145/2556195.2556266
|View full text |Cite
|
Sign up to set email alerts
|

Using linked data to mine RDF from wikipedia's tables

Abstract: The tables embedded in Wikipedia articles contain rich, semi-structured encyclopaedic content. However, the cumulative content of these tables cannot be queried against. We thus propose methods to recover the semantics of Wikipedia tables and, in particular, to extract facts from them in the form of RDF triples. Our core method uses an existing Linked Data knowledge-base to find pre-existing relations between entities in Wikipedia tables, suggesting the same relations as holding for other entities in analogous… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
50
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 56 publications
(50 citation statements)
references
References 16 publications
0
50
0
Order By: Relevance
“…Any predicates found in the query result are considered as relations between the two entities. The work is later extended in Muñoz et al [24] by adding a machine learning process to filter triples that are likely to be incorrect, exploiting features derived from both the knowledge base and the text content from the target cells.…”
Section: Semantic Table Interpretationmentioning
confidence: 99%
See 3 more Smart Citations
“…Any predicates found in the query result are considered as relations between the two entities. The work is later extended in Muñoz et al [24] by adding a machine learning process to filter triples that are likely to be incorrect, exploiting features derived from both the knowledge base and the text content from the target cells.…”
Section: Semantic Table Interpretationmentioning
confidence: 99%
“…Current methods are non-efficient because they typically adopt an exhaustive strategy that examines the entire table content, e.g., column classification depends on every cell in the column. This results in quadratic growth of the number of computations and knowledge base queries with respect to the size of tables, as such operations are usually required for every pair of candidates, e.g., candidate relation lookup between every pair of entities on the same row [26,23,24], or similarity computation between every pair of candidate entity and concept in a column [21]. This can be redundant as Zwicklbauer et al [46] have empirically shown that comparable accuracy can be obtained by using only a fraction of data (i.e., sample) from the column.…”
Section: Remarkmentioning
confidence: 99%
See 2 more Smart Citations
“…Therefore, several approaches for interpreting tables from Wikipedia with LOD have been proposed. Munoz et al [47,48] propose methods for triplifying Wikipedia tables, called WikiTables, using existing LOD knowledge bases, like DBpedia and YAGO. Following the idea of the previous approaches, this approach starts by extracting entities from the tables, and then discovering existing relations between them.…”
Section: Using Lod To Interpret Semi-structured Datamentioning
confidence: 99%