Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER) 2019
DOI: 10.18653/v1/d19-6607
|View full text |Cite
|
Sign up to set email alerts
|

Scalable Knowledge Graph Construction from Text Collections

Abstract: We present a scalable, open-source platform that "distills" a potentially large text collection into a knowledge graph. Our platform takes documents stored in Apache Solr and scales out the Stanford CoreNLP toolkit via Apache Spark integration to extract mentions and relations that are then ingested into the Neo4j graph database. The raw knowledge graph is then enriched with facts extracted from an external knowledge graph. The complete product can be manipulated by various applications using Neo4j's native Cy… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 19 publications
(6 citation statements)
references
References 13 publications
0
5
0
Order By: Relevance
“…While some works refer to knowledge found in texts or other resources as evidence for a fact [9,46,127,153] and call it fact only after the truthfulness has been determined and that knowledge is entered into a knowledge base, other works assume the truthfulness of the mentions and refer to them or the knowledge they represent as facts directly [32,71]. Very related is the task of Truth Discovery.…”
Section: Facts Vs Evidencementioning
confidence: 99%
“…While some works refer to knowledge found in texts or other resources as evidence for a fact [9,46,127,153] and call it fact only after the truthfulness has been determined and that knowledge is entered into a knowledge base, other works assume the truthfulness of the mentions and refer to them or the knowledge they represent as facts directly [32,71]. Very related is the task of Truth Discovery.…”
Section: Facts Vs Evidencementioning
confidence: 99%
“…Data-to-Text Generation Data-to-Text Generation has several benchmark datasets with slightly different objectives-WebNLG (Gardent et al, 2017) to convert a group of triples to text, E2ENLG (Dušek et al, 2018) (Etzioni et al, 2008;Angeli et al, 2015;Clancy et al, 2019) inherently create such a corpus but these works generally do not release the extracted KG triples.…”
Section: Related Workmentioning
confidence: 99%
“…There is a vast literature on the inverse task of automatic KG construction from text (Etzioni et al, 2008;Angeli et al, 2015;Clancy et al, 2019), however these works generally describe the methodology and do not release the corresponding dataset. Figure 2: KG verbalization process.…”
Section: Kg-text Alignmentmentioning
confidence: 99%