2004
DOI: 10.1007/978-3-540-30475-3_20
|View full text |Cite
|
Sign up to set email alerts
|

An Evaluation of Knowledge Base Systems for Large OWL Datasets

Abstract: Abstract. In this paper, we present our work on evaluating knowledge base systems with respect to use in large OWL applications. To this end, we have developed the Lehigh University Benchmark (LUBM). The benchmark is intended to evaluate knowledge base systems with respect to extensional queries over a large dataset that commits to a single realistic ontology. LUBM features an OWL ontology modeling university domain, synthetic OWL data generation that can scale to an arbitrary size, fourteen test queries repre… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
75
0
1

Year Published

2005
2005
2008
2008

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 104 publications
(77 citation statements)
references
References 12 publications
1
75
0
1
Order By: Relevance
“…BigOWLIM 8 is a scalable repository supporting structured queries but uses its own proprietary storage and index format. LUBM [18] benchmark are developed alongside those work to evaluate semantic web knowledge base systems [21].…”
Section: Related Workmentioning
confidence: 99%
“…BigOWLIM 8 is a scalable repository supporting structured queries but uses its own proprietary storage and index format. LUBM [18] benchmark are developed alongside those work to evaluate semantic web knowledge base systems [21].…”
Section: Related Workmentioning
confidence: 99%
“…The first weakness is addressed by using efficient and large-scale ontology repositories [17] in combination with Lucene 3 . Lucene indexes the semantic entities in the online and distributed back-end repositories into one or more indexes, and is used as our fast search engine 4 , which supports fuzzy searches based on the Lavenshtein Distance, or Edit Distance algorithm.…”
Section: Phase I: Syntactic Mappingmentioning
confidence: 99%
“…− a big set of 255Mb, containing 3,196,692 statements, 813,479 unique resources and the average node degree of 3.90 in the biggest connected component. − a small synthetic dataset, generated to include three ontologies (business, sports, and entertainment); 14Mb in size, containing 104,891 statements, 29,825 unique resources and the average node degree of 3.86 in the biggest component, and − a big synthetic set, generated as Univ(50, 0) using the Lehigh University Benchmark [10], 556Mb in size, containing 6,888,642 statements, 1,082,818 unique resources and the average node degree of 6.09.…”
Section: Data Setsmentioning
confidence: 99%