Abstract:The World Wide Web and the Semantic Web are designed as a network of distributed services and datasets. In this network and its genesis, collaboration played and still plays a crucial role. But currently we only have central collaboration solutions for RDF data, such as SPARQL endpoints and wiki systems, while decentralized solutions can enable applications for many more use-cases. Inspired by a successful distributed source code management methodology in software engineering a framework to support distributed… Show more
“…sections 10.4 to 10.7). With the improved implementation we could even reach a 3.6× improvement of the query throughput over the speed of the old implementation, at the state of development as published in our previous paper [2]. This improvement can mainly be led back to a completely reworked architecture, which is presented in section 9, and a number of optimizations based on profiling results, especially by reducing I/O overhead.…”
Section: Discussionmentioning
confidence: 78%
“…section 10.5) and 641QM pH for the baseline. This is an improvement of 3.7× over the speed of the old implementation of 67QM pH (old quit), at the state of development as published in our previous paper [2].…”
Section: Query Throughputmentioning
confidence: 75%
“…In our case we find the most recent common ancestor δ := B({G}). Now we have to compare (2) the graphs between the merge base and the individual branches and calculate their difference:…”
Section: Three-way-merge: An Unsupervised Approach To Merge Branched ...mentioning
confidence: 99%
“…Apart from documents, datasets are gaining more attention on the World Wide Web. An increasing number of the datasets on the Web are available as Linked Data, also called the Linked Open Data Cloud 1 or Giant Global Graph 2 . Collaboration of people and machines is a major aspect of the World Wide Web and as well of the Semantic Web.…”
Section: Introductionmentioning
confidence: 99%
“…For this we first introduce a formal model to describe and transform the operations implemented in Git to operations on RDF datasets. The formal model was published in [2]. In this paper, we can present a more elaborated formal model that was especially extended and improved regarding the atomic partitioning.…”
The World Wide Web and the Semantic Web are designed as a network of distributed services and datasets. The distributed character of the Web brings manifold collaborative possibilities to interchange data. The commonly adopted collaborative solutions for RDF data are centralized (e. g. SPARQL endpoints and wiki systems). But to support distributed collaboration, a system is needed, that supports divergence of datasets, brings the possibility to conflate diverged states, and allows distributed datasets to be synchronized. In this paper, we present Quit Store, it was inspired by and it builds upon the successful Git system. The approach is based on a formal expression of evolution and consolidation of distributed datasets. During the collaborative curation process, the system automatically versions the RDF dataset and tracks provenance information. It also provides support to branch, merge, and synchronize distributed RDF datasets. The merging process is guarded by specific merge strategies for RDF data. Finally, we use our reference implementation to show overall good performance and demonstrate the practical usability of the system. (Edgard Marx) URL: http://aksw.org/NatanaelArndt (Natanael Arndt), http://aksw.org/NormanRadtke (Norman Radtke), http://aksw.org/MichaelMartin (Michael Martin), http://aksw.org/EdgardMarx (Edgard Marx) 1 http://lod-cloud.net/ 2
“…sections 10.4 to 10.7). With the improved implementation we could even reach a 3.6× improvement of the query throughput over the speed of the old implementation, at the state of development as published in our previous paper [2]. This improvement can mainly be led back to a completely reworked architecture, which is presented in section 9, and a number of optimizations based on profiling results, especially by reducing I/O overhead.…”
Section: Discussionmentioning
confidence: 78%
“…section 10.5) and 641QM pH for the baseline. This is an improvement of 3.7× over the speed of the old implementation of 67QM pH (old quit), at the state of development as published in our previous paper [2].…”
Section: Query Throughputmentioning
confidence: 75%
“…In our case we find the most recent common ancestor δ := B({G}). Now we have to compare (2) the graphs between the merge base and the individual branches and calculate their difference:…”
Section: Three-way-merge: An Unsupervised Approach To Merge Branched ...mentioning
confidence: 99%
“…Apart from documents, datasets are gaining more attention on the World Wide Web. An increasing number of the datasets on the Web are available as Linked Data, also called the Linked Open Data Cloud 1 or Giant Global Graph 2 . Collaboration of people and machines is a major aspect of the World Wide Web and as well of the Semantic Web.…”
Section: Introductionmentioning
confidence: 99%
“…For this we first introduce a formal model to describe and transform the operations implemented in Git to operations on RDF datasets. The formal model was published in [2]. In this paper, we can present a more elaborated formal model that was especially extended and improved regarding the atomic partitioning.…”
The World Wide Web and the Semantic Web are designed as a network of distributed services and datasets. The distributed character of the Web brings manifold collaborative possibilities to interchange data. The commonly adopted collaborative solutions for RDF data are centralized (e. g. SPARQL endpoints and wiki systems). But to support distributed collaboration, a system is needed, that supports divergence of datasets, brings the possibility to conflate diverged states, and allows distributed datasets to be synchronized. In this paper, we present Quit Store, it was inspired by and it builds upon the successful Git system. The approach is based on a formal expression of evolution and consolidation of distributed datasets. During the collaborative curation process, the system automatically versions the RDF dataset and tracks provenance information. It also provides support to branch, merge, and synchronize distributed RDF datasets. The merging process is guarded by specific merge strategies for RDF data. Finally, we use our reference implementation to show overall good performance and demonstrate the practical usability of the system. (Edgard Marx) URL: http://aksw.org/NatanaelArndt (Natanael Arndt), http://aksw.org/NormanRadtke (Norman Radtke), http://aksw.org/MichaelMartin (Michael Martin), http://aksw.org/EdgardMarx (Edgard Marx) 1 http://lod-cloud.net/ 2
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.