We introduce Semantic Ontology-Controlled application for web Content Management Systems (SOCCOMAS), a development framework for FAIR (‘findable’, ‘accessible’, ‘interoperable’, ‘reusable’) Semantic Web Content Management Systems (S-WCMSs). Each S-WCMS run by SOCCOMAS has its contents managed through a corresponding knowledge base that stores all data and metadata in the form of semantic knowledge graphs in a Jena tuple store. Automated procedures track provenance, user contributions and detailed change history. Each S-WCMS is accessible via both a graphical user interface (GUI), utilizing the JavaScript framework AngularJS, and a SPARQL endpoint. As a consequence, all data and metadata are maximally findable, accessible, interoperable and reusable and comply with the FAIR Guiding Principles. The source code of SOCCOMAS is written using the Semantic Programming Ontology (SPrO). SPrO consists of commands, attributes and variables, with which one can describe an S-WCMS. We used SPrO to describe all the features and workflows typically required by any S-WCMS and documented these descriptions in a SOCCOMAS source code ontology (SC-Basic). SC-Basic specifies a set of default features, such as provenance tracking and publication life cycle with versioning, which will be available in all S-WCMS run by SOCCOMAS. All features and workflows specific to a particular S-WCMS, however, must be described within an instance source code ontology (INST-SCO), defining, e.g. the function and composition of the GUI, with all its user interactions, the underlying data schemes and representations and all its workflow processes. The combination of descriptions in SC-Basic and a given INST-SCO specify the behavior of an S-WCMS. SOCCOMAS controls this S-WCMS through the Java-based middleware that accompanies SPrO, which functions as an interpreter. Because of the ontology-controlled design, SOCCOMAS allows easy customization with a minimum of technical programming background required, thereby seamlessly integrating conventional web page technologies with semantic web technologies. SOCCOMAS and the Java Interpreter are available from (https://github.com/SemanticProgramming).
The landscape of currently existing repositories of specimen data consists of isolated islands, with each applying its own underlying data model. Using standardized protocols such as DarwinCore or ABCD, specimen data and metadata are exchanged and published on web portals such as GBIF. However, data models differ across repositories. This can lead to problems when comparing and integrating content from different systems. for example, in one system there is a field with the label 'determination', in another there is a field with the label 'taxonomic identification'. Both might refer to the same concepts of organism identification process (e.g., 'obi:organism identification assay'; http://purl.obolibrary.org/obo/OBI_0001624), but the intuitive meaning of the content is not clear and the understanding of the providers of the information might differ from that of the users. Without additional information, data integration across isolated repositories is thus difficult and error-prone. As a consequence, interoperability and retrievability of data across isolated repositories is difficult. Linked Open Data (LOD) promises an improvement. URIs can be used for concepts that are ideally created and accepted by a community and that provide machine-readable meanings. LOD thereby supports transfer of data into information and then into knowledge, thus making the data FAIR (Findable, Accessible, Interoperable, Reusable; Wilkinson et al. 2016). Annotating specimen associated data with LOD, therefore, seems to be a promising approach to guarantee interoperability across different repositories. However, all currently used specimen collection management systems are based on relational database systems, which lack semantic transparency and thus do not provide easily accessible, machine-readable meanings for the terms used in their data models. As a consequence, transferring their data contents into an LOD framework may lead to loss or misinterpretation of information. This discrepancy between LOD and relational databases results from the lack of semantic transparency and machine-readability of data in relational databases. Storing specimen collection data as semantic Knowledge Graphs provides semantic transparency and machine-readability of data. Semantic Knowledge Graphs are graphs that are based on the syntax of ‘Subject – Property – Object’ of the Resource Description Framework (RDF). The ‘Subject’ and ‘Property’ position is taken by URIs and the ‘Object’ position can be taken either by a URI or by a label or value. Since a given URI can take the ‘Subject’ position in one RDF statement and the ‘Object’ position in another RDF statement, several RDF statements can be connected to form a directed labeled graph, i.e. a semantic graph. Semantic Knowledge Graphs are graphs in which each described specimen and its parts and properties possess their own URI and thus can be individually referenced. These URIs are used to describe the respective specimen and its properties using the RDF syntax. Additional RDF statements specify the ontology class that each part and property instantiates. The reference to the URIs of the instantiated ontology classes guarantees the Findability, Interoperability, and Reusability of information contained in semantic Knowledge Graphs. Specimen collection data contained in semantic Knowledge Graphs can be made Accessible in a human-readable form through an interface and in a machine-readable form through a SPARQL endpoint (https://en.wikipedia.org/wiki/SPARQL). As a consequence, semantic Knowledge Graphs comply with the FAIR guiding principles. By using URIs for the semantic Knowledge Graph of each specimen in the collection, it is also available as LOD. With semantic Morph·D·Base, we have implemented a prototype to this approach that is based on Semantic Programming. We present the prototype and discuss different aspects of how specimen collection data are handled. By using community created terminologies and standardized methods for the contents created (e.g. species identification) as well as URIs for each expression, we make the data and metadata semantically transparent and communicable. The source code for Semantic Programming and for semantic Morph·D·Base is available from https://github.com/SemanticProgramming. The prototype of semantic Morph·D·Base can be accessed here: https://proto.morphdbase.de.
Creating an application for recording and documenting morphological data in a semantically transparent and reproducible way used to be a challenging task due to the heterogeneous nature of data within this domain. To provide a system for morphologists and taxonomists to work with their research data, collaborate and publish it, we built a Web-based semantic content management system for creating formalized morphological descriptions that are stored in a Linked Open Data ready tuple store. The application is completely controlled by a set of application ontologies that define the user interface, input and data views as well as database workflows, access rights, etc. Domain experts can create their own instance of the content management system that is tailored to their special needs by altering or extending the existing application ontologies.
We demonstrate the early prototype of a new module for Morph·D·Base that allows the generation of highly formalized semantic morphological descriptions (http:// escience.biowikifarm.net/wiki/EScience-Compliant_Standards_for_Morphology).The resulting morphological descriptions follow the individuals-based Instance Anatomy data scheme (as opposed to the class-based Semantic Phenotypes data scheme). The module allows the description of a specimen's anatomy by generating a granular representation of the parts of the specimen to be described, using ontology-terms from known ontologies. This results in a hierarchy of parts and subparts (partonomy), which serves as organizational backbone of the entire description, with each part representing a section of the description to which you can navigate using the partonomy. The module allows the description of each part from the partonomy using (1) a set of formalized input forms, which also allow the specification of metadata for each input field, (2) a text-widget for providing conventional free-text descriptions, which can be semantically enriched through annotating them with ontology-terms of (user-)selected ontologies, and (3) an image-widget for linking images, which allows semantically enriching each image by specifying regions of interest and annotating them with ontology-terms of (user-)selected ontologies. This new module is ‡ ‡ § § § § ‡
Ontologies are usually utilized for representing knowledge. Here, we extend this use and demonstrate that ontologies also can be used for describing and controlling semantic WebContent-Management-Systems (WCMS). We call the resulting application SOCCOMAS: a self-describing and content-independent application for semantic ontology-controlled WebContent-Management-Systems (http://escience.biowikifarm.net/wiki/SOCCOMAS). SOCCOMAS manages the contents of its WCMS through a corresponding knowledge base that is stored in a tuple store. It makes its contents accessible via a flexible and interactive graphical user interface (GUI), utilizing the JavaScript framework AngularJS. Central to SOCCOMAS is a set of application ontologies and a Java middleware, both of which have been developed by us. The ontologies contain the descriptions that control the behavior of the WCMS and the Java middleware interprets these descriptions as an operation-language with commands, attributes and variables. This language is used for describing and therewith defining the function and design of the GUI of the WCMS, including all its user interactions, the underlying data schemes and representations, and all workflow processes of the WCMS. The core application implements a set of default features, such as versioning, provenance and access rights management. These features ‡ ‡ § § § § ‡
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.