Abstract:Background: The changing landscape of genomics research and clinical practice has created a need for computational pipelines capable of efficiently orchestrating complex analysis stages while handling large volumes of data across heterogeneous computational environments. Workflow Management Systems (WfMSs) are the software components employed to fill this gap.
Results: This work provides an approach and systematic evaluation of key features of popular bioinformatics WfMSs in use today: Nextflow, CWL, and WDL … Show more
“…In neuroscience, this can include processing of functional Magnetic Resonance Imaging (fMRI) files requiring specialised techniques to handle motion correction (correcting the movement of the subject within the scanner) and smoothing of the data (to average out the noise present in the measurement). Finally, the datasets, tools, and computational platforms available to a neuroscientist are highly domain-specific such as repositories of mouse brain scans 2 .…”
Section: Domain-specificitymentioning
confidence: 99%
“…managers using the categories of ease of use, expressiveness, portability, scalability, learning resources, and pipeline initiatives [115]. Admed et al mention modularity and reproducibility amongst others [2], while Kortelainen adds the important characteristics of licensing and maturity [63]. Other factors may be connection to specialised tools or computing platforms such as the Hermes middleware platform for increased scalability [61].…”
Domain experts are increasingly employing machine learning to solve their domain-specific problems. This article presents six key challenges that a domain expert faces in transforming their problem into a computational workflow, and then into an executable implementation. These challenges arise out of our conceptual framework which presents the "route" of options that a domain expert may choose to take while developing their solution.To ground our conceptual framework in the stateof-the-practice, this article discusses a selection of available textual and graphical workflow systems and their support for these six challenges. Case studies from the literature in various domains are also examined to highlight the tools used by the domain experts as well as a classification of the domain-specificity and machine learning usage of their problem, workflow, and implementation.The state-of-the-practice informs our discussion of the six key challenges, where we identify which challenges are not sufficiently addressed by available tools. We also suggest possible research directions for software engineering researchers to increase the automation of these tools and disseminate best-practice techniques between software engineering and various scientific domains.
“…In neuroscience, this can include processing of functional Magnetic Resonance Imaging (fMRI) files requiring specialised techniques to handle motion correction (correcting the movement of the subject within the scanner) and smoothing of the data (to average out the noise present in the measurement). Finally, the datasets, tools, and computational platforms available to a neuroscientist are highly domain-specific such as repositories of mouse brain scans 2 .…”
Section: Domain-specificitymentioning
confidence: 99%
“…managers using the categories of ease of use, expressiveness, portability, scalability, learning resources, and pipeline initiatives [115]. Admed et al mention modularity and reproducibility amongst others [2], while Kortelainen adds the important characteristics of licensing and maturity [63]. Other factors may be connection to specialised tools or computing platforms such as the Hermes middleware platform for increased scalability [61].…”
Domain experts are increasingly employing machine learning to solve their domain-specific problems. This article presents six key challenges that a domain expert faces in transforming their problem into a computational workflow, and then into an executable implementation. These challenges arise out of our conceptual framework which presents the "route" of options that a domain expert may choose to take while developing their solution.To ground our conceptual framework in the stateof-the-practice, this article discusses a selection of available textual and graphical workflow systems and their support for these six challenges. Case studies from the literature in various domains are also examined to highlight the tools used by the domain experts as well as a classification of the domain-specificity and machine learning usage of their problem, workflow, and implementation.The state-of-the-practice informs our discussion of the six key challenges, where we identify which challenges are not sufficiently addressed by available tools. We also suggest possible research directions for software engineering researchers to increase the automation of these tools and disseminate best-practice techniques between software engineering and various scientific domains.
BACKGROUND
Data Interoperability in healthcare implies smooth information from multiple health- care systems, diverse providers, and multiple patient settings. Data interoperability has become increasingly important since technological advances propose a new landscape to achieve it. However, several barriers exist, such as inconsistent information across multiple sources or organizational resistance to sharing data under the value chain perspective.
OBJECTIVE
This study outlines a systematic literature review of the current research on data interoperability in the healthcare value chain.
METHODS
The study analyzed publication datasets from 2011 to 2022 from Scopus, Clarivates, IEEE Xplore, and PubMed/MEDLINE. The study methodology for querying, excluding, including, and selecting relevant research publications from scientific databases is detailed. The chosen research studies are classified and contrasted based on the scope and ontology.
RESULTS
Our study summarizes the state-of-the-art of healthcare value chain interoperability. The study included published articles from 2012 to 2022 after filtering from 5,834 studies gathered before the inclusion and exclusion criteria. The study found that 60% of the research studies were published from 2012 to 2022, which shows a considerable increment use of considering the interoperability topic in the last decade.
CONCLUSIONS
Our study is one of the initial efforts to develop a systematic review investigating data interoperability in the healthcare value chain holistically. It was shown that achieving interoperability is challenging because of the different data schemes, data legacy infrastructure, and need for data standardization, among other factors.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.