2016
DOI: 10.1101/048744
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Ten Simple Rules for Taking Advantage of git and GitHub

Abstract: IntroductionBioinformatics is a broad discipline in which one common denominator is the need to produce and/or use software that can be applied to biological data in different contexts. To enable and ensure the replicability and traceability of scientific claims, it is essential that the scientific publication, the corresponding datasets, and the data analysis are made publicly available [1,2]. All software used for the analysis should be either carefully documented (e.g., for commercial software) or, better y… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
37
0
3

Year Published

2017
2017
2022
2022

Publication Types

Select...
6
3

Relationship

2
7

Authors

Journals

citations
Cited by 37 publications
(40 citation statements)
references
References 19 publications
0
37
0
3
Order By: Relevance
“…Most research data and data processing will fit comfortably within these limits (the largest file in the Portal database is currently <20 MB and it takes <15 minutes for all data checking and processing code to run), so we think this type of system will work for the majority of research projects. However, in cases where larger data files or longer run times are necessary, it is possible to adapt our general approach by using equivalent tools that can be run on local computing resources (e.g, GitLab for managing git repositories and Jenkins for continuous integration) and using tools that are designed for versioning large data (e.g., dat; Ogden, McKelvey, & Madsen, 2017;or git Large File Storage;Perez-Riverol et al 2016).…”
Section: Discussionmentioning
confidence: 99%
“…Most research data and data processing will fit comfortably within these limits (the largest file in the Portal database is currently <20 MB and it takes <15 minutes for all data checking and processing code to run), so we think this type of system will work for the majority of research projects. However, in cases where larger data files or longer run times are necessary, it is possible to adapt our general approach by using equivalent tools that can be run on local computing resources (e.g, GitLab for managing git repositories and Jenkins for continuous integration) and using tools that are designed for versioning large data (e.g., dat; Ogden, McKelvey, & Madsen, 2017;or git Large File Storage;Perez-Riverol et al 2016).…”
Section: Discussionmentioning
confidence: 99%
“…To keep track of these changes, EXSIMOs are built on top of version control. In our example we take advantage of the features of git and GitHub (https://github.com/matthiaskoenig/exsimo, Figure 2H) [11]. By using version control we enable collaborative work (merging changes), managing different versions (creating branches), tracking of changes (analyzing diffs), reuse (forking the repository), and versioning (using tags) of an EXSIMO.…”
Section: "Talk Is Cheap Show Me the Code" -Linus Torvaldsmentioning
confidence: 99%
“…babette's development takes place on GitHub, https://github.com/richelbilderbeek/babette, which accommodates collaboration (Gorgolewski & Poldrack, ). and improves transparency (Perez‐Riverol et al., ) babette's GitHub facilitates feature requests and has guidelines how to do so.…”
Section: Babette Resourcesmentioning
confidence: 99%