HippoUnit: A software tool for the automated testing and systematic comparison of detailed models of hippocampal neurons based on electrophysiological data
Abstract:Anatomically and biophysically detailed data-driven neuronal models have become widely used tools for understanding and predicting the behavior and function of neurons. Due to the increasing availability of experimental data from anatomical and electrophysiological measurements as well as the growing number of computational and software tools that enable accurate neuronal modeling, there are now a large number of different models of many cell types available in the literature. These models were usually built t… Show more
“…After preprocessing, both datasets are passed to machine learning pipelines. StandardScaler is used to scale all the features; standardized values are useful for tracking the data, which are difficult to compare otherwise due to different magnitudes, metrics, or circumstances [ 30 ]. The Python sklearn package is used to create machine learning pipelines; these pipelines are an ensemble of several transformers with a final estimator [ 26 ].…”
Section: Resultsmentioning
confidence: 99%
“…After preprocessing, both datasets are passed to machine learning pipelines. StandardScaler is used to scale all the features; standardized values are useful for tracking the data, which are difficult to compare otherwise due to different magnitudes, metrics, or circumstances [30].…”
Purpose. Age can be an important clue in uncovering the identity of persons that left biological evidence at crime scenes. With the availability of DNA methylation data, several age prediction models are developed by using statistical and machine learning methods. From epigenetic studies, it has been demonstrated that there is a close association between aging and DNA methylation. Most of the existing studies focused on healthy samples, whereas diseases may have a significant impact on human age. Therefore, in this article, an age prediction model is proposed using DNA methylation biomarkers for healthy and diseased samples. Methods. The dataset contains 454 healthy samples and 400 diseased samples from publicly available sources with age (1–89 years old). Six CpG sites are identified from this data having a high correlation with age using Pearson’s correlation coefficient. In this work, the age prediction model is developed using four different machine learning techniques, namely, Multiple Linear Regression, Support Vector Regression, Gradient Boosting Regression, and Random Forest Regression. Separate models are designed for healthy and diseased data. The data are split randomly into 80 : 20 ratios for training and testing, respectively. Results. Among all the techniques, the model designed using Random Forest Regression shows the best performance, and Gradient Boosting Regression is the second best model. In the case of healthy samples, the model achieved a MAD of 2.51 years for training data and 4.85 for testing data. Also, for diseased samples, a MAD of 3.83 years is obtained for training and 9.53 years for testing. Conclusion. These results showed that the proposed model can predict age for healthy and diseased samples.
“…After preprocessing, both datasets are passed to machine learning pipelines. StandardScaler is used to scale all the features; standardized values are useful for tracking the data, which are difficult to compare otherwise due to different magnitudes, metrics, or circumstances [ 30 ]. The Python sklearn package is used to create machine learning pipelines; these pipelines are an ensemble of several transformers with a final estimator [ 26 ].…”
Section: Resultsmentioning
confidence: 99%
“…After preprocessing, both datasets are passed to machine learning pipelines. StandardScaler is used to scale all the features; standardized values are useful for tracking the data, which are difficult to compare otherwise due to different magnitudes, metrics, or circumstances [30].…”
Purpose. Age can be an important clue in uncovering the identity of persons that left biological evidence at crime scenes. With the availability of DNA methylation data, several age prediction models are developed by using statistical and machine learning methods. From epigenetic studies, it has been demonstrated that there is a close association between aging and DNA methylation. Most of the existing studies focused on healthy samples, whereas diseases may have a significant impact on human age. Therefore, in this article, an age prediction model is proposed using DNA methylation biomarkers for healthy and diseased samples. Methods. The dataset contains 454 healthy samples and 400 diseased samples from publicly available sources with age (1–89 years old). Six CpG sites are identified from this data having a high correlation with age using Pearson’s correlation coefficient. In this work, the age prediction model is developed using four different machine learning techniques, namely, Multiple Linear Regression, Support Vector Regression, Gradient Boosting Regression, and Random Forest Regression. Separate models are designed for healthy and diseased data. The data are split randomly into 80 : 20 ratios for training and testing, respectively. Results. Among all the techniques, the model designed using Random Forest Regression shows the best performance, and Gradient Boosting Regression is the second best model. In the case of healthy samples, the model achieved a MAD of 2.51 years for training data and 4.85 for testing data. Also, for diseased samples, a MAD of 3.83 years is obtained for training and 9.53 years for testing. Conclusion. These results showed that the proposed model can predict age for healthy and diseased samples.
“…The NFE exploits the BluePyEfe and the eFEL Python libraries (see section “Methods”) by hiding from the user the technical details of configuration file writing and data management (as required when both tools are run in a standalone manner on a local machine) and exposing a user-friendly point-and-click interface, instead. The models available for the optimization are fetched from the EBRAINS Model Catalog (see Table 1 ), which provides details and links related to the modeling work carried out in the framework of the HBP/EBRAINS research infrastructure as well as results concerning the validation of models against experimental observations ( Sáray et al, 2021 ). A tight integration is also in place between the HHNB and the HPC systems available for job optimization: via the UNICORE Python library or dedicated APIs for the interaction with the CSCS-DAINT system and the NSG, respectively, and using an intuitive web interface, the HHNB seamlessly allows to authenticate to the remote platforms, configure the job execution files and fetch the optimization results.…”
In the last decades, brain modeling has been established as a fundamental tool for understanding neural mechanisms and information processing in individual cells and circuits at different scales of observation. Building data-driven brain models requires the availability of experimental data and analysis tools as well as neural simulation environments and, often, large scale computing facilities. All these components are rarely found in a comprehensive framework and usually require ad hoc programming. To address this, we developed the EBRAINS Hodgkin-Huxley Neuron Builder (HHNB), a web resource for building single cell neural models via the extraction of activity features from electrophysiological traces, the optimization of the model parameters via a genetic algorithm executed on high performance computing facilities and the simulation of the optimized model in an interactive framework. Thanks to its inherent characteristics, the HHNB facilitates the data-driven model building workflow and its reproducibility, hence fostering a collaborative approach to brain modeling.
“…This approach based on feature statistics is now supported by Neuroptimus. To provide access to a diverse array of electrophysiological features, and ensure compatibility with some common workflows [5,7,44,45], Neuroptimus utilizes the Electrophys Feature Extraction Library (eFEL; https://github.com/BlueBrain/eFEL) [27] to characterize the voltage responses of the models. The target data in this case contain the experimental mean and standard deviation values of a predefined set of eFEL features extracted from voltage responses to specific current step inputs, stored in a JSON file created from the recordings using the BluePyEfe tool (https://github.com/BlueBrain/BluePyEfe) and a custom script that converts the output of BluePyEfe to the format expected by Neuroptimus.…”
Finding optimal parameters for detailed neuronal models is a ubiquitous challenge in neuroscientific research. Recently, manual model tuning has been replaced by automated parameter search using a variety of different tools and methods. However, using most of these software tools and choosing the most appropriate algorithm for a given optimization task require substantial technical expertise, which prevents the majority of researchers from using these methods effectively. To address these issues, we developed a generic platform (called Neuroptimus) that allows users to set up neural parameter optimization tasks via a graphical interface, and to solve these tasks using a wide selection of state-of-the-art parameter search methods implemented by five different Python packages. Neuroptimus also offers several features to support more advanced usage, including the ability to run most algorithms in parallel, which allows it to take advantage of high-performance computing architectures. We used the common interface provided by Neuroptimus to conduct a detailed comparison of more than twenty different algorithms (and implementations) on six distinct benchmarks that represent typical scenarios in neuronal parameter search. We quantified the performance of the algorithms in terms of the best solutions found and in terms of convergence speed. We identified several algorithms, including covariance matrix adaptation evolution strategy and particle swarm optimization, that consistently found good solutions in all of our use cases. By contrast, some other algorithms including all local search methods provided good solutions only for the simplest use cases, and failed completely on more complex problems. Finally, we created an online database that allows uploading, querying and analyzing the results of optimization runs performed by Neuroptimus, which enables all researchers to update and extend the current benchmarking study. The tools and analysis we provide should aid members of the neuroscience community to apply parameter search methods more effectively in their research.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.