2022
DOI: 10.48550/arxiv.2207.03928
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Accelerating Material Design with the Generative Toolkit for Scientific Discovery

Abstract: With the growing availability of data within various scientific domains, generative models hold enormous potential to accelerate scientific discovery at every step of the scientific method. Perhaps their most valuable application lies in the speeding up of what has traditionally been the slowest and most challenging step of coming up with a hypothesis. Powerful representations are now being learned from large volumes of data to generate novel hypotheses, which is making a big impact on scientific discovery app… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 19 publications
0
4
0
Order By: Relevance
“…Moreover, in the Generative Toolkit for Scientific Discovery (GT4SD), we provide an example on leveraging the affinity predictor as a reward function in a protein-driven molecular generative model: .…”
Section: Data and Software Availabilitymentioning
confidence: 99%
“…Moreover, in the Generative Toolkit for Scientific Discovery (GT4SD), we provide an example on leveraging the affinity predictor as a reward function in a protein-driven molecular generative model: .…”
Section: Data and Software Availabilitymentioning
confidence: 99%
“…Additional training epochs did not improve the performance (see Appendix A: Table A7). To train these models, we relied on the Generative Toolkit for Scientific Discovery (GT4SD) library 24 and its LM trainer.…”
Section: Resultsmentioning
confidence: 99%
“…We evaluate the model's performance on five tasks: forward and backward reaction prediction in chemistry, textconditional de novo molecule generation and molecule captioning across domains, and paragraph-to-action in the language domain. The training process is carried out using the language modeling trainer based on Hugging Face transformers (Wolf et al, 2020) and PyTorch Lightning (Falcon and The PyTorch Lightning team, 2019) from the GT4SD library (Manica et al, 2022). To initialize our transformer model, we choose to use the natural language domain, as it has the most available data.…”
Section: Methodsmentioning
confidence: 99%