2019
DOI: 10.26434/chemrxiv.8168354.v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Automatic Retrosynthetic Pathway Planning Using Template-free Models

Abstract: We present an attention-based Transformer model for automatic retrosynthesis route planning. Our approach starts from reactants prediction of single-step organic reactions for given products, followed by Monte Carlo tree search-based automatic retrosynthetic pathway prediction.Trained on two datasets from the United States patent literature, our models achieved a top-1 prediction accuracy of over 54.6% and 63.0% with more than 95% and 99.6% validity rate of SMILES, respectively, which is the best up to now to … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

1
14
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 12 publications
(15 citation statements)
references
References 44 publications
(47 reference statements)
1
14
0
Order By: Relevance
“…Duan et al [37] increased the batch size and the training time for their Transformer model and were able to achieve a top-1 accuracy of 54.1% on the 50k USPTO data set [44]. Later on, the same architecture was reported to have a top-1 accuracy of 43.8% [36], in line with the three previous transformer-based approaches [32,33,35] but significantly lower than the accuracy previously reported by Duan et al [37]. Interestingly, the transformer model was also trained on a proprietary data set [36], including only reactions with two reactants with a Tanimoto similarity distribution peaked at 0.75, characteristic of an excessive degree of similarity (roughly 2 times higher than the USPTO).…”
Section: Introductionsupporting
confidence: 60%
See 4 more Smart Citations
“…Duan et al [37] increased the batch size and the training time for their Transformer model and were able to achieve a top-1 accuracy of 54.1% on the 50k USPTO data set [44]. Later on, the same architecture was reported to have a top-1 accuracy of 43.8% [36], in line with the three previous transformer-based approaches [32,33,35] but significantly lower than the accuracy previously reported by Duan et al [37]. Interestingly, the transformer model was also trained on a proprietary data set [36], including only reactions with two reactants with a Tanimoto similarity distribution peaked at 0.75, characteristic of an excessive degree of similarity (roughly 2 times higher than the USPTO).…”
Section: Introductionsupporting
confidence: 60%
“…While this extensive production of AI models for Organic chemistry was made possible by the availability of public data [28,29], the noise contained in this data and generated by the text-mining extraction process is heavily reducing their potential. In fact, while rule-based systems [30] demonstrated, through wet-lab experiments, the capability to design target molecules with less purification steps and hence, leading to savings in time and cost [31], the AI approaches [6,9,12,16,[32][33][34][35][36][37][38] still have a long way to go.…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations