2019
DOI: 10.26434/chemrxiv.8058464.v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Transformer Model for Retrosynthesis

Abstract: <div><div><div><p>We describe a Transformer model for a retrosynthetic reaction prediction task. The model is trained on 45 033 experimental reaction examples extracted from USA patents. It can successfully predict the reactants set for 42.7% of cases on the external test set. During the training procedure, we applied different learning rate schedules and snapshot learning. These techniques can prevent overfitting and thus can be a reason to get rid of internal validation dataset that i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
117
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 59 publications
(119 citation statements)
references
References 19 publications
(41 reference statements)
2
117
0
Order By: Relevance
“…Other rules for handling most common two-letters elements, charges, and stereochemistry also are used for preparing the input for the neural network. According to our experience, the use of more complicated schemes instead of simple character-level tokenization did not increase the accuracy of models [30]. Therefore a simple character-level tokenization was used in this study.…”
Section: Model Inputmentioning
confidence: 99%
See 3 more Smart Citations
“…Other rules for handling most common two-letters elements, charges, and stereochemistry also are used for preparing the input for the neural network. According to our experience, the use of more complicated schemes instead of simple character-level tokenization did not increase the accuracy of models [30]. Therefore a simple character-level tokenization was used in this study.…”
Section: Model Inputmentioning
confidence: 99%
“…Each pair contained on the left side a non-canonical, and on the right side-a canonical SMILES for the same molecule. Such an arrangement of the training dataset allowed us to reuse the previous Transformer code, which was originally applied for retrosynthetic tasks [30]. For completeness, we added for every compound a line where both left and right sides were identical, i.e.…”
Section: Smiles Canonicalization Model Datasetmentioning
confidence: 99%
See 2 more Smart Citations
“…In contrast to the reaction prediction task in the forward direction, where a defined set of reaction conditions should lead to a single distribution of product molecules, a single-step retrosynthetic prediction takes the form of a one-to-many mapping, where the target could theoretically be made through a variety of different individual reaction steps. Retrosynthesis has seen increased attention from the data science and cheminformatics communities recently with a number of machine learning efforts leveraging reaction templates or rules, [1][2][3][4] techniques adapted from natural language processing, [5][6][7][8] and graph based models. 9,10 However, only the template and rulebased methods are capable of making a connection from the prediction directly back to the source of the template or rule, which is most likely a reaction that was successfully performed in a laboratory.…”
Section: Introductionmentioning
confidence: 99%