2020
DOI: 10.1186/s13321-020-00425-8
|View full text |Cite
|
Sign up to set email alerts
|

GEN: highly efficient SMILES explorer using autodidactic generative examination networks

Abstract: Recurrent neural networks have been widely used to generate millions of de novo molecules in defined chemical spaces. Reported deep generative models are exclusively based on LSTM and/or GRU units and frequently trained using canonical SMILES. In this study, we introduce Generative Examination Networks (GEN) as a new approach to train deep generative networks for SMILES generation. In our GENs, we have used an architecture based on multiple concatenated bidirectional RNN units to enhance the validity of genera… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
20
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 25 publications
(20 citation statements)
references
References 35 publications
(53 reference statements)
0
20
0
Order By: Relevance
“…Even though it is possible to generate novel compounds with the desired properties, the resulting solutions often lack chemical diversity [ 23 – 25 ]. Deursen et al proposed to address this issue with the introduction of Generative Examination Networks (GEN), which perform statistical analysis of the generated compounds during training [ 26 ]. However, their study does did not include the application of this approach in any pre-defined optimization scenario.…”
Section: Introductionmentioning
confidence: 99%
“…Even though it is possible to generate novel compounds with the desired properties, the resulting solutions often lack chemical diversity [ 23 – 25 ]. Deursen et al proposed to address this issue with the introduction of Generative Examination Networks (GEN), which perform statistical analysis of the generated compounds during training [ 26 ]. However, their study does did not include the application of this approach in any pre-defined optimization scenario.…”
Section: Introductionmentioning
confidence: 99%
“…In principle, any kind of deep network might be used for the encoding, and the same or any other kind for the decoding [115]. In this case, the input (encoder) network [27] was mainly a CNN while the output used a specific type of RNN called a gated recurrent unit [116,117]. The latent space used [27] was mainly of 196 dimensions, and the VAE was trained to reproduce its inputs at the outputs (another module from RDKit was used to filter invalid SMILES strings).…”
Section: A Specific Examplementioning
confidence: 99%
“…More recently, it was recognised that various kinds of architectures could, in fact, permit the reversal of this numerical encoding so as to return a molecule (or its SMILES string encoding a unique structure). These are known as generative methods [ 41 , 42 , 43 , 44 , 45 , 46 , 47 , 48 , 49 , 50 , 51 , 52 , 53 , 54 , 55 ], and at heart their aim to generate a suitable and computationally useful representation [ 56 ] of the input data. It is common (but cf.…”
Section: Introductionmentioning
confidence: 99%
“…[ 90 , 91 , 92 , 93 , 94 ]), we use backpropagation to update the network so as to minimise the difference between the predicted and the desired output, subject to any other constraints that we may apply. We also recognise the importance of various forms of regularisation, that are all designed to prevent overfitting [ 49 , 95 , 96 , 97 , 98 ].…”
Section: Introductionmentioning
confidence: 99%