2023
DOI: 10.1101/2023.09.04.556234
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Poor Generalization by Current Deep Learning Models for Predicting Binding Affinities of Kinase Inhibitors

Wern Juin Gabriel Ong,
Palani Kirubakaran,
John Karanicolas

Abstract: The extreme surge of interest over the past decade surrounding the use of neural networks has inspired many groups to deploy them for predicting binding affinities of drug-like molecules to their receptors. A model that can accurately make such predictions has the potential to screen large chemical libraries and help streamline the drug discovery process. However, despite reports of models that accurately predict quantitative inhibition using protein kinase sequences and inhibitors' SMILES strings, it is still… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

1
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 63 publications
1
2
0
Order By: Relevance
“…This would ensure a comprehensive evaluation, highlighting the strengths and limitations of each method in various contexts. Randomly splitting compound-kinase pairs into training and validation sets results in overoptimistic performance in terms of generalization to previously unseen data, as was also observed in other work [OKK23].…”
Section: Discussionsupporting
confidence: 66%
“…This would ensure a comprehensive evaluation, highlighting the strengths and limitations of each method in various contexts. Randomly splitting compound-kinase pairs into training and validation sets results in overoptimistic performance in terms of generalization to previously unseen data, as was also observed in other work [OKK23].…”
Section: Discussionsupporting
confidence: 66%
“…To avoid this, all 5 models for a given complex were together placed in either the training set or the validation set or the test set [31]. As noted earlier, all five AF2 models for a given complex non-redundant active complexes with experimentally derived structures were obtained from DockGround, and five AF2 models were built from each of these.…”
Section: Dataset For Training/testing Ppiscreenmlmentioning
confidence: 99%
“…The dataset includes 5 AF2 models for each (active or decoy) complex: training on one AF2 model of a given complex then using a different model of the same complex in the test set would introduce obvious information leakage. To avoid this, all 5 models for a given complex were together placed in either the training set or the validation set or the test set [31]. As noted earlier, all five AF2 models for a given complex were included in the validation and testing sets; however, only those AF2 models of active complexes close to the experimentally derived structures were included in the training set ( Figure S2 ).…”
Section: Introductionmentioning
confidence: 99%