2006
DOI: 10.1021/ci050519k
|View full text |Cite
|
Sign up to set email alerts
|

Assessing Different Classification Methods for Virtual Screening

Abstract: How well do different classification methods perform in selecting the ligands of a protein target out of large compound collections not used to train the model? Support vector machines, random forest, artificial neural networks, k-nearest-neighbor classification with genetic-algorithm-optimized feature selection, trend vectors, naïve Bayesian classification, and decision tree were used to divide databases into molecules predicted to be active and those predicted to be inactive. Training and predicted activitie… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
97
0

Year Published

2006
2006
2014
2014

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 102 publications
(102 citation statements)
references
References 35 publications
(58 reference statements)
5
97
0
Order By: Relevance
“…Up to now, all the computational algorithms have only used single machine learning methods for the analysis and prediction of protein-protein interactions [156][157][158][159][160], or the statistical analysis of interacting patches of protein surfaces [75,149,161,162]. Our experience clearly supports the idea that each machine learning algorithm performs better for selected types of training data [163,164]. Some have very high specificity, others focus more on sensitivity.…”
Section: Resultssupporting
confidence: 56%
“…Up to now, all the computational algorithms have only used single machine learning methods for the analysis and prediction of protein-protein interactions [156][157][158][159][160], or the statistical analysis of interacting patches of protein surfaces [75,149,161,162]. Our experience clearly supports the idea that each machine learning algorithm performs better for selected types of training data [163,164]. Some have very high specificity, others focus more on sensitivity.…”
Section: Resultssupporting
confidence: 56%
“…This technique has already gained recognition as one of the most robust and efficient classifiers (21,(56)(57)(58)63). It can tackle nontrivial problems by projecting the original descriptor vectors to a higher dimensional feature space where a clearer division between the two classes of data becomes feasible.…”
Section: Classification Proceduresmentioning
confidence: 99%
“…We note two such recent studies. Plewczynski et al report the use of support vector machines, random forests, neural networks, k-nearest neighbour classification, trend vectors, naive Bayesian classification and decision trees to categorise sets of ligands for five pharmaceutically important biological targets [79]. Their study focused on the performance of the individual methods but they also investigated how voting might affect performance.…”
Section: Raymond Et Al Have Recently Introduced a New Fusion Rule Cmentioning
confidence: 99%