2008
DOI: 10.1002/qsar.200810072
|View full text |Cite
|
Sign up to set email alerts
|

Prediction of hERG Potassium Channel Blockade Using kNN‐QSAR and Local Lazy Regression Methods

Abstract: We have collated hERG inhibition data of 165 compounds from literature and employed two regression procedures, namely, Local Lazy Regression (LLR) and k-Nearest Neighbor (kNN)-QSAR regression methods in combination with Genetic Algorithms (GAs) to select significant and independent molecular descriptors and to build robust predictive models. This methodology helped us to derive four, optimal 2D-and 3D-QSPR models, M1 -M4, based on five descriptors. Extensive validation tests using leave-one-out method and 61 c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
14
0
1

Year Published

2009
2009
2018
2018

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(15 citation statements)
references
References 81 publications
0
14
0
1
Order By: Relevance
“…k ‐Nearest neighbor (kNN) : kNN is a type of instance‐based learning and is thought to be one of the simplest machine learning algorithms. In this case, Manhattan distance was used to calculate the distance matrices, k was set to 28 to reduce the effect of noise, and the final result was an inverse distance weighted average of the k nearest neighbors.…”
Section: Methodsmentioning
confidence: 99%
“…k ‐Nearest neighbor (kNN) : kNN is a type of instance‐based learning and is thought to be one of the simplest machine learning algorithms. In this case, Manhattan distance was used to calculate the distance matrices, k was set to 28 to reduce the effect of noise, and the final result was an inverse distance weighted average of the k nearest neighbors.…”
Section: Methodsmentioning
confidence: 99%
“…This approach described in [12,13] is based on the method of k the nearest neighbors (kNN), which employs clusterization and generation of own regres sion model in each cluster.…”
Section: Introductionmentioning
confidence: 99%
“…36,37 Deciding the optimal number k of nearest neighbors to be used for the identification of the best model is of great relevance and vital for the prediction ability of local modeling. 38 When a prediction is required for a query point, LL proceeds by identifying a set of local model candidates with different polynomial degrees and different number of neighbors, by letting k vary among k min and k max (bandwidth, where k min and k max control the minimal and maximal number of neighbors to be used for identifying and validating models, respectively).…”
Section: Introductionmentioning
confidence: 99%
“…38 When a prediction is required for a query point, LL proceeds by identifying a set of local model candidates with different polynomial degrees and different number of neighbors, by letting k vary among k min and k max (bandwidth, where k min and k max control the minimal and maximal number of neighbors to be used for identifying and validating models, respectively). The prediction ability of each model is commonly assessed through a local leave-one-out cross-validation (LOO-CV) procedure, [36][37][38][39] i.e. the model is built using the k min resultant neighborhood, which is validated using LOO-CV to generate a Q 2 LOO value or prediction error.…”
Section: Introductionmentioning
confidence: 99%