2013
DOI: 10.1016/j.neucom.2012.07.003
|View full text |Cite
|
Sign up to set email alerts
|

GFO: A data driven approach for optimizing the Gaussian function based similarity metric in computational biology

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
7
0

Year Published

2013
2013
2018
2018

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 13 publications
(7 citation statements)
references
References 34 publications
0
7
0
Order By: Relevance
“…There are two important parameters, c and γ, that require optimization: c represents the cost factor controlling the trade-off for maximizing the margin and minimizing the error rate, and γ regulates model generalization. In order to optimize these parameters during model training, the grid-search strategy and the GFO 98 algorithm were used, as well as five-fold cross-validation on the training dataset, to fully optimize the model performance. In addition, we adopted the following strategies to avoid potential overfitting problems:

At each cross-validation step, addition of each of the features to train the model was achieved by using four folds of the dataset, validating the performance of the trained model on the singled-out fold of the dataset.

…”
Section: Methodsmentioning
confidence: 99%
“…There are two important parameters, c and γ, that require optimization: c represents the cost factor controlling the trade-off for maximizing the margin and minimizing the error rate, and γ regulates model generalization. In order to optimize these parameters during model training, the grid-search strategy and the GFO 98 algorithm were used, as well as five-fold cross-validation on the training dataset, to fully optimize the model performance. In addition, we adopted the following strategies to avoid potential overfitting problems:

At each cross-validation step, addition of each of the features to train the model was achieved by using four folds of the dataset, validating the performance of the trained model on the singled-out fold of the dataset.

…”
Section: Methodsmentioning
confidence: 99%
“…The range values for C, γ , and ε were taken from the LIBSVM grid search [31] and extended to suit all methods assessed in this study. We also considered using a data-driven approach for optimising the kernel width parameter ( γ ) [32], however, for the relatively small size of our data set, the grid search was a sufficient solution.…”
Section: Methodsmentioning
confidence: 99%
“…The range values for C, γ , and ε were taken from the LIBSVM grid search [ 31 ] and extended to suit all methods assessed in this study. We also considered using a data-driven approach for optimising the kernel width parameter ( γ ) [ 32 ], however, for the relatively small size of our data set, the grid search was a sufficient solution.…”
Section: Methodsmentioning
confidence: 99%