2010
DOI: 10.1093/bioinformatics/btq037
|View full text |Cite
|
Sign up to set email alerts
|

Small-sample precision of ROC-related estimates

Abstract: edward@mail.ece.tamu.edu.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
143
0

Year Published

2010
2010
2022
2022

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 268 publications
(153 citation statements)
references
References 16 publications
2
143
0
Order By: Relevance
“…Hanczar et al [8] show that the variance in ROC curves computed over small data sets can significantly impact scientific conclusions. Bouckaert and Frank [9] study the consistency of statistical tests on individual data sets and recommend a corrected t-test [10] across ten iterations of ten-fold cross-validation as the least sensitive to the order of the data.…”
Section: Introductionmentioning
confidence: 99%
“…Hanczar et al [8] show that the variance in ROC curves computed over small data sets can significantly impact scientific conclusions. Bouckaert and Frank [9] study the consistency of statistical tests on individual data sets and recommend a corrected t-test [10] across ten iterations of ten-fold cross-validation as the least sensitive to the order of the data.…”
Section: Introductionmentioning
confidence: 99%
“…For C4.5, the optimised parameter was the confidence factor (CF) that configures its error-based pruning method, varying it within its boundary values [0, 0.5], in steps of 0.05. For CART, the optimised parameter was the number of folds in the crossvalidation procedure that is executed within the cost-complexity pruning method (ranging within [2,20]). For REPTree, the optimised parameter was the size of the 1.00…”
Section: Methodsmentioning
confidence: 99%
“…The machine learning community most often uses the AUC statistic for model comparison, even though this practice has recently been questioned based upon new research that shows that AUC is quite noisy as a performance measure for classification [20] and has some other significant problems in model comparison [21,23].…”
Section: Aucmentioning
confidence: 99%
“…ROC analysis is now an integral part of the evaluation of machine learning algorithms (Bradley, 1997). Whereas ROC curves are widely (and rightly so) considered useful, both theoretical and practical shortcomings of the AUC have been pointed out (Hilden, 1991;Adams & Hand, 1999;Bengio, Mariéthoz, & Keller, 2005;Webb & Ting, 2005;Lobo et al, 2008;Hand, 2009;Hanczar, Hua, Sima, Weinstein, Bittner, & Dougherty, 2010;Hand & Anagnostopoulos, 2013;Parker, 2013). A particular problem of the AUC is that it can be incoherent, in the sense that it assumes different cost distributions for different classifiers (Hand, 2009).…”
Section: Area Under the Roc Curve (Auc)mentioning
confidence: 99%