2015
DOI: 10.1371/journal.pone.0118432
|View full text |Cite
|
Sign up to set email alerts
|

The Precision-Recall Plot Is More Informative than the ROC Plot When Evaluating Binary Classifiers on Imbalanced Datasets

Abstract: Binary classifiers are routinely evaluated with performance measures such as sensitivity and specificity, and performance is frequently illustrated with Receiver Operating Characteristics (ROC) plots. Alternative measures such as positive predictive value (PPV) and the associated Precision/Recall (PRC) plots are used less frequently. Many bioinformatics studies develop and evaluate classifiers that are to be applied to strongly imbalanced datasets in which the number of negatives outweighs the number of positi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

29
1,943
0
16

Year Published

2016
2016
2024
2024

Publication Types

Select...
9
1

Relationship

0
10

Authors

Journals

citations
Cited by 2,873 publications
(2,154 citation statements)
references
References 47 publications
29
1,943
0
16
Order By: Relevance
“…Namely, such setting adopts the k-fold cross-validation (CV) procedure (k = 5) to assess the generalization abilities of the compared methods, and the Area Under the ROC Curve (AUC) and the Precision at different Recall levels (PXR) to measure the corresponding performance. Furthermore, as done by authors of NWGP on the same data, we also computed for our method the Area Under the Precision-Recall Curve (AUPRC), being AUPRC more informative than AUC on unbalanced settings [68].…”
Section: Resultsmentioning
confidence: 99%
“…Namely, such setting adopts the k-fold cross-validation (CV) procedure (k = 5) to assess the generalization abilities of the compared methods, and the Area Under the ROC Curve (AUC) and the Precision at different Recall levels (PXR) to measure the corresponding performance. Furthermore, as done by authors of NWGP on the same data, we also computed for our method the Area Under the Precision-Recall Curve (AUPRC), being AUPRC more informative than AUC on unbalanced settings [68].…”
Section: Resultsmentioning
confidence: 99%
“…These results are also confirmed by the AUROC 50 , AUROC 100 , and AUROC 1000 results, where hyperSMURF with tuned parameters largely and significantly outperforms hyperSMURF with default parameters ( Table 1). Note that we do not report AUROC results, since in this highly imbalanced context pure AUROC results are not as significant as AUPRC or AUROC limited to the top ranked SNVs [Saito and Rehmsmeier, 2015]. Table 1: Comparison of hyperSMURF results obtained respectively with default parameters (n = 100, f = 2, m = 3) and with the best parameters obtained by internal cross-validation on the training data (n = 300, f = 1, m = 10).…”
Section: Resultsmentioning
confidence: 99%
“…The sample genotypes were analyzed using the ΔΔCt [22] method in accordance with the description [8], the data was processed with the SPSS statistics package [23] using the ROCanalysis [24]. The use of other binary classifiers [25] and statistical methods [18] to increase the reliability of results can be promising in such studies.…”
Section: Introductionmentioning
confidence: 99%