2023
DOI: 10.1186/s13040-023-00322-4
|View full text |Cite
|
Sign up to set email alerts
|

The Matthews correlation coefficient (MCC) should replace the ROC AUC as the standard metric for assessing binary classification

Abstract: Binary classification is a common task for which machine learning and computational statistics are used, and the area under the receiver operating characteristic curve (ROC AUC) has become the common standard metric to evaluate binary classifications in most scientific fields. The ROC curve has true positive rate (also called sensitivity or recall) on the y axis and false positive rate on the x axis, and the ROC AUC can range from 0 (worst result) to 1 (perfect result). The ROC AUC, however, has several flaws … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
37
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 113 publications
(67 citation statements)
references
References 66 publications
0
37
0
Order By: Relevance
“…Generally, all annotators showed high correlation [ 20 ] to “gold standard” annotations of CXR text reports ( Table 2 a,b). This finding was comparable to a previous study which showed a similar level of agreement between radiologists and non-radiological physicians and medical students when reading and comprehending radiology reports [ 26 ].…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Generally, all annotators showed high correlation [ 20 ] to “gold standard” annotations of CXR text reports ( Table 2 a,b). This finding was comparable to a previous study which showed a similar level of agreement between radiologists and non-radiological physicians and medical students when reading and comprehending radiology reports [ 26 ].…”
Section: Discussionmentioning
confidence: 99%
“…Matthew’s correlation coefficient (MCC) [ 20 ] was used to compare annotator performance to “gold standard” labeling and to compare annotators’ performance to each other. The MCC was based on values selected for a 2 × 2 confusion matrix ( Table 1 ) where true positive (TP) described the number of labels that matched “gold standard” labels for all positive and negative findings separately.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Matthews Correlation Coefficient is used as a single metric for direct comparison between the developed models. Because MCC is known to be superior to any other performance metric for ranking the binary classification models …”
Section: Methodsmentioning
confidence: 99%
“…Because MCC is known to be superior to any other performance metric for ranking the binary classification models. 25…”
Section: Machine Learning Algorithmsmentioning
confidence: 99%