2011
DOI: 10.2478/s13531-011-0022-9
|View full text |Cite
|
Sign up to set email alerts
|

Assessing the quality of classification models: Performance measures and evaluation procedures

Abstract: This article systematically reviews techniques used for the evaluation of classification models and provides guidelines for their proper application. This includes performance measures assessing the model’s performance on a particular dataset and evaluation procedures applying the former to appropriately selected data subsets to produce estimates of their expected values on new data. Their common purpose is to assess model generalization capabilities, which are crucial for judging the applicability and usefuln… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0
2

Year Published

2012
2012
2023
2023

Publication Types

Select...
5
1
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(12 citation statements)
references
References 7 publications
0
10
0
2
Order By: Relevance
“…Recall = TP TP + FN (7) In this way, the arithmetic mean of all F1 score values obtained class-wise (denominated as the macro-averaged F1 score) provides a key figure of merit to compare different indicators + FE + classifier configurations [33,34]. Therefore, Figure 7 shows the results of the macro F1 score, where each boxplot is generated from the 100 iterations performed for each configuration.…”
Section: Discussionmentioning
confidence: 99%
“…Recall = TP TP + FN (7) In this way, the arithmetic mean of all F1 score values obtained class-wise (denominated as the macro-averaged F1 score) provides a key figure of merit to compare different indicators + FE + classifier configurations [33,34]. Therefore, Figure 7 shows the results of the macro F1 score, where each boxplot is generated from the 100 iterations performed for each configuration.…”
Section: Discussionmentioning
confidence: 99%
“…A 10–fold cross–validation was used. This approach has the advantage of using all available data, while balancing the tradeoff of bias and variance [ 17 , 18 ]. At each fold, ca.…”
Section: Methodsmentioning
confidence: 99%
“…Figure 12 shows a confusion matrix for DGA multiclass classification. The confusion matrix allows us to deduce several other evaluation parameters such as the classification accuracy, classification error, sensitivity, precision and the f1-score [70][71][72][73][74][75].…”
Section: Confusion Matrix (Cm)mentioning
confidence: 99%
“…The classification accuracy gives us a measure of how often the classifier is correct [70][71][72][73][74][75][76]. Equation (7) gives the formula for calculating the classification accuracy.…”
Section: Classification Accuracymentioning
confidence: 99%