2012
DOI: 10.1186/1471-2164-13-s4-s2
|View full text |Cite
|
Sign up to set email alerts
|

How to evaluate performance of prediction methods? Measures and their interpretation in variation effect analysis

Abstract: BackgroundPrediction methods are increasingly used in biosciences to forecast diverse features and characteristics. Binary two-state classifiers are the most common applications. They are usually based on machine learning approaches. For the end user it is often problematic to evaluate the true performance and applicability of computational tools as some knowledge about computer science and statistics would be needed.ResultsInstructions are given on how to interpret and compare method evaluation results. For s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
245
1
1

Year Published

2012
2012
2022
2022

Publication Types

Select...
8
1

Relationship

2
7

Authors

Journals

citations
Cited by 241 publications
(252 citation statements)
references
References 36 publications
(41 reference statements)
1
245
1
1
Order By: Relevance
“…For quantitative results, some of these zones were manually segmented to obtain ground truth. The detection results are presented in Table 1 using different standard evaluation metrics as described in (Vihinen, 2012). The analysis was conducted with 3D points.…”
Section: Experiments Results and Evaluationmentioning
confidence: 99%
“…For quantitative results, some of these zones were manually segmented to obtain ground truth. The detection results are presented in Table 1 using different standard evaluation metrics as described in (Vihinen, 2012). The analysis was conducted with 3D points.…”
Section: Experiments Results and Evaluationmentioning
confidence: 99%
“…Using these values of Similarity threshold and u threshold along with L = 2 m, e tol = (0.000125 0.000125 0.000125) T (in m 3 ) and nreset = 3 (maximum number of passages possible in our case is 4), we then, evaluated the performance of our method, for all three neighborhoods, using different standard evaluation metrics as described in (Vihinen, 2012) (see Table 2). Although, all these metrics are commonly used to evaluate such algorithms, MCC (Matthews Correlation Coefficient) is regarded as most balanced measure as it is insensitive to different class sizes (like in our application the number of changed cells (changes) is generally quite less as compared to unchanged cells in the urban environment).…”
Section: Results Evaluation and Discussionmentioning
confidence: 99%
“…These studies explicitly reported values for %Acc, BER, and MCC, thereby allowing for comparison with our approach. While we utilize an additional performance metric (PPV) to evaluate our approach, %Acc, BER, and MCC is the largest set of metrics shared by all studies [27]. A comparative assessment of all these methods is shown in Table 3, where we also report the size of the data used (Size of DB) as well as the percentage of the data used for training (%Use) so that an easy comparison of these factors can be made.…”
Section: Comparison Of T-am-al With Other Learning Methodsmentioning
confidence: 99%
“…Overall performance is derived by averaging performance metrics computed over the k splits. [27] The learning fold is used to train the base classifier / learner (A priori or Ripper) in order to learn association rules while tuning takes place over the validation folds. The final association rules learned are run on the (sequestered) test fold partition and results are tabulated.…”
Section: Performance Evaluation and Metricsmentioning
confidence: 99%