2022
DOI: 10.5194/hess-2022-380
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

When best is the enemy of good – critical evaluation of performance criteria in hydrological models

Abstract: Abstract. Performance criteria play a key role in the calibration and evaluation of hydrological models and have been extensively developed and studied, but some of the most used criteria still have unknown pitfalls. This study set out to examine counterbalancing errors, which are inherent to the Kling-Gupta Efficiency (KGE) and its variants. A total of nine performance criteria – including the KGE and its variants, as well as the Nash-Sutcliffe Efficiency (NSE) and the refined version of the Willmott’s index … Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 34 publications
0
2
0
Order By: Relevance
“…Overall, the overestimation in autumn and the underestimation in spring can still lead to an adequate Qmean over the entire time period. This and other counterbalancing errors (Cinkus et al., 2023) are one reason why KGE is unlikely to lead to adequate model structures that capture relevant hydrological mechanisms in a catchment.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Overall, the overestimation in autumn and the underestimation in spring can still lead to an adequate Qmean over the entire time period. This and other counterbalancing errors (Cinkus et al., 2023) are one reason why KGE is unlikely to lead to adequate model structures that capture relevant hydrological mechanisms in a catchment.…”
Section: Discussionmentioning
confidence: 99%
“…Thus, it may be a sign of an adequate model choice or simply luck if a KGE calibrated model manages to reproduce a more nuanced understanding of hydrological processes as described for example, through the hydrological signatures tested. This is another reason why aggregated metrics have been criticized (Cinkus et al., 2023; Clark et al., 2021) and the calls for using additional methods (e.g., Bouaziz et al., 2021; Knoben et al., 2020; Pool et al., 2017) or better metrics (Fowler et al., 2018; Pool et al., 2018; Schwemmle et al., 2021) for evaluating model performance accumulate.…”
Section: Discussionmentioning
confidence: 99%
“…Previous studies (Ambroise et al, 1995;Refsgaard, 1997) have highlighted that calibrating a distributed hydrological model solely against single-point hydrological parameters may not yield satisfactory results for the entire catchment. The criteria to calibrate and evaluate the hydrological models play an important role in its performance (Cinkus et al, 2022). To achieve better performance, it is recommended to employ multiple variables and multiple site calibration strategies in distributed hydrological modelling.…”
Section: Calibration and Validationmentioning
confidence: 99%
“…Ignoring these assumptions and limitations can lead to contradictory results and confusion about model evaluation (Bennett et al, 2013; Castaneda‐Gonzalez et al, 2018; Criss & Winston, 2008; Knoben et al, 2019; Koskinen et al, 2017). Although several modifications regarding the existing metrics of R 2 (Legates & McCabe Jr, 1999; Onyutha, 2022), NSE (Criss & Winston, 2008; Duc & Sawada, 2023; Mathevet et al, 2006), and KGE (Cinkus et al, 2023; Kling et al, 2012; Lamontagne et al, 2020; Liu, 2020; Pool et al, 2018) have been proposed, these revised versions have not been widely accepted and there is still no broad consensus on how to evaluate the performance of hydrologic and hydraulic models by using an appropriate criterion given the availability and accuracy of observed hydrologic data, epistemic uncertainty in the modeling process (Beven & Lane, 2022; Clark et al, 2021; Huang & Merwade, 2023b; Knoben et al, 2019). Therefore, in order to evaluate the reliability and accuracy of flood model predictions, the pros and cons of multiple commonly used evaluation metrics for ensemble flood modeling are investigated and demonstrated in this study.…”
Section: Introductionmentioning
confidence: 99%