2017
DOI: 10.1093/jamia/ocx030
|View full text |Cite
|
Sign up to set email alerts
|

Calibration drift in regression and machine learning models for acute kidney injury

Abstract: Efficient and effective updating protocols will be essential for maintaining accuracy of, user confidence in, and safety of personalized risk predictions to support decision-making. Model updating protocols should be tailored to account for variations in calibration drift across methods and respond to periods of rapid performance drift rather than be limited to regularly scheduled annual or biannual intervals.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
177
0
2

Year Published

2017
2017
2021
2021

Publication Types

Select...
9
1

Relationship

0
10

Authors

Journals

citations
Cited by 203 publications
(185 citation statements)
references
References 64 publications
0
177
0
2
Order By: Relevance
“…Although study sizes have increased over time (among studies published in 2017 and 2018, the median [IQR] sample size was 3464 [286-21,498]), studies based on fewer than 1000 patients have been regularly published in recent years. Six studies used data on more than 100,000 patients: one in 2015, one in 2017, and four in 2018 [2,[12][13][14][15][16].…”
Section: Type Of Machine Learningmentioning
confidence: 99%
“…Although study sizes have increased over time (among studies published in 2017 and 2018, the median [IQR] sample size was 3464 [286-21,498]), studies based on fewer than 1000 patients have been regularly published in recent years. Six studies used data on more than 100,000 patients: one in 2015, one in 2017, and four in 2018 [2,[12][13][14][15][16].…”
Section: Type Of Machine Learningmentioning
confidence: 99%
“…Demonstrating utility or validity on retrospective, in silico settings does not guarantee that the product will perform well in a different setting 57 or in a different time period. 86,87 The utility and validity of the product must be reassessed across time (Step 2b) and space (Step 2d). Evaluating a machine learning product on a hold-out and temporal validation set (Step 2a) is recommended before integrating a product into clinical care.…”
Section: Evaluate and Validatementioning
confidence: 99%
“…In other domains, like acute kidney injury prediction, for example, calibration has been shown to worsen over time. [50] Event rate shift alone was a primary reason for such drift, so practical frameworks incorporating these shifts in program evaluation are important. Our findings underline the cross-sectional impacts of event rate differences not only in calibration but in usefulness and expected utility analyses.…”
Section: Discussionmentioning
confidence: 99%