2021
DOI: 10.48550/arxiv.2109.07827
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Enabling risk-aware Reinforcement Learning for medical interventions through uncertainty decomposition

Abstract: Reinforcement Learning (RL) is emerging as tool for tackling complex control and decision-making problems. However, in high-risk environments such as healthcare, manufacturing, automotive or aerospace, it is often challenging to bridge the gap between an apparently optimal policy learned by an agent and its real-world deployment, due to the uncertainties and risk associated with it. Broadly speaking RL agents face two kinds of uncertainty, 1. aleatoric uncertainty, which reflects randomness or noise in the dyn… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 16 publications
0
3
0
Order By: Relevance
“…An emerging new avenue in the field is to augment AI models so that they can quantify their own confidence or uncertainty over their recommendations. 19 Going forward, it may be helpful to algorithmically combine the communication of uncertainty that a system has about itself, which reflects the risk of unwanted behaviour as we have shown in other domains of risk-aware control by medical devices, 20 with its safety features, that we have shown here.…”
Section: Discussionmentioning
confidence: 84%
“…An emerging new avenue in the field is to augment AI models so that they can quantify their own confidence or uncertainty over their recommendations. 19 Going forward, it may be helpful to algorithmically combine the communication of uncertainty that a system has about itself, which reflects the risk of unwanted behaviour as we have shown in other domains of risk-aware control by medical devices, 20 with its safety features, that we have shown here.…”
Section: Discussionmentioning
confidence: 84%
“…, (x L−1 , u L−1 )} generated by following the clinician's policy H ∼ µ CP . Approaches for off-policy evaluation are an active topic of research in the context of related reinforcement learning frameworks (Hanna et al, 2017;Thomas et al, 2015;Festor et al, 2021). It should be noted that the factors driving µ CP may differ from the cost function chosen to derive µ OP , which is why effective calibration is important to objectively compare the two policies.…”
Section: Policy Cost Estimationmentioning
confidence: 99%
“…Recently, there has been an increased volume of research which try to learn optimal treatment strategies for critically ill and in particular for septic patients (Komorowski et al, 2018;Chen et al, 2019;Raghu et al, 2017;Li et al, 2019;Peng et al, 2018;Festor et al, 2021;Nanayakkara et al, 2022b), using Reinforcement Learning (RL) methods. Given the enormous mortality, morbidity and economic burden (Liu et al, 2014;Rhee et al, 2017;Paoli et al, 2018), the ambiguity regarding optimal treatment strategies and lack of accepted guidelines for treatment (Marik, 2015;Jarczak et al, 2021), such attempts are certainly justified.…”
Section: Introductionmentioning
confidence: 99%