2022
DOI: 10.48550/arxiv.2205.14691
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On the Robustness of Safe Reinforcement Learning under Observational Perturbations

Abstract: Safe reinforcement learning (RL) trains a policy to maximize the task reward while satisfying safety constraints. While prior works focus on the performance optimality, we find that the optimal solutions of many safe RL problems are not robust and safe against carefully designed observational perturbations. We formally analyze the unique properties of designing effective state adversarial attackers in the safe RL setting. We show that baseline adversarial attack techniques for standard RL tasks are not always … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
10
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
3

Relationship

2
4

Authors

Journals

citations
Cited by 6 publications
(12 citation statements)
references
References 36 publications
1
10
0
Order By: Relevance
“…Similarly, the PPO agent that achieves the best route completion (Comp) score presents, however, the highest RR and SS scores, which means that it may run red lights and stop signs most frequently. This observation suggests the inherent contradiction between some safety metrics and functionality metrics, which is also unveiled in some previous studies [46,38,39].…”
Section: Benchmark Resultssupporting
confidence: 52%
“…Similarly, the PPO agent that achieves the best route completion (Comp) score presents, however, the highest RR and SS scores, which means that it may run red lights and stop signs most frequently. This observation suggests the inherent contradiction between some safety metrics and functionality metrics, which is also unveiled in some previous studies [46,38,39].…”
Section: Benchmark Resultssupporting
confidence: 52%
“…Beyond robotics continuous control tasks and simulated games, robust RL is also tested in mobile robot tasks and autonomous driving scenarios. Liu et al [108] propose a safe and robust benchmark containing mobile robot tasks based on Bullet safety gym [60] environments. Jaafra et al [77] propose to test in CARLA simulator [41] with different conditions, including the traffic density, such as the number of dynamic objects, and visual effects such as weather and lightening conditions.…”
Section: Application Benchmarks and Resourcesmentioning
confidence: 99%
“…To better describe the unique properties of a safe RL problem, we provide the feasibility, optimality, and temptation definitions following the previous work [108]. Their figure illustrations for one CMDP are presented in Fig.…”
Section: Problem Formulation Of Safe Reinforcement Learningmentioning
confidence: 99%
See 2 more Smart Citations