2023
DOI: 10.1007/978-3-031-30333-3_41
|View full text |Cite
|
Sign up to set email alerts
|

Investigating High-Level Decision Making for Automated Driving

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 7 publications
0
0
0
Order By: Relevance
“…Motivated by the shortcomings of other popular policy gradients algorithms, such as Trust Region Policy Optimization (TRPO) [41] and A2C [42], which suffered from training stability issues and slow policy convergence, PPO introduces smaller policy update steps and a clipped objective function to ensure stable, generalized and efficient learning. Thanks to these features and its implementation simplicity, PPO has quickly gained popularity in a wide range of DRL-related research fields, including game-playing and robotic control [43]- [45].…”
Section: Reinforcement Learning and Highway-env Development Environmentmentioning
confidence: 99%
“…Motivated by the shortcomings of other popular policy gradients algorithms, such as Trust Region Policy Optimization (TRPO) [41] and A2C [42], which suffered from training stability issues and slow policy convergence, PPO introduces smaller policy update steps and a clipped objective function to ensure stable, generalized and efficient learning. Thanks to these features and its implementation simplicity, PPO has quickly gained popularity in a wide range of DRL-related research fields, including game-playing and robotic control [43]- [45].…”
Section: Reinforcement Learning and Highway-env Development Environmentmentioning
confidence: 99%