Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence 2019
DOI: 10.24963/ijcai.2019/452
|View full text |Cite
|
Sign up to set email alerts
|

An Atari Model Zoo for Analyzing, Visualizing, and Comparing Deep Reinforcement Learning Agents

Abstract: Much human and computational effort has aimed to improve how deep reinforcement learning (DRL) algorithms perform on benchmarks such as the Atari Learning Environment. Comparatively less effort has focused on understanding what has been learned by such methods, and investigating and comparing the representations learned by different families of DRL algorithms. Sources of friction include the onerous computational requirements, and general logistical and architectural complications for running DRL algorithms at… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 4 publications
0
3
0
Order By: Relevance
“…In RL settings, in contrast, uncertainty generally arises from a combination of state ambiguity, insufficient exploration, and intrinsic stochasticity 123 , all of which complicate the problem of learning from limited experience. Distributional RL excels in partitioning out this intrinsic uncertainty from other sources, potentially allowing for improvements in state representation 77,78 , exploration [79][80][81][82] , value estimation 124 , model-based learning 125 , off-policy learning 126 , and risk sensitivity [127][128][129][130] .…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…In RL settings, in contrast, uncertainty generally arises from a combination of state ambiguity, insufficient exploration, and intrinsic stochasticity 123 , all of which complicate the problem of learning from limited experience. Distributional RL excels in partitioning out this intrinsic uncertainty from other sources, potentially allowing for improvements in state representation 77,78 , exploration [79][80][81][82] , value estimation 124 , model-based learning 125 , off-policy learning 126 , and risk sensitivity [127][128][129][130] .…”
Section: Discussionmentioning
confidence: 99%
“…In addition to supporting our mechanistic REDRL model, the selective disruption of variance coding by 6-OHDA gives us an experimental tool with which to probe the function of distributional RL in the brain. When paired with deep neural networks, distributional RL is thought to boost system performance mainly by improving state representations 1,4,78 . Due to multiplexing of odor-specific representations alongside distribution information within the striatum (Extended Data Fig.…”
Section: Dopamine Is Necessary For Distributional Rlmentioning
confidence: 99%
See 1 more Smart Citation