2018
DOI: 10.48550/arxiv.1812.07069
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

An Atari Model Zoo for Analyzing, Visualizing, and Comparing Deep Reinforcement Learning Agents

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 0 publications
0
4
0
Order By: Relevance
“…Learned Representations We further investigate what type of world model can emerge from an evolutionary process that does not directly optimize for forward prediction or reconstruction loss. To gain insights into the learned representations we employ the t-SNE dimensionality reduction technique (Maaten and Hinton 2008), which has proven valuable for visualizing the inner workings of deep neural networks (Such et al 2018;Mnih et al 2015). We are particularly interested in the information contained in the compressed 32-dimensional vector of the VC and the information stored in the hidden states of the MDN-RNN (which are both fed into the controller that decides on the agent's action).…”
Section: Resultsmentioning
confidence: 99%
“…Learned Representations We further investigate what type of world model can emerge from an evolutionary process that does not directly optimize for forward prediction or reconstruction loss. To gain insights into the learned representations we employ the t-SNE dimensionality reduction technique (Maaten and Hinton 2008), which has proven valuable for visualizing the inner workings of deep neural networks (Such et al 2018;Mnih et al 2015). We are particularly interested in the information contained in the compressed 32-dimensional vector of the VC and the information stored in the hidden states of the MDN-RNN (which are both fed into the controller that decides on the agent's action).…”
Section: Resultsmentioning
confidence: 99%
“…Manipulating the output image of deep generative models is achieved by editing the disentangled hidden features [5,57,66]. On the other hand, visuomotor control tasks can also acquire visual understanding to explain the behavior chosen by agent with visual input through saliency map [59,3,47,12,26,22].…”
Section: Related Workmentioning
confidence: 99%
“…Diverse policy seeking methods The work of Such et al shows that different RL algorithms may converge to different policies for the same task [30]. On the contrary, we are interested in how to learn different policies through a single learning algorithm with the capability of avoiding local optimum.…”
Section: Related Workmentioning
confidence: 99%