2019 IEEE Conference on Games (CoG) 2019
DOI: 10.1109/cig.2019.8848048
|View full text |Cite
|
Sign up to set email alerts
|

MazeExplorer: A Customisable 3D Benchmark for Assessing Generalisation in Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 4 publications
0
6
0
Order By: Relevance
“…In the authors' opinion, the results obtained in this paper indicate that those techniques could be successfully utilized in order to solve non-trivial problems within such domain, given sufficient learning time. However, due to the high complexity of both of those algorithms and complete randomisation of the quest location generation (proven to impair the reinforcement learning process [66]), the presented solution is far from convergence. Applying this solution to a real-life scenario would yield improved results due to the partially predictable patterns of location distribution-resulting from population density, consumption patterns and more fixed pick-up points.…”
Section: Discussionmentioning
confidence: 99%
“…In the authors' opinion, the results obtained in this paper indicate that those techniques could be successfully utilized in order to solve non-trivial problems within such domain, given sufficient learning time. However, due to the high complexity of both of those algorithms and complete randomisation of the quest location generation (proven to impair the reinforcement learning process [66]), the presented solution is far from convergence. Applying this solution to a real-life scenario would yield improved results due to the partially predictable patterns of location distribution-resulting from population density, consumption patterns and more fixed pick-up points.…”
Section: Discussionmentioning
confidence: 99%
“…More recent work has produced game-like environments with procedurally generated elements, such as the Procgen Benchmark [18], MazeExplorer [30], and the Obstacle Tower environment [38]. However, we argue that, compared to NetHack or Minecraft, these environments do not provide the depth likely necessary to serve as long-term RL testbeds due to limited number of entities and environment interactions that agents have to learn to master.…”
Section: Related Workmentioning
confidence: 91%
“…Experimental setup The experiments are conducted to evaluate the success ratio, sample-efficiency, and generalization of our method, as well as compare our work to competitive baseline methods in multi-target environments. We develop and conduct experiments on (1) visual navigation tasks based on ViZDoom [23,18], and (2) robot arm manipulation tasks based on MuJoCo [42]. These multitarget tasks involve N targets which are placed at positions p 1 , p 2 , • • • , p N , and provide the index of one target as the goal z ∈ {1, 2, • • • , N }.…”
Section: Methodsmentioning
confidence: 99%
“…Reinforcement learning (RL) has been expanding to various fields including robotics, to solve increasingly complex problems. For instance, RL has been gradually mastering skills such as robot arm/hand manipulation on an object [2,48,36,22] and navigation to a target destination [18,39]. However, to benefit humans like the R2-D2 robot in the Star Wars, RL must extend to realistic settings that require interaction with multiple objects or destinations, which is still challenging for RL.…”
Section: Introductionmentioning
confidence: 99%