2020
DOI: 10.3389/frobt.2020.00098
|View full text |Cite
|
Sign up to set email alerts
|

Efficacy of Modern Neuro-Evolutionary Strategies for Continuous Control Optimization

Abstract: We analyze the efficacy of modern neuro-evolutionary strategies for continuous control optimization. Overall, the results collected on a wide variety of qualitatively different benchmark problems indicate that these methods are generally effective and scale well with respect to the number of parameters and the complexity of the problem. Moreover, they are relatively robust with respect to the setting of hyper-parameters. The comparison of the most promising methods indicates that the OpenAI-ES algorithm outper… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
4
2

Relationship

3
7

Authors

Journals

citations
Cited by 31 publications
(36 citation statements)
references
References 30 publications
0
24
0
Order By: Relevance
“…More specifically, the length and the mass of the second pole corresponds to # $ of the length of the first pole (instead of # #: ), and the range of variation of the environmental conditions is much larger than in the classic double-pole balancing problem. These variations increase significantly the complexity of the problem (for more details see Pagliuca, Milano andNolfi, 2018 andPagliuca andNolfi, 2019).…”
Section: The Adaptive Problemsmentioning
confidence: 99%
“…More specifically, the length and the mass of the second pole corresponds to # $ of the length of the first pole (instead of # #: ), and the range of variation of the environmental conditions is much larger than in the classic double-pole balancing problem. These variations increase significantly the complexity of the problem (for more details see Pagliuca, Milano andNolfi, 2018 andPagliuca andNolfi, 2019).…”
Section: The Adaptive Problemsmentioning
confidence: 99%
“…The connection weights and the biases of the control network are trained through the Open-AI evolutionary strategy [ 1 ], i.e. one of the most effective evolutionary methods for continuous problem optimization [ 2 ]. Agents are evaluated for 1 episode in the case of the Walker2DBullet, HalfCheetahBullet, BipedalWalkerHardcore, and MIT racecar and for 5 episodes in the case of the BipedalWalkerHardcore.…”
Section: Experimental Settingmentioning
confidence: 99%
“…Gaming similarly to DRL, gaming represents one of the main testbeds for ES. Most of the literature on ES reviewed in this survey test their algorithms on Atari games [4,5,29,32,256,257,258,259]. These are considered to be challenging as they present the agents with high dimensional visual inputs and a diverse and interesting set of tasks that were designed to be difficult for humans players [14].…”
Section: B Evolution Strategy Applicationsmentioning
confidence: 99%