2019
DOI: 10.1007/978-3-030-35699-6_1
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Run Faster in a Humanoid Robot Soccer Environment Through Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
30
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
1
1

Relationship

3
4

Authors

Journals

citations
Cited by 33 publications
(32 citation statements)
references
References 13 publications
2
30
0
Order By: Relevance
“…This is possible since in a kick scenario, moving the head joints brings little Early attempts tried to directly relate the values of the action space to the angular velocities, using variations on a simple proportional or a complete PID controller. Nevertheless, as also verified in previous work [12], a noticeable improvement is obtained through interpreting each term of the action space as relating to an angle, and from there deriving an angular velocity to feed as an input. In the end, the desired objective angle θ goal , can be obtained from the action space output as: With θ goal now defined we can now obtain from it the angular velocity as:…”
Section: Action Spacesupporting
confidence: 73%
See 3 more Smart Citations
“…This is possible since in a kick scenario, moving the head joints brings little Early attempts tried to directly relate the values of the action space to the angular velocities, using variations on a simple proportional or a complete PID controller. Nevertheless, as also verified in previous work [12], a noticeable improvement is obtained through interpreting each term of the action space as relating to an angle, and from there deriving an angular velocity to feed as an input. In the end, the desired objective angle θ goal , can be obtained from the action space output as: With θ goal now defined we can now obtain from it the angular velocity as:…”
Section: Action Spacesupporting
confidence: 73%
“…This forward motion corresponds to one of several skills developed for usage in a game environment. These correspond to a sprinting and running pace [12].…”
Section: Motivationmentioning
confidence: 99%
See 2 more Smart Citations
“…Each oscillator has a set of parameters that are generally tuned by some trail-intensive method like trials and error, machine learning (ML) algorithms or both of them [25,33]. Some other approaches in this class are designed based on learning from scratch and mostly are based on Reinforcement Learning (RL) algorithms [1,19] which need many samples to be able to generate walking that takes a considerable amount of time. Unlike the model-free approaches, the fundamental core of the model-based approach is a dynamics model of the robot.…”
Section: Introductionmentioning
confidence: 99%