2021
DOI: 10.48550/arxiv.2104.10592
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Robust Biped Locomotion Using Deep Reinforcement Learning on Top of an Analytical Control Approach

Mohammadreza Kasaei,
Miguel Abreu,
Nuno Lau
et al.

Abstract: This paper proposes a modular framework to generate robust biped locomotion using a tight coupling between an analytical walking approach and deep reinforcement learning. This framework is composed of six main modules which are hierarchically connected to reduce the overall complexity and increase its flexibility. The core of this framework is a specific dynamics model which abstracts a humanoid's dynamics model into two masses for modeling upper and lower body. This dynamics model is used to design an adaptiv… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 44 publications
0
3
0
Order By: Relevance
“…In addition to learning skills from scratch, reinforcement learning algorithms are also used to optimize existing skills. Kasaei et al [102] used a combination of reinforcement learning and robot dynamics models to generate and optimize the omnidirectional walking skill of robots. This study used the ZMP as the main criterion for robot stability and designed a linear inverted pendulum model considering the motion of the center of mass.…”
Section: Reinforcement Learning Methodsmentioning
confidence: 99%
“…In addition to learning skills from scratch, reinforcement learning algorithms are also used to optimize existing skills. Kasaei et al [102] used a combination of reinforcement learning and robot dynamics models to generate and optimize the omnidirectional walking skill of robots. This study used the ZMP as the main criterion for robot stability and designed a linear inverted pendulum model considering the motion of the center of mass.…”
Section: Reinforcement Learning Methodsmentioning
confidence: 99%
“…The reference trajectory of the robot can be used to regulate robot behavior and narrow the exploration space of the RL-based policy. While many prior approaches use dynamics models to generate physically feasible reference trajectories for bipedal locomotion [11,12], using fixed reference trajectories proved to be feasible for quadrupedal locomotion [4] as the control policy is capable of modifying the trajectory to stabilize the robot. In our framework, we use a reference trajectory generator to build trajectory, regardless of the robot's current state.…”
Section: Control Frameworkmentioning
confidence: 99%
“…Reinforcement learning has been employed to control low-speed quadrupedal walking on uneven terrains [4][5][6][7], or high-speed running on flat ground [8,9]. RL is also implemented in bipedal locomotion controllers, resulting in a better performance regarding dynamic model uncertainty [10], fast yet robust maneuvers [11,12], and the precise control of stepping point [13] or gait [14].…”
Section: Introduction 1backgroundmentioning
confidence: 99%