2022
DOI: 10.48550/arxiv.2203.14912
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Advanced Skills through Multiple Adversarial Motion Priors in Reinforcement Learning

Abstract: In recent years, reinforcement learning (RL) has shown outstanding performance for locomotion control of highly articulated robotic systems. Such approaches typically involve tedious reward function tuning to achieve the desired motion style. Imitation learning approaches such as adversarial motion priors aim to reduce this problem by encouraging a pre-defined motion style. In this work, we present an approach to augment the concept of adversarial motion prior-based RL to allow for multiple, discretely switcha… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(16 citation statements)
references
References 12 publications
0
16
0
Order By: Relevance
“…This allows the learning agent to execute tasks that may not be portrayed in the original demonstrations. To enable active style control, Multi-AMP allows for the switching of multiple different style rewards by training multiple discriminators encoding different reference motions in parallel [39].…”
Section: Related Workmentioning
confidence: 99%
“…This allows the learning agent to execute tasks that may not be portrayed in the original demonstrations. To enable active style control, Multi-AMP allows for the switching of multiple different style rewards by training multiple discriminators encoding different reference motions in parallel [39].…”
Section: Related Workmentioning
confidence: 99%
“…It can also build large-scale realistic complex scenes, and its underlying PhysX engine can accurately and realistically model and simulate the motion of objects. Therefore, more researchers have begun to use Isaac Gym as the implementation and verification platform of DRL algorithm [35][36][37][38] .…”
Section: Simulatormentioning
confidence: 99%
“…Lower-cost hardware platforms allow DRL algorithms to be more widely used. More recently, a wheel-legged quadruped robot [38] demonstrated skills learned from existing DRL controllers and trajectory optimization, such as ducking and walking, and new skills, such as switching between quadrupedal and humanoid configurations.…”
Section: Hardware Platformmentioning
confidence: 99%
See 2 more Smart Citations