2017 IEEE International Conference on Robotics and Automation (ICRA) 2017
DOI: 10.1109/icra.2017.7989079
|View full text |Cite
|
Sign up to set email alerts
|

Deep reinforcement learning for tensegrity robot locomotion

Abstract: Tensegrity robots, composed of rigid rods connected by elastic cables, have a number of unique properties that make them appealing for use as planetary exploration rovers. However, control of tensegrity robots remains a difficult problem due to their unusual structures and complex dynamics. In this work, we show how locomotion gaits can be learned automatically using a novel extension of mirror descent guided policy search (MDGPS) applied to periodic locomotion movements, and we demonstrate the effectiveness o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
45
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 73 publications
(49 citation statements)
references
References 26 publications
0
45
0
Order By: Relevance
“…Due to the inherent coupled, nonlinear dynamics of the robot, multi-cable actuation policies render robotic control a challenging intellectual task, providing a launch point for future work. We look forward to exploring the integration of artificial intelligence (particularly evolutionary algorithms and deep reinforcement learning architectures) in this robotic platform to optimize locomotive gaits on varied inclines, and even generate optimal tensegrity topologies, areas which have proven promising in prior work [17], [18]. We hope to leverage learning algorithms to achieve more fluid and efficient locomotion using a robust and fully autonomous control policy.…”
Section: Discussionmentioning
confidence: 99%
“…Due to the inherent coupled, nonlinear dynamics of the robot, multi-cable actuation policies render robotic control a challenging intellectual task, providing a launch point for future work. We look forward to exploring the integration of artificial intelligence (particularly evolutionary algorithms and deep reinforcement learning architectures) in this robotic platform to optimize locomotive gaits on varied inclines, and even generate optimal tensegrity topologies, areas which have proven promising in prior work [17], [18]. We hope to leverage learning algorithms to achieve more fluid and efficient locomotion using a robust and fully autonomous control policy.…”
Section: Discussionmentioning
confidence: 99%
“…Model-based closed-loop control has been mostly limited to low-dimensional structures [23], [24], [25], [13], [26]. More complex and high dimensional systems have been addressed with model-free methods [16], [22], [18], [27], [28], [29] or open-loop control [30], [31], [32], [20]. In order to use a tensegrity spine with Laika, a model-based closed-loop tracking controller was developed by the authors in [12] and is improved upon in this work.…”
Section: A Tensegrity Robots and Controlmentioning
confidence: 99%
“…(27)(28)(29), in the context of robotics and control systems, did not address these rank issues when implemented for this 2D spine. Reducing (27)(28)(29) to a cablesonly formulation by optimizing only over q s as suggested in [20] only exacerbates these rank issues by defining A with fewer columns. Additionally, the relaxation of this equalityconstrained problem to an inequality-constrained formulation, as used in [20], did not make the problem feasible.…”
Section: Existence Of Solutions and Rank Deficiencymentioning
confidence: 99%
See 1 more Smart Citation
“…Further details are presented in Alg. 2 2) Adversarial transfer of encoder from sim-to-real: Once we have a policy that is performing well in the simulator, we aim to learn an encoder that generates the same distribution of latent states over real images as the pre-trained encoder. To achieve this we begin by freezing the source encoder's learned weights.…”
Section: B Policy Transfer To the Real Robotmentioning
confidence: 99%