2019
DOI: 10.1007/978-3-030-29135-8_4
|View full text |Cite
|
Sign up to set email alerts
|

Artificial Intelligence for Prosthetics: Challenge Solutions

Abstract: In the NeurIPS 2018 Artificial Intelligence for Prosthetics challenge, participants were tasked with building a controller for a musculoskeletal model with a goal of matching a given time-varying velocity vector. Top participants were invited to describe their algorithms. In this work, we describe the challenge and present thirteen solutions that used deep reinforcement learning approaches. Many solutions use similar relaxations and heuristics, such as reward shaping, frame skipping, discretization of the acti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
26
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 28 publications
(26 citation statements)
references
References 37 publications
0
26
0
Order By: Relevance
“…Various RL techniques have been effectively used since the first competition [124,125], including frame skipping, discretization of the action space, and reward shaping. These are practical techniques that constrain the problem in certain ways to encourage an agent to search successful regions faster in the initial stages of training.…”
Section: Top Solutions and Resultsmentioning
confidence: 99%
“…Various RL techniques have been effectively used since the first competition [124,125], including frame skipping, discretization of the action space, and reward shaping. These are practical techniques that constrain the problem in certain ways to encourage an agent to search successful regions faster in the initial stages of training.…”
Section: Top Solutions and Resultsmentioning
confidence: 99%
“…Moco attempts to produce well-scaled optimization problems to facilitate rapid convergence (e.g., using normalized tendon force as a state variable), but providing automated problem scaling ( [20], section 4.8) could further improve convergence. Lastly, the direct collocation method itself is not well-suited to simulations that are interactive (e.g., users perturbing the model) or include randomness (e.g., uneven terrain); such problems are better suited to single shooting [12] or reinforcement learning approaches [77].…”
Section: Availability and Future Directionsmentioning
confidence: 99%
“…In the NeurIPS 2018 Artificial Intelligence for Prosthetics challenge, participants were asked to build a controller for a transtibial amputee model with the goal of moving it forward, and were encouraged to use DRL. Kidzinski et al, [8] use a musculoskeletal model of 19 muscles with one leg having the below-knee leg replaced by a prosthesis. Results have shown that DRL can find solutions in which the agent learns a policy to efficiently move forward.…”
Section: Transfemoral Prostheses and Drlmentioning
confidence: 99%
“…A state-of-the-art DRL algorithm (PPO) is implemented and a DRL algorithm (PPO with imitation learning) is proposed and implemented on a musculoskeletal model of a healthy subject in the open-source simulation software OpenSim [7]. This paper uses the model presented in [8], which consists of two healthy legs including 18 healthy muscles (9 per leg) to control 10 degrees of freedom. The DRL algorithms (PPO and PPO with imitation learning) are validated on a public data-set [9].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation