2020 3rd IEEE International Conference on Soft Robotics (RoboSoft) 2020
DOI: 10.1109/robosoft48309.2020.9116003
|View full text |Cite
|
Sign up to set email alerts
|

Continuous Control of a Soft Continuum Arm using Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 43 publications
(23 citation statements)
references
References 17 publications
0
23
0
Order By: Relevance
“…The Q -learning controller proposed in this work uses reinforcement learning to control the soft arm and its environment as a whole. You et al (2017), Satheeshbabu et al (2019, 2020), and Wu et al (2020) use reinforcement learning to implement the control of soft arms under influence of unknown environments or hardware failures. By contrast, our proposed method to increase data by generating virtual goals allows the Q -learning controller to be updated quickly, as a result, our Q -learning controller can deal with greater environmental influence and uncertainty.…”
Section: Discussion and Future Workmentioning
confidence: 99%
See 2 more Smart Citations
“…The Q -learning controller proposed in this work uses reinforcement learning to control the soft arm and its environment as a whole. You et al (2017), Satheeshbabu et al (2019, 2020), and Wu et al (2020) use reinforcement learning to implement the control of soft arms under influence of unknown environments or hardware failures. By contrast, our proposed method to increase data by generating virtual goals allows the Q -learning controller to be updated quickly, as a result, our Q -learning controller can deal with greater environmental influence and uncertainty.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…To better deal with this situation, the states need to contain variables representing the absolute position. For example, Satheeshbabu et al (2020) added the current actuation of the arm to the relative states. However, this will greatly increase the size of the state space, which will mean the controller requires more training data.…”
Section: Discussion and Future Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In contrast to Q-learning, SARSA employs the same policy to sample and optimize, which is an onpolicy learning algorithm. In the study by Satheeshbabu et al (2020), the BR 2 soft continuum arm was improved with vision feedback and deep deterministic policy gradient (DDPG), which is a family of actor-critic algorithms. Compared to the previous open-loop control scheme, the closed-loop control method could not only decrease the error obviously but also enable the soft manipulator to track the relatively complex curve path.…”
Section: Reinforcement Learning Without Kinematics/ Dynamics Modelmentioning
confidence: 99%
“…In contrast to policy search reinforcement learning, valuebased methods generate the optimal control policy by optimizing the value function, including SARSA (Ansari et al, 2017b), Q-learning (You et al, 2017;Jiang et al, 2021), DQN (Satheeshbabu et al, 2019;Wu et al, 2020) and its various extensions (e.g., DDQN (You et al, 2019) and Double DQN). The actor-critic approach is a combination of policy-based and value-based reinforcement learning, where the actor executes referring to the policy; thereby the critic calculates the value function to evaluate the actor (Satheeshbabu et al, 2020). Some algorithms (Satheeshbabu et al, 2019;Satheeshbabu et al, 2020;Wu et al, 2020) can be regarded as deep reinforcement learning, which means complex deep neural networks were applied in the control policy, rather than a simple state-action-reward table.…”
Section: Policy-based Vs Value-based Reinforcement Learningmentioning
confidence: 99%