2020
DOI: 10.3390/app10249013
|View full text |Cite
|
Sign up to set email alerts
|

A Parametric Study of a Deep Reinforcement Learning Control System Applied to the Swing-Up Problem of the Cart-Pole

Abstract: In this investigation, the nonlinear swing-up problem associated with the cart-pole system modeled as a multibody dynamical system is solved by developing a deep Reinforcement Learning (RL) controller. Furthermore, the sensitivity analysis of the deep RL controller applied to the cart-pole swing-up problem is carried out. To this end, the influence of modifying the physical properties of the system and the presence of dry friction forces are analyzed employing the cumulative reward during the task. Extreme lim… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0
1

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 47 publications
(7 citation statements)
references
References 67 publications
0
6
0
1
Order By: Relevance
“…9(a). 31,32) The weight coefficient was fixed at +1 and the bit width of data was fixed at 8. Since this was a fulldigital implementation, the entire neural net was written in HDL, and the layout could be created only by synthesis and placement and wiring operations.…”
Section: Experimental Setup and Resultsmentioning
confidence: 99%
“…9(a). 31,32) The weight coefficient was fixed at +1 and the bit width of data was fixed at 8. Since this was a fulldigital implementation, the entire neural net was written in HDL, and the layout could be created only by synthesis and placement and wiring operations.…”
Section: Experimental Setup and Resultsmentioning
confidence: 99%
“…RL thereby achieves long-term results, which are otherwise very difficult to achieve. Deep RL has recently been used in robotic manipulation controllers [13,14]. A deep learning controller based on RL is also implemented in [15] for the application of DL in industrial process control.…”
Section: Output Layermentioning
confidence: 99%
“…Many numerical studies have implemented an inverted pendulum virtual environment as a benchmark to test RL algorithms [ 15 – 22 ], but to our knowledge, there is no study that provides successful RL implementations in experiments. First, except for a few studies that have discussed non ideal systems [ 16 , 17 ], most of these numerical implementations discard the effects associated to realistic (and thus more complex) control methods: in experiments, the control of the cart is subject to delay, hysteresis, biases and noise that can significantly alter the learning process. Second, most of the existing virtual environments consider only motion of the pendulum in a small angle range around the upward and unstable position and do not treat the whole control from the downward and stable position as expected in experiments.…”
Section: Introductionmentioning
confidence: 99%