2020
DOI: 10.1080/24751839.2020.1833137
|View full text |Cite
|
Sign up to set email alerts
|

Heuristic and deep reinforcement learning-based PID control of trajectory tracking in a ball-and-plate system

Abstract: The manual tuning of controller parameters, for example, tuning proportional integral derivative (PID) gains often relies on tedious human engineering. To curb the aforementioned problem, we propose an artificial intelligence-based deep reinforcement learning (RL) PID controller (three variants) compared with genetic algorithm-based PID (GA-PID) and classical PID; a total of five controllers were simulated for controlling and trajectory tracking of the ball dynamics in a linearized ball-and-plate (B&P) system.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
17
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 11 publications
(17 citation statements)
references
References 23 publications
(30 reference statements)
0
17
0
Order By: Relevance
“…The deep reinforcement learning architecture proposed is motivated by [37], which yielded the best performance than the heuristic or classical methods when tracking the ball dynamics in a ball and plate system. In the current study, we modify the architecture to deal with the MPP tracking task.…”
Section: ) Proposed Deep Reinforcement Learning Architecturementioning
confidence: 99%
See 1 more Smart Citation
“…The deep reinforcement learning architecture proposed is motivated by [37], which yielded the best performance than the heuristic or classical methods when tracking the ball dynamics in a ball and plate system. In the current study, we modify the architecture to deal with the MPP tracking task.…”
Section: ) Proposed Deep Reinforcement Learning Architecturementioning
confidence: 99%
“…The layers containing {300, 2} network nodes employ exponential linear unit (ELU) activation function, which can be defined mathematically using Eq. 18 [37], [38].…”
Section: Calculate Rewardmentioning
confidence: 99%
“…It is a learning paradigm concerned with learning to control a system in order to maximize a cumulative expected reward performance measure that expresses a long-term objective [14] and can determine the optimal policy for decisions in a real environment. Recently, research into RL methods has been extended into multiple control fields such as trajectory tracking [15][16][17][18], path-following [19,20], etc. [21].…”
Section: Introductionmentioning
confidence: 99%
“…It is noted that RL is capable of coping with a control problem without knowing information about the objective dynamics and presents control with good performance under the influence of external disturbances [15,16,20]. Reinforcement learning can be combined with other classical control methods to solve tracking problems [17]. Considering the PID method, a Q-learning-PID control approach has been proposed to solve the trajectory tracking control problem for mobile robots, with better results than the single approach [18].…”
Section: Introductionmentioning
confidence: 99%
“…The B&P system finds practical applications in several dynamic systems, such as robotics, rocket systems, and unmanned aerial vehicles. The enumerated systems are often expected to follow a time parameterized reference path [2]. Due to the BP complexity and it's nonlinearity, the mathematical model of the BPS presents uncertainties which increase the difficulty of designing a suitable controller.…”
Section: Introductionmentioning
confidence: 99%