2019
DOI: 10.1007/s11276-019-02225-x
|View full text |Cite
|
Sign up to set email alerts
|

Design and application of adaptive PID controller based on asynchronous advantage actor–critic learning method

Abstract: To address the problems of the slow convergence and inefficiency in the existing adaptive PID controllers, we propose a new adaptive PID controller using the asynchronous advantage actor-critic (A3C) algorithm. Firstly, the controller can train the multiple agents of the actor-critic structures in parallel exploiting the multi-thread asynchronous learning characteristics of the A3C structure. Secondly, in order to achieve the best control effect, each agent uses a multilayer neural network to approach the stra… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 25 publications
(11 citation statements)
references
References 22 publications
(21 reference statements)
0
7
0
Order By: Relevance
“…It is also possible to determine the values with the perceived dynamics. The advantages of this situation are [37]:…”
Section: Methods 21 Pid Control Systemsmentioning
confidence: 99%
“…It is also possible to determine the values with the perceived dynamics. The advantages of this situation are [37]:…”
Section: Methods 21 Pid Control Systemsmentioning
confidence: 99%
“…Another relevant line of work applied RL to tuning simple parameterized policies, such as PID controllers Dogru et al (2022); Shipman (2021); Carlucho et al (2017); Sun et al (2021); Sedighizadeh and Rezazadeh (2008). In Dogru et al (2022) the authors framed the problem as a contextual bandit problem, where the actions represent PID controller parameters, and an entire episode outcome is treated as a single reward value.…”
Section: Related Workmentioning
confidence: 99%
“…In Carlucho et al (2017) the authors applied tabular Q-learning to output the PI controller parameters to control a mobile robot. In Sun et al (2021); Sedighizadeh and Rezazadeh (2008) the authors applied online RL algorithms to train agents that output PID controller parameters in real time. In real production systems, changing the base policy parameters dynamically at every time step may require additional infrastructure investment on one hand, and may increase safety risks and policy explainability costs on the other.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, soft computing techniques such as neural networks have been successfully used in control systems. The use of artificial intelligence (AI) to replace the traditional PID controller can significantly simplify the tuning process and improve the overall performance and robustness of the control system [27][28][29]. Ghoniem et al [30] replaced a PID controller with a neural network in order to control a new low-cost semi-active vehicle suspension system, and it proved to have good accuracy in terms of vibration reduction and response time.…”
Section: Introductionmentioning
confidence: 99%