2022
DOI: 10.1002/asjc.2830
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive reinforcement learning in control design for cooperating manipulator systems

Abstract: In this paper, an optimal motion/force hybrid control strategy based on adaptive reinforcement learning (ARL) is proposed for cooperating manipulator systems.The optimal trajectory control and constraint force factor control, by using the Moore-Penrose pseudoinverse, are addressed to design the controller corresponding to the manipulator dynamic model. In addition, a frame of a different auxiliary term and an appropriate state-variable vector are presented to address the non-autonomous closed system with a tim… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
22
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6

Relationship

1
5

Authors

Journals

citations
Cited by 29 publications
(24 citation statements)
references
References 35 publications
0
22
0
Order By: Relevance
“…The proposed algorithm was evaluated in the literature using different intelligent techniques, including genetic algorithms and cuckoo search algorithms with PID, proving its superiority in following trajectories. Dao and Liu 53 proposed an adaptive reinforcement learning‐based optimal motion/force hybrid strategy for a collaborative robotic. The Moore‐Penrose pseudo‐inverse algorithm is used for optimal trajectory and binding factor, and a controller corresponding to the robotic arm dynamics model is designed.…”
Section: Related Workmentioning
confidence: 99%
“…The proposed algorithm was evaluated in the literature using different intelligent techniques, including genetic algorithms and cuckoo search algorithms with PID, proving its superiority in following trajectories. Dao and Liu 53 proposed an adaptive reinforcement learning‐based optimal motion/force hybrid strategy for a collaborative robotic. The Moore‐Penrose pseudo‐inverse algorithm is used for optimal trajectory and binding factor, and a controller corresponding to the robotic arm dynamics model is designed.…”
Section: Related Workmentioning
confidence: 99%
“…In current research, RL based dynamic Programming has been used for evaluating the Optimal Reward Function to be incorporated with conventional DDPG Algorithm, which makes it novel and unique as compared to other approaches. In another paper, Dao et al [26] employed Actor Critic Network through ARL. This work is in accordance with the current work in which both DDPG and PPO have been used which are also from a family of Actor-Critic networks, with prime difference of a Reward Function.…”
Section: Relevant Studiesmentioning
confidence: 99%
“…Therefore, many scholars investigate control designs for each agent with the Euler-Lagrange (EL) dynamic equation, as well as MAS control systems. However, due to the difficulty in unifying the optimal control task and the trajectory tracking problem, there exist a few optimal control approaches, [1][2][3][4] but many typical robust adaptive control schemes [5][6][7][8][9][10] for each agent are represented under the EL dynamic equation. Several authors utilize the description of the Hamiltonian function to find the updating laws of actor/critic (AC) neural networks (NN) [1][2][3][4] after establishing tracking error models under a time-invariant representation for cooperating manipulators, 1,2 surface vehicles (SVs), 3 or wheeled mobile robots (WMRs).…”
Section: Introductionmentioning
confidence: 99%
“…However, due to the difficulty in unifying the optimal control task and the trajectory tracking problem, there exist a few optimal control approaches, [1][2][3][4] but many typical robust adaptive control schemes [5][6][7][8][9][10] for each agent are represented under the EL dynamic equation. Several authors utilize the description of the Hamiltonian function to find the updating laws of actor/critic (AC) neural networks (NN) [1][2][3][4] after establishing tracking error models under a time-invariant representation for cooperating manipulators, 1,2 surface vehicles (SVs), 3 or wheeled mobile robots (WMRs). 4 On the other hand, based on the foundation of traditional nonlinear control laws, some remarkable progress is mentioned, including the extension of the event-triggered strategy 6 and the development of special transformation functions in adaptive backstepping methods to handle constraint requirements.…”
Section: Introductionmentioning
confidence: 99%