2022
DOI: 10.1016/j.artint.2022.103771
|View full text |Cite
|
Sign up to set email alerts
|

Q-Learning-based model predictive variable impedance control for physical human-robot collaboration

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
18
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 32 publications
(18 citation statements)
references
References 69 publications
0
18
0
Order By: Relevance
“…In the human-robot collaborative motion experiment, after completing a point-to-point motion, each participant answered the questions in Table V according to the motion's naturalness, smoothness, stability, overall performance, etc. [12]. Questions 1-4 and 6 adopt 5-point Likert scale [36].…”
Section: Experiments Taskmentioning
confidence: 99%
See 1 more Smart Citation
“…In the human-robot collaborative motion experiment, after completing a point-to-point motion, each participant answered the questions in Table V according to the motion's naturalness, smoothness, stability, overall performance, etc. [12]. Questions 1-4 and 6 adopt 5-point Likert scale [36].…”
Section: Experiments Taskmentioning
confidence: 99%
“…Such as many papers are estimating the properties of the environment online to determine the control action guaranteeing stability or other performance [9,10]. Considering human-robot interaction, some papers are specifically estimating online the human-robot interaction dynamics to modulate the control action [11], guaranteeing the stability of the controller [12]. Therefore, the admittance control with fixed parameters is only applicable to the situation where the environment is fixed and the modeling is simple.To address the changing environment and improve the robot's interaction with the external environment, more researchers have studied variable admittance control of the robot.…”
mentioning
confidence: 99%
“…The problem of adapting the damping and stiffness terms is faced in [3], where Adversarial Inverse Reinforcement Learning (AIRL) is exploited to infer, starting from human "expert" demonstrations, the objective function against which the variable terms are optimized. Recent research has shown how Neural Networks (NNs) can be exploited to learn interaction dynamics [16], whether the contacts happens with a working environment [17] or a human being [18]. Estimating this dynamics can be exploited for the computation of optimal impedance gains for a safe and compliant interaction [19].…”
Section: B Related Workmentioning
confidence: 99%
“…In this scope, the methodology proposed in this article takes advantage of AI techniques to develop an optimizationbased control law. In literature, NNs learning the unknown interaction dynamics are mainly exploited for optimizing the robot compliance [17], [18], whereas our objective is delivering accurate force tracking. As regards the optimization of robot-environment interaction tasks, the state-of-art strategies suppose that an "expert" system is employed beforehand to register a policy, from which an optimal control configuration is devised [3], [15], with the policy being strongly taskdependent.…”
Section: B Related Workmentioning
confidence: 99%
See 1 more Smart Citation