2021
DOI: 10.1109/lra.2021.3092685
|View full text |Cite
|
Sign up to set email alerts
|

Learning Kinematic Feasibility for Mobile Manipulation Through Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
23
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1

Relationship

2
5

Authors

Journals

citations
Cited by 37 publications
(28 citation statements)
references
References 13 publications
0
23
0
Order By: Relevance
“…Recently, several RL algorithms were proposed for solving MM tasks as interactive navigation, where the hierarchical structure would be employed to decide possible sub-goals for the arm or the base in [7], [8], that would be executed either through RL policies or by motion planning respectively, both however employing precomputed policies for executing the manipulation tasks like pushing or door-opening. On the other part, [9] learns a policy that controls the base velocity using an augmented state-space while maintaining a reward function that would account for the kinematic feasibility of the next end-effector pose, which was, interestingly, sampled from the path that connected the end-effector and the goal. Learning whole-body control for MM seems to benefit from structural information, as shown in [26], [27].…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Recently, several RL algorithms were proposed for solving MM tasks as interactive navigation, where the hierarchical structure would be employed to decide possible sub-goals for the arm or the base in [7], [8], that would be executed either through RL policies or by motion planning respectively, both however employing precomputed policies for executing the manipulation tasks like pushing or door-opening. On the other part, [9] learns a policy that controls the base velocity using an augmented state-space while maintaining a reward function that would account for the kinematic feasibility of the next end-effector pose, which was, interestingly, sampled from the path that connected the end-effector and the goal. Learning whole-body control for MM seems to benefit from structural information, as shown in [26], [27].…”
Section: Related Workmentioning
confidence: 99%
“…SAC-hybrid, which is, in essence, the implementation of HyRL with maximum entropy exploration; SAC with no discrete action space (SAC-continuous), which resembles the method of [8], where the selection of the embodiment can be tackled as thresholding of a continuous value, but this does not affect the learning as it happens outside the MDP; iv. a variant of learning kinematic feasibility (LKF) [9], in which instead of predicting base velocities, we predict only base sub-goals, so that it can be directly comparable to ours, but there is no policy for arm activation -IK is checked at every step. Please note, that we extended all methods to consider 6D goal poses to be reached by the robot.…”
Section: B Evaluation 1) MM -Reachmentioning
confidence: 99%
See 2 more Smart Citations
“…However, this severely limits the utility of this method for diverse applications. For example, it can neither be used for mobile robots [14], nor for manipulation tasks where a wrist camera is required for precise alignment of the gripper [15].…”
Section: Introductionmentioning
confidence: 99%