2019
DOI: 10.1007/s10846-019-01030-0
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Hybrid-Supervised Deep Reinforcement Learning for Person Following Robot

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
19
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 20 publications
(19 citation statements)
references
References 28 publications
0
19
0
Order By: Relevance
“…Pang et al [9] applied a hybrid method of supervised learning (SL) and DRL to train an agent with deep Q-network (DQN) for developing a robot which is able to continuously follow a specified target person from behind. In order to generate the robot's actions, the study uses images obtained from an RGB camera attached to the robot as states which are then fed directly to the neural networks inside the agent.…”
Section: A Drl For Person-following Robotsmentioning
confidence: 99%
See 2 more Smart Citations
“…Pang et al [9] applied a hybrid method of supervised learning (SL) and DRL to train an agent with deep Q-network (DQN) for developing a robot which is able to continuously follow a specified target person from behind. In order to generate the robot's actions, the study uses images obtained from an RGB camera attached to the robot as states which are then fed directly to the neural networks inside the agent.…”
Section: A Drl For Person-following Robotsmentioning
confidence: 99%
“…Instead, the policy is updated through rewards obtained from the interaction of an RL agent with its environment during the training process. Recently, several studies such as [6] and [9] have shown that DRL can be successfully implemented to train agents for obtaining the policy for a robot to generate appropriate actions to follow a specified target person. However, the person-following task to be performed by the robot in [9] is not seen as an integrated substantial task so that another important sub task, namely the obstacle avoidance is neglected.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…We use continuous action space [23] so that the agent will generate more dimension output actions resulting more smooth robot's trajectories when it navigates [47]. The continuous action space a ∈ A is defined as…”
Section: ) Action Spacementioning
confidence: 99%
“…Thus, the motion trajectory of the robot is not smooth. In addition to the above traditional following methods, [31] proposed an approach in which an agent is trained by hybrid-supervised deep reinforcement learning to perform a person following task in end-to-end manner. However, like some traditional methods, this method does not consider the avoidance of dynamic obstacles, and it's difficult to control the robot to follow people smoothly with action instructions directly generated from discrete images.…”
Section: B Person Followingmentioning
confidence: 99%