2013
DOI: 10.1007/978-3-319-02675-6_46
|View full text |Cite
|
Sign up to set email alerts
|

Training a Robot via Human Feedback: A Case Study

Abstract: Abstract. We present a case study of applying a framework for learning from numeric human feedback-TAMER-to a physically embodied robot. In doing so, we also provide the first demonstration of the ability to train multiple behaviors by such feedback without algorithmic modifications and of a robot learning from free-form human-generated feedback without any further guidance or evaluative feedback. We describe transparency challenges specific to a physically embodied robot learning from human feedback and adjus… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
88
0

Year Published

2014
2014
2023
2023

Publication Types

Select...
6
4

Relationship

0
10

Authors

Journals

citations
Cited by 106 publications
(88 citation statements)
references
References 17 publications
0
88
0
Order By: Relevance
“…Their algorithm uses reinforcement learning to improve over the initial sequences provided by the user, and it incorporates on-line feedback from the user during the learning process creating a novel dynamic reward shaping mechanism to converge faster to an optimal policy. Furthermore, in [18], the human trainer, an author of that study, followed a predetermined algorithm of giving positive reward for desired actions and negative reward otherwise. They explored how the Interactive Reinforcement Learning algorithm that enables a human trainer to provide both rewards and anticipatory guidance for the learner can be applied to a real-world robotic system.…”
Section: Related Workmentioning
confidence: 99%
“…Their algorithm uses reinforcement learning to improve over the initial sequences provided by the user, and it incorporates on-line feedback from the user during the learning process creating a novel dynamic reward shaping mechanism to converge faster to an optimal policy. Furthermore, in [18], the human trainer, an author of that study, followed a predetermined algorithm of giving positive reward for desired actions and negative reward otherwise. They explored how the Interactive Reinforcement Learning algorithm that enables a human trainer to provide both rewards and anticipatory guidance for the learner can be applied to a real-world robotic system.…”
Section: Related Workmentioning
confidence: 99%
“…The doll is designed to carry on an extensive conversation with young girls or boys in areas of their interest. Knox, Breazeal, and Stone (2013) present a case study of applying a framework for learning from human feedback to an interactive robot. They claim this application as a first demonstration of the ability to train multiple behaviors in robot learning from free-form human-generated feedback without any further guidance or evaluative feedback.…”
Section: Human-robot Social Interactionmentioning
confidence: 99%
“…Some approaches use human feedback as shaping signals to teach a system how to achieve a task. In such approaches, the source of the feedback is considered as an observer of the system who evaluates each of the system's actions [26,27] or the system's entire policy [28]. The TAMER ( Training an Agent Manually via Evaluative Reinforcement) framework [26] proposes a method to shape a learning robot by giving positive and negative signals (as for a domestic dog).…”
Section: User Feedback For Smarter Homesmentioning
confidence: 99%