Companion Proceedings of the 20th International Conference on Intelligent User Interfaces 2015
DOI: 10.1145/2732158.2732180
|View full text |Cite
|
Sign up to set email alerts
|

Towards Integrating Real-Time Crowd Advice with Reinforcement Learning

Abstract: Reinforcement learning is a powerful machine learning paradigm that allows agents to autonomously learn to maximize a scalar reward. However, it often suffers from poor initial performance and long learning times. This paper discusses how collecting on-line human feedback, both in real time and post hoc, can potentially improve the performance of such learning systems. We use the game Pac-Man to simulate a navigation setting and show that workers are able to accurately identify both when a sub-optimal action i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(2 citation statements)
references
References 10 publications
0
2
0
Order By: Relevance
“…While Legion provided generalized user interface control, Salisbury et al (Salisbury, Stein, and Ramchurn 2015b) introduced additional control mediators that improved performance by focusing specifically on robotics applications. de la Cruz et al (de la Cruz et al 2015) got feedback from a crowd of workers in ∼0.3 seconds for mistakes made by an automated agent. Chung et al (Chung et al 2014) explored learning from initial demonstrations using crowd feedback for motion planning problems.…”
Section: Robotics and Autonomous Controlmentioning
confidence: 99%
“…While Legion provided generalized user interface control, Salisbury et al (Salisbury, Stein, and Ramchurn 2015b) introduced additional control mediators that improved performance by focusing specifically on robotics applications. de la Cruz et al (de la Cruz et al 2015) got feedback from a crowd of workers in ∼0.3 seconds for mistakes made by an automated agent. Chung et al (Chung et al 2014) explored learning from initial demonstrations using crowd feedback for motion planning problems.…”
Section: Robotics and Autonomous Controlmentioning
confidence: 99%
“…Practically, a weight could be computed for each human which indicates the competence level of that human at the task, and all trajectories supplied by that human would be weighted accordingly. This approach would be useful for example in a crowdsourcing framework [22].…”
Section: E Human Elicited Priorsmentioning
confidence: 99%