2020
DOI: 10.48550/arxiv.2010.04689
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

LaND: Learning to Navigate from Disengagements

Abstract: Fig. 1: LaND is a learning-based approach for autonomous mobile robot navigation that directly learns from disengagements-any time a human monitor disengages the robot's autonomy. These disengagement datasets are ubiquitous because they are naturally collected during the process of testing these autonomous systems. LaND is able to navigate in a diverse set of sidewalk environments, including parked bicycles, dense foliage, parked cars, sun glare, sharp turns, and unexpected obstacles.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 31 publications
0
2
0
Order By: Relevance
“…To reduce the burden on the supervisor, several robot-gated interactive IL algorithms such as SafeDAgger [9], EnsembleDAgger [5], LazyDAgger [10], and ThriftyDAgger [8] have been proposed, in which the robot actively solicits human interventions when certain criteria are met. Interactive reinforcement learning (RL) [53,54,55,56] is another active area of research in which robots learn from both online human feedback and their own experience. However, these interactive learning algorithms are designed for and primarily studied in the single-robot, single-human setting.…”
Section: Single-robot Single-human Interactive Learningmentioning
confidence: 99%
“…To reduce the burden on the supervisor, several robot-gated interactive IL algorithms such as SafeDAgger [9], EnsembleDAgger [5], LazyDAgger [10], and ThriftyDAgger [8] have been proposed, in which the robot actively solicits human interventions when certain criteria are met. Interactive reinforcement learning (RL) [53,54,55,56] is another active area of research in which robots learn from both online human feedback and their own experience. However, these interactive learning algorithms are designed for and primarily studied in the single-robot, single-human setting.…”
Section: Single-robot Single-human Interactive Learningmentioning
confidence: 99%
“…Kurenkov et al [26] and Xie et al [48] leverage interventions from suboptimal supervisors to accelerate policy learning, but assume that the supervisors are algorithmic and thus can be queried cheaply. Amir et al [2], Kahn et al [23], Kelly et al [24], Spencer et al [42], and Wang et al [47] instead consider learning from human supervisors and present learning algorithms which utilize the timing and nature of human interventions to update the learned policy. By giving the human control for multiple timesteps in a row, these algorithms show improvements over methods that only hand over control on a state-by-state basis [6].…”
Section: Background and Related Workmentioning
confidence: 99%