2018
DOI: 10.1177/0278364918776060
|View full text |Cite
|
Sign up to set email alerts
|

Shared autonomy via hindsight optimization for teleoperation and teaming

Abstract: In shared autonomy, a user and autonomous system work together to achieve shared goals. To collaborate effectively, the autonomous system must know the user's goal. As such, most prior works follow a predict-then-act model, first predicting the user's goal with high confidence, then assisting given that goal. Unfortunately, confidently predicting the user's goal may not be possible until they have nearly achieved it, causing predict-then-act methods to provide little assistance. However, the system can often p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
152
1

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 165 publications
(166 citation statements)
references
References 76 publications
2
152
1
Order By: Relevance
“…As mentioned in the introduction, a key motivation behind learning agent models is to enable effective human-machine interaction. The algorithms presented in this work can be used both by a collaborative robot (for learning models of human behavior to improve robot decision-making) and by a human (to learn transparent models of robot behavior to better calibrate trust) (Javdani et al 2018;Yang et al 2017). Both of these use cases offer novel research avenues, such as the efficient specification of decision-theoretic models for interaction (such as, POMDPs and decentralized-POMDPs) and evaluation of the utility of aligned models as compared to purely predictive models for human-machine interaction (Oliehoek, Amato, and others 2016).…”
Section: Discussionmentioning
confidence: 99%
“…As mentioned in the introduction, a key motivation behind learning agent models is to enable effective human-machine interaction. The algorithms presented in this work can be used both by a collaborative robot (for learning models of human behavior to improve robot decision-making) and by a human (to learn transparent models of robot behavior to better calibrate trust) (Javdani et al 2018;Yang et al 2017). Both of these use cases offer novel research avenues, such as the efficient specification of decision-theoretic models for interaction (such as, POMDPs and decentralized-POMDPs) and evaluation of the utility of aligned models as compared to purely predictive models for human-machine interaction (Oliehoek, Amato, and others 2016).…”
Section: Discussionmentioning
confidence: 99%
“…Memory-based inference is utilized in previous works involving shared autonomy and human-robot systems [8,9,30]. Another approach uses Laplace's approximation [9] and formulates the problem as one of optimizing a partially observable Markov decision process (POMDP) over the user's goal to arbitrate control over a distribution of possible outcomes [20]. The approach considers user inputs for the prediction model and uses a hand-specified distance-based user cost function to achieve a closed-form value function computation.…”
Section: Related Workmentioning
confidence: 99%
“…In addition to a distancebased likelihood observation used in the prior works, our approach incorporates a fusion of multiple observations and introduces a probabilistic modeling of user control inputs as goal-directed actions that customizes the rationality index value to each individual user, and thus can account for their particular behavior. The motivation for our approach are prior studies that show that users vary greatly in their performance, preferences, and desires [9,20,37]-suggesting a need for assistive systems to customize to individual users.…”
Section: Related Workmentioning
confidence: 99%
“…Intent can also be inferred from the user's control signals and other environmental cues using various algorithms [11]. Within the context of shared autonomy a Bayesian scheme for user intent prediction models the user within the Markov Decision Process framework [12], [13], [14] and is typically assumed to be noisily optimizing some cost function for their intended goal. In low-dimensional spaces, this cost function can be learned from expert demonstrations using Inverse Reinforcement Learning [15].…”
Section: Related Workmentioning
confidence: 99%