2020
DOI: 10.3389/fnbot.2020.567571
|View full text |Cite
|
Sign up to set email alerts
|

Toward Shared Autonomy Control Schemes for Human-Robot Systems: Action Primitive Recognition Using Eye Gaze Features

Abstract: The functional independence of individuals with upper limb impairment could be enhanced by teleoperated robots that can assist with activities of daily living. However, robot control is not always intuitive for the operator. In this work, eye gaze was leveraged as a natural way to infer human intent and advance action recognition for shared autonomy control schemes. We introduced a classifier structure for recognizing low-level action primitives that incorporates novel three-dimensional gaze-related features. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 11 publications
(24 citation statements)
references
References 61 publications
0
12
0
Order By: Relevance
“…Our study follows similar motivations as and Wang et al (2020) and complements those results, while not being aimed specifically at assistive applications, but rather trying to leverage human dexterity and eye-hand coordination to improve performance in teleoperated manipulation tasks. To investigate human oculomotor behavior during teleoperation in a more controlled scenario and with a more natural input interface, we designed an experiment in simulation, where the participant would control the remote robot arm by means of their own arm movements via motion tracking.…”
Section: Related Workmentioning
confidence: 64%
See 3 more Smart Citations
“…Our study follows similar motivations as and Wang et al (2020) and complements those results, while not being aimed specifically at assistive applications, but rather trying to leverage human dexterity and eye-hand coordination to improve performance in teleoperated manipulation tasks. To investigate human oculomotor behavior during teleoperation in a more controlled scenario and with a more natural input interface, we designed an experiment in simulation, where the participant would control the remote robot arm by means of their own arm movements via motion tracking.…”
Section: Related Workmentioning
confidence: 64%
“…To evaluate the intention recognition performance, we looked on the one hand at how accurate was the prediction whenever a prediction was indeed available (that is, the proportion of correct predictions over the overall number of delivered predictions). On the other hand, as stated in Ellis et al (2013) and Wang et al (2020), there is a trade-off between accuracy and observational latency. Indeed, the more evidence is accumulated before making a prediction, the more accurate the prediction is going to be.…”
Section: Intention Recognition Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…However, one can reasonably expect similar outcomes in other cases, even when the choice of coordinates acting as contextual information is suboptimal. Following a similar approach, recent reports [ 22 , 23 ] have investigated how various forms of knowledge about the user’s movement intention or goal can improve the control of a prosthetic or assistive robotic arm.…”
Section: Introductionmentioning
confidence: 99%