2021
DOI: 10.1016/j.neucom.2020.12.069
|View full text |Cite
|
Sign up to set email alerts
|

Predicting short-term next-active-object through visual attention and hand position

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(11 citation statements)
references
References 21 publications
0
8
0
Order By: Relevance
“…[19] also explored NAO prediction using cues from visual attention and hand position, but by only using a single frame for the prediction. That approach [19] is not able to differentiate between the past or future active object, since it does not account for the temporal information acquired by the videos. Furnari et al [12] also explored the NAO problem by taking into account the active/passive objects definition of [29].…”
Section: Action Anticipation In Egocentric Videosmentioning
confidence: 99%
“…[19] also explored NAO prediction using cues from visual attention and hand position, but by only using a single frame for the prediction. That approach [19] is not able to differentiate between the past or future active object, since it does not account for the temporal information acquired by the videos. Furnari et al [12] also explored the NAO problem by taking into account the active/passive objects definition of [29].…”
Section: Action Anticipation In Egocentric Videosmentioning
confidence: 99%
“…The output is composed of the next action label, a "hotspot" which indicates the area of the object in which there will be a contact, and the hand trajectory. The two egocentric datasets ADL [70] and EPIC-Kitchens [15] have been re-annotaded by [46] to tackle the problem of short-term next-active object detection. They proposed a novel human-centerd approach composed of two pathways: 1) the first pathway generates a human visual attention probability map and 2) the second one generates a human hand position probability map.…”
Section: Next Active Objects Detectionmentioning
confidence: 99%
“…Due to the limited number of public datasets explicitly annotated to study the future intentions of humans, few past works explored the task of predicting the next-active objects considering the first person point of view [32], using both RGB and depth signals [3], focusing on the hands [46] or estimating also the time to contact with the future active objects [41].…”
Section: Next Active Object Annotationsmentioning
confidence: 99%
See 1 more Smart Citation
“…Previous works have investigated different forms of anticipation tasks, including next-active object predictions [5,11,19,21,32], predicting future actions [10, 12-14, 30, 35,37], forecasting human-object interactions [26], predicting future hands [6,20] or user trajectory prediction [31].…”
Section: Introductionmentioning
confidence: 99%