2023
DOI: 10.1109/tpami.2021.3055233
|View full text |Cite
|
Sign up to set email alerts
|

Forecasting Action Through Contact Representations From First Person Video

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
32
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 42 publications
(32 citation statements)
references
References 43 publications
0
32
0
Order By: Relevance
“…Action Anticipation Different works have tackled this task [12,41,6,29]. Previous approaches have considered baselines designed for action recognition [4], defined custom losses [11], modeled the evolution of scene attributes and action over time [31], disentangled the tasks of encoding and anticipation [12], aggregated features over time [41], predicted motor attention [29], leveraged contact representations [6], mimicked intuitive and analytical thinking [53], and predicted future representations [51]. While these approaches have been designed to maximize performance when predicting the future, they have never been evaluated in a streaming scenario.…”
Section: Egocentricmentioning
confidence: 99%
See 1 more Smart Citation
“…Action Anticipation Different works have tackled this task [12,41,6,29]. Previous approaches have considered baselines designed for action recognition [4], defined custom losses [11], modeled the evolution of scene attributes and action over time [31], disentangled the tasks of encoding and anticipation [12], aggregated features over time [41], predicted motor attention [29], leveraged contact representations [6], mimicked intuitive and analytical thinking [53], and predicted future representations [51]. While these approaches have been designed to maximize performance when predicting the future, they have never been evaluated in a streaming scenario.…”
Section: Egocentricmentioning
confidence: 99%
“…Wearable devices equipped with egocentric cameras are recently attracting attention as an ideal platform to implement intelligent agents able to provide assistance to humans in a natural way [22]. Among the different problems studied in egocentric vision, the task of action anticipation, which consists in predicting a plausible future action before it is performed by the camera wearer, has attracted a lot of attention [2,4,6,12,29,34,41,51,53]. Indeed, from a practical point of view, being able to predict future events is fundamental when designing technologies which can assist humans in their daily and working activities [24,43].…”
Section: Introductionmentioning
confidence: 99%
“…The main focus of these works is to extract relevant information from the observations to predict the label of the action starting in τ seconds, varying between zero [32] to 10s of seconds [33]. Other models leverage external cues such as hand movements to help with the anticipation task [34,35].…”
Section: Related Workmentioning
confidence: 99%
“…Recognizing how hands interact with objects is crucial to understand how we interact with the world. Hand-object interaction analysis contributes to several fields such as action prediction [10], rehabilitation [28], robotics [38], and virtual reality [17].…”
Section: Introductionmentioning
confidence: 99%