2016
DOI: 10.1016/j.imavis.2016.06.006
|View full text |Cite
|
Sign up to set email alerts
|

Action recognition using saliency learned from recorded human gaze

Abstract: This paper addresses the problem of recognition and localization of actions in image sequences, by utilizing, in the training phase only, gaze tracking data of people watching videos depicting the actions in question. First, we learn discriminative action features at the areas of gaze fixation and train a Convolutional Network that predicts areas of fixation (i.e. salient regions) from raw image data. Second, we propose a Support Vector Machine-based recognition method for joint recognition and localization, i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 31 publications
(56 reference statements)
0
3
0
Order By: Relevance
“…Thi et al in [39]proposeda method for action classification and localization by representing human action as a complex set of local features. Stefic and Patras in [40] proposed a method for action recognition using saliency learned from recorded human gaze. Instead of using gaze information as side information, they have trained a model that predicts where people look when presented with image sequences.…”
Section: Image and Vision Computingmentioning
confidence: 99%
“…Thi et al in [39]proposeda method for action classification and localization by representing human action as a complex set of local features. Stefic and Patras in [40] proposed a method for action recognition using saliency learned from recorded human gaze. Instead of using gaze information as side information, they have trained a model that predicts where people look when presented with image sequences.…”
Section: Image and Vision Computingmentioning
confidence: 99%
“…In recent years, some contributions in the field of action recognition have been proposed, such as in Liu, Xu, Qiu, Qing, and Tao (2016), Shi, Laganière, and Petriu (2016), and Stefic and Patras (2016). A summary report of the state-of-the-arts was discussed and presented in González et al (2015) and Ziaeefard and Bergevin (2015).…”
Section: Related Workmentioning
confidence: 99%
“…These images were then used as input to a multilayer CNN which automatically extracted features from the images that were fed in to a multilayer perceptron for classification [ 21 ]. Stefic and Patras utilized CNNs to extract areas of gaze fixation in raw image training data as participants watched videos of multiple activities [ 22 ]. This produced strong results in identifying salient regions of images that were then used for action recognition.…”
Section: Introductionmentioning
confidence: 99%