2009
DOI: 10.1109/tpami.2009.43
|View full text |Cite
|
Sign up to set email alerts
|

Human Action Recognition by Semilatent Topic Models

Abstract: We propose two new models for human action recognition from video sequences using topic models. Video sequences are represented by a novel "bag-of-words" representation, where each frame corresponds to a "word." Our models differ from previous latent topic models for visual recognition in two major aspects: first of all, the latent topics in our models directly correspond to class labels; second, some of the latent variables in previous topic models become observed in our case. Our models have several advantag… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2011
2011
2019
2019

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 280 publications
(21 citation statements)
references
References 40 publications
0
21
0
Order By: Relevance
“…Here, the experimental result of proposed approach is presented. As KTH human action database [39] has been used for benchmarking the accuracy of consistency with set of experiments used in [2, 22, 27, 43, 44], we made a set of our training map and test set for proposed technique on the entire dataset, in which the mixture of four scenarios videos was together. The dataset split into a set of training maps with five randomly selected subjects and a test part by residual subjects.…”
Section: Evaluation and Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Here, the experimental result of proposed approach is presented. As KTH human action database [39] has been used for benchmarking the accuracy of consistency with set of experiments used in [2, 22, 27, 43, 44], we made a set of our training map and test set for proposed technique on the entire dataset, in which the mixture of four scenarios videos was together. The dataset split into a set of training maps with five randomly selected subjects and a test part by residual subjects.…”
Section: Evaluation and Resultsmentioning
confidence: 99%
“…Using optical flow (as it is aforementioned) is one of the effective methods in human action recognition (see [27]). For having the features regarding motion of subject, layer-wise optical flow estimation has been done.…”
Section: System Overviewmentioning
confidence: 99%
“…Similarity of the presented model has been deliberated in the assessment. KTH human action dataset is used for benchmarking and the evaluation assessment compared with state-of-the-art methods in the same dataset for consistency in the experimental results (see Figure 4) [11, 4246]. Also it is noticeable, as previously mentioned, that the training map and action prototypes obtained from the random selection of the human action set in four different scenario videos from KTH and excluded from the testing set have no overlap between these two sets.…”
Section: Resultsmentioning
confidence: 99%
“…In this paper, we use optical flow to detect the relative direction and magnitude of environmental motion observed in reference to an observer and also describe the movement of object from current image with respect to the last image. The optical flow [22, 29] equation can be assumed to hold for all pixels within a window centered at p , the local image flow (velocity) vector ( V x , V y ) must be satisfied, and we define some equations as follows. Ix(s1)Vx+Iy(s1)Vy=It(s1),Ix(s2)Vx+Iy(s2)Vy=It(s2),Ix(sd)Vx+Iy(sd)Vy=It(sd), where s 1 , s 2 ,…, s d are the pixels inside the window and I x ( s i ), I y ( s i ), and I t ( s i ) are the partial derivatives of the image I with respect to position x , y and time t , evaluated at the point S i and the current time.…”
Section: Feature Representationmentioning
confidence: 99%
“…In order to capture the correlation of topics, we model the hyperparameter of topic prior distribution as multivariate normal distribution instead of Dirichlet, similar to [29, 32] and the structure topics of dependencies by covariance matrix, and then we use the logistic normal function: f(θie)=(expθifalse∑j=normal1Kexpθj) to project the multivariate normal to topic proportions for each image; here K is the topic number.…”
Section: Action Classificationmentioning
confidence: 99%