Proceedings of the 24th ACM International Conference on Multimedia 2016
DOI: 10.1145/2964284.2967191
|View full text |Cite
|
Sign up to set email alerts
|

Action Recognition Based on Joint Trajectory Maps Using Convolutional Neural Networks

Abstract: Recently, Convolutional Neural Networks (ConvNets) have shown promising performances in many computer vision tasks, especially imagebased recognition. How to effectively use ConvNets for video-based recognition is still an open problem. In this paper, we propose a compact, effective yet simple method to encode spatiotemporal information carried in 3D skeleton sequences into multiple 2D images, referred to as Joint Trajectory Maps (JTM), and ConvNets are adopted to exploit the discriminative features for real-t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
212
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 257 publications
(214 citation statements)
references
References 22 publications
2
212
0
Order By: Relevance
“…To verify the differences from the other DTW algorithms, we compared DTW and modified DTW. In comparison with the general methods, we used discrete HMM, SVM (for MSRC-12 datasets only) [22], and two CNN methods (SOS, JTM) [23,24]. The 1-NN method was used to find the most similar movement in the gesturerecognition system, and the threshold-value setting of the gesture-recognition system was used to filter out insignificant gestures.…”
Section: Comparison and Analysis Of Dtw Performance Based On Dsmentioning
confidence: 99%
“…To verify the differences from the other DTW algorithms, we compared DTW and modified DTW. In comparison with the general methods, we used discrete HMM, SVM (for MSRC-12 datasets only) [22], and two CNN methods (SOS, JTM) [23,24]. The 1-NN method was used to find the most similar movement in the gesturerecognition system, and the threshold-value setting of the gesture-recognition system was used to filter out insignificant gestures.…”
Section: Comparison and Analysis Of Dtw Performance Based On Dsmentioning
confidence: 99%
“…We compare it to other skeleton representations in the literature. Besides the classical skeleton image representation of Du et al [4], we compare with other representations used by state-ofthe-art approaches [32,10,13,14,34] as baselines on NTU RGB+D 60 [25]. We also compare our approach to sate-of-the-art methods on the NTU RGB+D 120 [15].…”
Section: Resultsmentioning
confidence: 99%
“…Wang et al [32,30] present a skeleton representation to represent both spatial configuration and dynamics of joint trajectories into three texture images through color encoding, named Joint Trajectory Maps (JTMs). The authors apply rotations to the skeleton data to mimicking multi-views and also for data enlargement to overcome the drawback of CNNs usually being not view invariant.…”
Section: Related Workmentioning
confidence: 99%
“…To prove that a good structural organization of joints is important to preserve the spatial relations of the skeleton data, we compare our approach with a baseline employing random joints order when creating the representation (i.e., the creation of the chains' order C t does not take into account any semantic meaning of adjacent joints). Moreover, we also compare with the classical skeleton image representations used by state-ofthe-art approaches [12], [13], [15], [16], [18], [19] as well as to estate-of-the-art methods on the NTU RGB+D 120 [21].…”
Section: Resultsmentioning
confidence: 99%
“…Table I presents a comparison of our approach with skeleton image representations of the literature. For the methods that have more than one "image" per representation ( [13] and [15]), we stacked them to be used as input to the network. The same was performed for our TSRJI (Stacked) approach considering the images for each reference joint (i.e., S a , ... [18] achieving 70.8% of accuracy 69.5%.…”
Section: Discussionmentioning
confidence: 99%