2017
DOI: 10.1155/2017/8191537
|View full text |Cite
|
Sign up to set email alerts
|

Improved Collaborative Representation Classifier Based on l2-Regularized for Human Action Recognition

Abstract: Human action recognition is an important recent challenging task. Projecting depth images onto three depth motion maps (DMMs) and extracting deep convolutional neural network (DCNN) features are discriminant descriptor features to characterize the spatiotemporal information of a specific action from a sequence of depth images. In this paper, a unified improved collaborative representation framework is proposed in which the probability that a test sample belongs to the collaborative subspace of all classes can … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 21 publications
0
2
0
Order By: Relevance
“…The deep convolutional features of video frames are extracted through deep neural networks; then a pyramidlike structure of video feature expression is constructed, and then the feature expression of the entire video sequence is obtained; finally, the video is classified and recognized by the classifier [7,8]. A multi-source deep neural network model is constructed, which non-linearly represents image features from different sources, and estimates the human body's action posture by extracting high-level human joint features from video information [9]. By fine-tuning the deep convolutional neural network trained on the Imagen net dataset, several feature fusion strategies are proposed based on the idea of fusing the features of the convolutional neural network.…”
Section: Introductionmentioning
confidence: 99%
“…The deep convolutional features of video frames are extracted through deep neural networks; then a pyramidlike structure of video feature expression is constructed, and then the feature expression of the entire video sequence is obtained; finally, the video is classified and recognized by the classifier [7,8]. A multi-source deep neural network model is constructed, which non-linearly represents image features from different sources, and estimates the human body's action posture by extracting high-level human joint features from video information [9]. By fine-tuning the deep convolutional neural network trained on the Imagen net dataset, several feature fusion strategies are proposed based on the idea of fusing the features of the convolutional neural network.…”
Section: Introductionmentioning
confidence: 99%
“…Pedestrian re-identification refers to the similarity matching of pedestrian targets under the surveillance of multiple cameras without the overlapping of viewing angles. Its application in monitoring systems is a hot topic in the field of computer vision [5], [36]. As shown in Figure 1 [6], the target (the pedestrian, point 1), can be identified and locked in multiple cameras in different locations or in the same video (point 1, dotted circles).…”
Section: Introductionmentioning
confidence: 99%