2019
DOI: 10.1016/j.patrec.2018.05.018
|View full text |Cite
|
Sign up to set email alerts
|

A review of Convolutional-Neural-Network-based action recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
113
0
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 278 publications
(117 citation statements)
references
References 27 publications
0
113
0
2
Order By: Relevance
“…Human action recognition [1][2][3][4][5][6] is one of the most important research fields in computer vision. Although recognizing the motion of human action in video can provide discriminative clues for classifying one specific action, many human actions (e.g., "Phoning," "InteractingWithComputer," and "Shooting," as shown in Figure 1), can be represented by one single still image [2]. In particular, certain actions (e.g., "Play-ingGuitar," "RidingHorse," and "Running," as shown in Figure 1) may require static cue-based approaches even if those motions in videos are available [2].…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…Human action recognition [1][2][3][4][5][6] is one of the most important research fields in computer vision. Although recognizing the motion of human action in video can provide discriminative clues for classifying one specific action, many human actions (e.g., "Phoning," "InteractingWithComputer," and "Shooting," as shown in Figure 1), can be represented by one single still image [2]. In particular, certain actions (e.g., "Play-ingGuitar," "RidingHorse," and "Running," as shown in Figure 1) may require static cue-based approaches even if those motions in videos are available [2].…”
Section: Introductionmentioning
confidence: 99%
“…Although recognizing the motion of human action in video can provide discriminative clues for classifying one specific action, many human actions (e.g., "Phoning," "InteractingWithComputer," and "Shooting," as shown in Figure 1), can be represented by one single still image [2]. In particular, certain actions (e.g., "Play-ingGuitar," "RidingHorse," and "Running," as shown in Figure 1) may require static cue-based approaches even if those motions in videos are available [2]. To recognize these human actions with video-based approaches mentioned above [5,6,8] may be inappropriate due to their slight action changes without distinguishability.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Video content representation, that is feature extraction, is the core of video action recognition [Yao, Lei and Zhong (2019)]. Then, whether the feature extraction and effective characterization of the video content can be better realized will directly determine the motion recognition effect.…”
Section: Deep Featurementioning
confidence: 99%
“…As said in (G Yao et al, 2019), we have two ways for representation of action recognition; the Handcrafted representation method (Caba Heilbron et al, 2016, Mettes et al, 2015, Yu, 2015, and the Deep Learning representation method (I. Goodfellow, 2016); in the first method, we extract features manually and it is generally used as a baseline to evaluate new Deep Learning representation; whereas, the deep learning representation method learns the trainable features automatically from videos (G Yao et al, 2019). So, talking about automatic human events recognition recently seems that researchers in this area aim to go beyond the human spirit.…”
Section: Introductionmentioning
confidence: 99%