2021
DOI: 10.1016/j.patrec.2021.04.002
|View full text |Cite
|
Sign up to set email alerts
|

A novel feature extractor for human action recognition in visual question answering

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 15 publications
0
6
0
Order By: Relevance
“…Reference [18] focused on contextual abnormal human behavior detection especially in video surveillance applications. Reference [19] proposed a new approach to human action recognition for visual question answering, using a novel feature extractor, multiperson 2D pose estimation, and machine learning technique. Reference [20] proposed a unique attention-based pipeline for human action recognition, utilizing both the spatial and the temporal features from a sequence of frames.…”
Section: Security-related Classification and Analyticsmentioning
confidence: 99%
“…Reference [18] focused on contextual abnormal human behavior detection especially in video surveillance applications. Reference [19] proposed a new approach to human action recognition for visual question answering, using a novel feature extractor, multiperson 2D pose estimation, and machine learning technique. Reference [20] proposed a unique attention-based pipeline for human action recognition, utilizing both the spatial and the temporal features from a sequence of frames.…”
Section: Security-related Classification and Analyticsmentioning
confidence: 99%
“…Holanda et al [3] proposed a rapid human action recognition method using visual question-answering techniques, processing a 2D image frame in a mere 190 milliseconds. Their approach involved an ensemble of classifiers, including Bayesian, Multinomial, Complement, and Gaussian classifiers.…”
Section: Vision-based Harmentioning
confidence: 99%
“…Most State-of-the-Art boxing punch classification methods use IMU sensor data to classify the punches. Some studies use visual data [3,5,6,7,8,9] instead, but either need costly depth measuring sensors or could be faster due to naive trajectory feature extraction. Also, those models use heavy Deep Learning classifiers, which require vast amounts of data, extensive for portable devices.…”
Section: Introductionmentioning
confidence: 99%
“…Run the timing model identification. This paper uses a fixed feature extraction network and an action classification loss function to explore the performance of four common time series modeling methods on this task, including temporal convolution, temporal clustering, temporal attention mechanism, recurrent neural network and its variant models, etc [13][14]. Through research and analysis of 11 time series models, fair and comprehensive performance comparison analysis data is provided for research in this field.…”
Section: Related Workmentioning
confidence: 99%