2018
DOI: 10.18178/ijmlc.2018.8.3.699
|View full text |Cite
|
Sign up to set email alerts
|

Local Feature Extraction from RGB and Depth Videos for Human Action Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(5 citation statements)
references
References 22 publications
0
5
0
Order By: Relevance
“…Rawya et al [46] MSR-Daily Activity 3D dataset and Online RGBD action dataset Spatio-temporal features were extracted using a Bag-of-Features (BoF) approach. Points of interest were detected, and motion history images were created to perform this research work.…”
Section: Rgb-d Datasets Methodology Classification Resultsmentioning
confidence: 99%
“…Rawya et al [46] MSR-Daily Activity 3D dataset and Online RGBD action dataset Spatio-temporal features were extracted using a Bag-of-Features (BoF) approach. Points of interest were detected, and motion history images were created to perform this research work.…”
Section: Rgb-d Datasets Methodology Classification Resultsmentioning
confidence: 99%
“…The Expectation step (E-step) computes the posterior probability of each data point belonging to each component the GMM, given the current estimates of the parameters. Mathematically, this is given by: 𝑃(𝑧 𝑖 = π‘˜|π‘₯ 𝑖 , πœƒ) = πœ‹ π‘˜ * 𝑁(π‘₯ 𝑖 |πœ‡ π‘˜ ,𝛴 π‘˜ ) 𝛴 𝑗 (πœ‹ 𝑗 * 𝑁(π‘₯ 𝑖 |πœ‡ 𝑗 ,𝛴 𝑗 )) (8) where zi is the latent variable indicating the elliptical component assignment for data point xi, ΞΈ is the set of all parameters (Ο€, ΞΌ, Ξ£) for the GMM, and N(xi | ΞΌk, Ξ£k) is the probability density function of the normal distribution for k th ellipse. The Maximization step (M-step) updates the estimates of the parameters by maximizing the expected complete data log-likelihood, given the posterior probabilities computed in the E-step.…”
Section: ) Gmm-em-based Elliptical Modelingmentioning
confidence: 99%
“…Working with drone cameras is more challenging as compared to grounded cameras because in the former case, the background is not static and there is more chance of noise induction into the data. Many previous works have focused on recognizing actions captured by conventional Red-Green-Blue (RGB) video cameras, e.g., [8], and [9]. However, these works have limitations such as coping with various lighting conditions and cluttered backgrounds because RGB data suffer from these variations.…”
Section: Introductionmentioning
confidence: 99%
“…Rawya Al-Akam and Dietrich Paulus investigated a new method for detecting human activities in 3D movies using RGB and depth data. The method suggested in [13] involves using Bag-of-Information techniques to extract local-spatial temporal features from all frames of a video and to distinguish between human activities. To achieve this, K-means clustering and multi-class Support Vector Machines are utilized for classification, and the system is designed to be invariant to scale, rotation, and lighting.…”
Section: Page 201mentioning
confidence: 99%