2018
DOI: 10.1109/lsp.2017.2778190
|View full text |Cite
|
Sign up to set email alerts
|

PMHI: Proposals From Motion History Images for Temporal Segmentation of Long Uncut Videos

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
11
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
3

Relationship

3
5

Authors

Journals

citations
Cited by 14 publications
(11 citation statements)
references
References 20 publications
0
11
0
Order By: Relevance
“…Schindler and Gool (2008) claimed that from one to seven frames are sufficient to recognise a basic action from a very short video shot. Unlike above, (Murtaza et al, 2018) claimed that the video content could adaptively determine the suitable number of frames, and in a typical video, requires between ten to twenty five frames. We chose ten for window size W similar to Paul et al (2017) algorithm because it reflects the texture and motion features within video shots effectively.…”
Section: Experiments and Results Analysismentioning
confidence: 99%
See 2 more Smart Citations
“…Schindler and Gool (2008) claimed that from one to seven frames are sufficient to recognise a basic action from a very short video shot. Unlike above, (Murtaza et al, 2018) claimed that the video content could adaptively determine the suitable number of frames, and in a typical video, requires between ten to twenty five frames. We chose ten for window size W similar to Paul et al (2017) algorithm because it reflects the texture and motion features within video shots effectively.…”
Section: Experiments and Results Analysismentioning
confidence: 99%
“…The key frames (KFs) extraction algorithm should provide a compact video summarization with less processing time, and preserve the sufficient information in the video with simple implementation (Cao et al, 2012;Murtaza et al, 2018;Paul et al, 2017). From literature reviews, the key frame extraction algorithm starts by detecting the shot change to segment a video to several shots, then extracts the key frames from each shot (Truong & Venkatesh, 2007).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Numerous feature extraction methods have been proposed for HAR using RGB video data, which achieved successful recognition results. Particularly, these methods include 3D gradientbased spatiotemporal descriptor [15], spatiotemporal interest point (STIP) detector [16], motion-energy images (MEIs) and motion history images (MHIs) [17], [18]. The evolution of deep learning schemes, i.e., deep learning based convolutional neural networks (CNN) and Long Short-Term Memory (LSTM) networks, has motivated the researchers to explore its application for action recognition from RGB videos [19]- [22].…”
Section: Introductionmentioning
confidence: 99%
“…Over the last few years, convolutional neural networks (CNNs) have led to improved accuracy of action recognition [5][6][7][8]. However, TAD methods [2,[10][11][12][13][14] still need improvement. In [10], Pyramid of Score Distribution Feature (PSDF) based TAD approach is proposed.…”
Section: Introductionmentioning
confidence: 99%