IEEE International Conference on Image Processing 2005 2005
DOI: 10.1109/icip.2005.1530618
|View full text |Cite
|
Sign up to set email alerts
|

Gesture-based video summarization

Abstract: A novel method for summarizing videos of gestures is presented. The gestures performed by the hands and the head are extracted through skin color segmentation and represented through Zernike moments. The gesture energy is calculated using the norms of the Zernike moments and monitored through time for local minima and maxima that indicate distinctive visual events and thus key-frames. The proposed scheme is not thresholddependent and therefore the number of extracted key-frames varies according to the complexi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2012
2012
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 19 publications
0
5
0
Order By: Relevance
“…The inference of meaning from low-level video cues is non-trivial and thus many approaches are specifc to domains such as sports, music, or news. Approaches for extracting key frames from a video include detecting gestures [34,35], color [25], motion activities [12,20] and specifc events, e.g., goals in sports or diferences between frames in surveillance videos [16]. Internal approaches are challenging for surgical videos which generally do not contain any easily detectable shot changes, events nor diferences in color.…”
Section: Video Summarizationmentioning
confidence: 99%
“…The inference of meaning from low-level video cues is non-trivial and thus many approaches are specifc to domains such as sports, music, or news. Approaches for extracting key frames from a video include detecting gestures [34,35], color [25], motion activities [12,20] and specifc events, e.g., goals in sports or diferences between frames in surveillance videos [16]. Internal approaches are challenging for surgical videos which generally do not contain any easily detectable shot changes, events nor diferences in color.…”
Section: Video Summarizationmentioning
confidence: 99%
“…Within the field of gesture classification, several studies have previously used optical datasets and ML, using peripherals such as the Leap Motion and Kinect [ 19 ]. Other studies have used images and videos to classify gestures [ 20 ]. Other studies in this area have compared ML algorithms using the Kinect/optical peripherals [ 21 ] in order to find the most efficient model.…”
Section: Literature Reviewmentioning
confidence: 99%
“…In recent years, hand gesture recognition [1] has attracted a growing interest due to its applications in many different fields, such as human-computer interaction, robotics, computer gaming, automatic sign-language interpretation and so on. The problem was originally tackled by the computer vision community by means of images and videos [1,2]. More recently the introduction of low cost consumer depth cameras, like Time-Of-Flight cameras and Microsoft's Kinect TM [3], has opened the way to several different approaches that exploit the depth information acquired by these devices for improving gesture recognition performance.…”
Section: Introductionmentioning
confidence: 99%