Proceedings of the International Conference on Multimedia Information Retrieval 2010
DOI: 10.1145/1743384.1743453
|View full text |Cite
|
Sign up to set email alerts
|

Controlling your TV with gestures

Abstract: Vision-based user interfaces enable natural interaction modalities such as gestures. Such interfaces require computationally intensive video processing at low latency. We demonstrate an application that recognizes gestures to control TV operations. Accurate recognition is achieved by using a new descriptor called MoSIFT, which explicitly encodes optical flow with appearance features. MoSIFT is computationally expensive -a sequential implementation runs 100 times slower than real time. To reduce latency suffici… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
17
0

Year Published

2010
2010
2019
2019

Publication Types

Select...
5
3
2

Relationship

2
8

Authors

Journals

citations
Cited by 39 publications
(17 citation statements)
references
References 21 publications
0
17
0
Order By: Relevance
“…a grab gesture). Chen et al [5] dispensed with the visible cursor and controlled TV channel and volume settings by moving the left or the right arm upwards or downwards. Dezfuli et al [9] assigned interactions to specific regions of the users´ non-dominant hand, triggered when tapped with the dominant hand which facilitates to control the TV blindly.…”
Section: Freehand Gesturesmentioning
confidence: 99%
“…a grab gesture). Chen et al [5] dispensed with the visible cursor and controlled TV channel and volume settings by moving the left or the right arm upwards or downwards. Dezfuli et al [9] assigned interactions to specific regions of the users´ non-dominant hand, triggered when tapped with the dominant hand which facilitates to control the TV blindly.…”
Section: Freehand Gesturesmentioning
confidence: 99%
“…The second application, TV control, provides an interface to control a television via gestures [5]. Each video frame is sent to two separate tasks, face detection and motion extraction, shown in Figure 2.…”
Section: Interactive Perception Applicationsmentioning
confidence: 99%
“…These include a gesturebased gaming system, a natural pointing interface [3], a gesture-based tv control application [10], object pose detection [2], and video action recognition [1] (see Figure 1). Key takeaways from this work are that it is indeed possible to run latency-sensitive interactive perception applications on a cluster of machines, and that there is sufficient coarse-grained parallelism in these applications to make good use of 10s to 100s of processor cores.…”
Section: A Programming Framework For Distributed Interactive Applicamentioning
confidence: 99%