Proceedings of the Second International Conference on Automatic Face and Gesture Recognition
DOI: 10.1109/afgr.1996.557249
|View full text |Cite
|
Sign up to set email alerts
|

Understanding manipulation in video

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
20
0

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 38 publications
(20 citation statements)
references
References 2 publications
0
20
0
Order By: Relevance
“…In the 'what is being learned' area, most imitation learning research focuses on learning assembly/pick-and-place operations (Brand, 1997), (Ehrenmann, 2002), (Kang, 1991), (Kuniyoshi, 1994), (Ogata, 1994), (Paul, 1996), (Tung, 1995). Such research generally looks at an input data stream, and attempts to segment the stream to identify actions performed by a human hand.…”
Section: Related Workmentioning
confidence: 99%
“…In the 'what is being learned' area, most imitation learning research focuses on learning assembly/pick-and-place operations (Brand, 1997), (Ehrenmann, 2002), (Kang, 1991), (Kuniyoshi, 1994), (Ogata, 1994), (Paul, 1996), (Tung, 1995). Such research generally looks at an input data stream, and attempts to segment the stream to identify actions performed by a human hand.…”
Section: Related Workmentioning
confidence: 99%
“…Other recent attempts to provide an analysis of video in restricted domains include the work of Mann et al [15] and Siskind et al [19] who propose methods for analyzing the physical interactions between objects in a video sequence, and that of Brand [6] who looks at understanding human actions in video for the purpose of video summarization. Kollnig et al [13] have defined a vocabulary of motion verbs that are used to analyze the behavior of cars in video sequences of traffic scenes.…”
Section: Related Workmentioning
confidence: 99%
“…In 1996 Brand created a blob-oriented 2D vision system that used 6 handcoded networks similar to HMMs to recognize actions Brand called "touching", "putting", "getting", "adding", and "removing", in highly constrained video of human activity [2]. In 2000 Brand and Kettnaker described work on a system that automatically learned HMMs (the states, transitions, and parameter values) from a similar 2D blob oriented input from a well-positione stationary desk camera in an office [3].…”
Section: Machines That Watch From Afarmentioning
confidence: 99%