2012
DOI: 10.1073/pnas.1207690109
|View full text |Cite
|
Sign up to set email alerts
|

From simple innate biases to complex visual concepts

Abstract: Early in development, infants learn to solve visual problems that are highly challenging for current computational methods. We present a model that deals with two fundamental problems in which the gap between computational difficulty and infant learning is particularly striking: learning to recognize hands and learning to recognize gaze direction. The model is shown a stream of natural videos and learns without any supervision to detect human hands by appearance and by context, as well as direction of gaze, in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

6
87
0
1

Year Published

2012
2012
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 69 publications
(94 citation statements)
references
References 44 publications
(65 reference statements)
6
87
0
1
Order By: Relevance
“…the infant probably has no concept of objects at this stage (with a few probable exceptions such as faces), and the theory of Ullman et al (8) does not require it. Instead, they demonstrate that simple action events can be detected by local analysis of motion flow patterns.…”
Section: Bootstrapping Visual Conceptsmentioning
confidence: 98%
See 3 more Smart Citations
“…the infant probably has no concept of objects at this stage (with a few probable exceptions such as faces), and the theory of Ullman et al (8) does not require it. Instead, they demonstrate that simple action events can be detected by local analysis of motion flow patterns.…”
Section: Bootstrapping Visual Conceptsmentioning
confidence: 98%
“…In PNAS, Ullman et al (8) suggest a twist to this story by emphasizing the key role of motions that cause actions as innate biases that help bootstrap the learning of complex visual concepts. They address the phenomenon that infants are quick to learn models of hands and estimate the gaze direction of the owner of the hand.…”
Section: Bootstrapping Visual Conceptsmentioning
confidence: 99%
See 2 more Smart Citations
“…In addition to this, computer can predict Observers Task (Kanan, et al, 2014), using a multi-fixation pattern analysis method. For saliency problem, computer need to analysis simple innate biases, and translate it to complex visual concepts (Ullman, et al, 2012). In this process, computer can learn a model to detect human hands and gaze direction.…”
Section: Introductionmentioning
confidence: 99%