2022 IEEE International Conference on Development and Learning (ICDL) 2022
DOI: 10.1109/icdl53763.2022.9962190
|View full text |Cite
|
Sign up to set email alerts
|

Toddler-inspired embodied vision for learning object representations

Abstract: Recent time-contrastive learning approaches manage to learn invariant object representations without supervision. This is achieved by mapping successive views of an object onto close-by internal representations. When considering this learning approach as a model of the development of human object recognition, it is important to consider what visual input a toddler would typically observe while interacting with objects. First, human vision is highly foveated, with high resolution only available in the central r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 18 publications
0
1
0
Order By: Relevance
“…In the past, computational models of cognitive development have often been restricted to isolated cognitive phenomena. Some examples are works on binocular vision [63], [64], visual object and category learning [65]- [67], gaze following [68], [69], learning to grasp objects [20], perservative reaching [70], word learning [71]- [73], and countless others. While such models have produced many important insights, they often work with simplified sensory inputs and it is not clear how to scale them to the rich multimodal sensory input provided by our sense organs.…”
Section: Discussionmentioning
confidence: 99%
“…In the past, computational models of cognitive development have often been restricted to isolated cognitive phenomena. Some examples are works on binocular vision [63], [64], visual object and category learning [65]- [67], gaze following [68], [69], learning to grasp objects [20], perservative reaching [70], word learning [71]- [73], and countless others. While such models have produced many important insights, they often work with simplified sensory inputs and it is not clear how to scale them to the rich multimodal sensory input provided by our sense organs.…”
Section: Discussionmentioning
confidence: 99%