2019
DOI: 10.1126/scirobotics.aaw6661
|View full text |Cite
|
Sign up to set email alerts
|

Does computer vision matter for action?

Abstract: Controlled experiments indicate that explicit intermediate representations help action.https://doi.org/10.1126/scirobotics.aaw6661Biological vision systems evolved to support action in physical environments [1, 2]. Action is also a driving inspiration for computer vision research. Problems in computer vision are often motivated by their relevance to robotics and their prospective utility for systems that move and act in the physical world. In contrast, a recent stream of research at the intersection of machine… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
58
1

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3
3

Relationship

1
9

Authors

Journals

citations
Cited by 91 publications
(61 citation statements)
references
References 7 publications
2
58
1
Order By: Relevance
“…D EPTH is among the most useful intermediate representations for action in physical environments [1]. Despite its utility, monocular depth estimation remains a challenging problem that is heavily underconstrained.…”
Section: Introductionmentioning
confidence: 99%
“…D EPTH is among the most useful intermediate representations for action in physical environments [1]. Despite its utility, monocular depth estimation remains a challenging problem that is heavily underconstrained.…”
Section: Introductionmentioning
confidence: 99%
“…Oxford Robotcar [27] was the first real-world large-scale dataset in which adverse visual conditions such as nighttime, rain and snow were significantly represented, but it did not feature semantic annotations. While more recent large-scale sets [2,30] that cover adverse conditions, such as Waymo Open [42] and nuScenes [3], include bounding boxes, they still lack dense pixel-level semantic annotations, which are vital for real-world autonomous agents [63]. BDD100K [55] is the only exception to this rule, with ca.…”
Section: Related Workmentioning
confidence: 99%
“…The field of unsupervised learning has explored different ways to learn state representations [27,28,29] for policy learning. Recently, several works [30,31,32] have studied the benefits of combining various mid-level visual representations for reinforcement learning. Different from previous works, we 1) demonstrate real-world results on manipulation while previous works present simulation results on navigation, 2) highlight the importance of using a pre-trained model for exploration, and 3) propose an initialization strategy without the need of data collection at target domain.…”
Section: Related Workmentioning
confidence: 99%