2019 IEEE/CVF International Conference on Computer Vision (ICCV) 2019
DOI: 10.1109/iccv.2019.00111
|View full text |Cite
|
Sign up to set email alerts
|

SplitNet: Sim2Sim and Task2Task Transfer for Embodied Visual Navigation

Abstract: We propose SplitNet, a method for decoupling visual perception and policy learning. By incorporating auxiliary tasks and selective learning of portions of the model, we explicitly decompose the learning objectives for visual navigation into perceiving the world and acting on that perception. We show dramatic improvements over baseline models on transferring between simulators, an encouraging step towards Sim2Real. Additionally, SplitNet generalizes better to unseen environments from the same simulator and tran… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
67
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 60 publications
(67 citation statements)
references
References 33 publications
0
67
0
Order By: Relevance
“…Finally, there are two really recent articles closely related to ours. The first one by Gordon et al [10] introduced SplitNet on which they explicitly decompose the learning scheme in finding features from perception task and use these features as input to their model-free RL agent. But their scheme is applied to a completely different task, robot navigation and scene exploration.…”
Section: Auxiliary Tasks and Learning Affordancesmentioning
confidence: 99%
“…Finally, there are two really recent articles closely related to ours. The first one by Gordon et al [10] introduced SplitNet on which they explicitly decompose the learning scheme in finding features from perception task and use these features as input to their model-free RL agent. But their scheme is applied to a completely different task, robot navigation and scene exploration.…”
Section: Auxiliary Tasks and Learning Affordancesmentioning
confidence: 99%
“…Thus, we should have pursued a more effective real-to-red.sim transfer. However, as indicated by [24], the separation of advanced visual perception and motion planning is advantageous for domain adaptation. For real-time MAV motion planning, where motion blurring is introduced, candidate applications are limited to non-agile and noncomplex trajectories.…”
Section: Discussionmentioning
confidence: 99%
“…Thus, it is not appropriate to use the existing sim-to-real type datasets and simulators or conventional methods directly for industrial applications in narrow or confined environments. Previous research [24] has handled this gap between real-worldbased simulators. This study has revealed the effectiveness of splitting visual perception and motion control.…”
Section: Sim-to-real Approachesmentioning
confidence: 99%
See 1 more Smart Citation
“…TD-A3C, I2A and Gated-LSTM-A3C are all first trained via behavioral cloning using ground-truth paths. After pre-training, we update the three policy layers using a shaped reward based on the geodesic distance to the goal, geo(x, g), as described in (Gordon et al 2019): r t = geo(x t−1 , g)−geo(x t , g)+ζ,where ζ = −0.01 is a small constant time penalty. More implementation details are provided in the supplemental material.…”
Section: Success Criteriamentioning
confidence: 99%