2016
DOI: 10.1177/0278364916679892
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal learning and inference from visual and remotely sensed data

Abstract: Autonomous vehicles are often tasked to explore unseen environments, aiming to acquire and understand large amounts of visual image data and other sensory information. In such scenarios, remote sensing data may be available a priori, and can help to build a semantic model of the environment and plan future autonomous missions. In this paper, we introduce two multimodal learning algorithms to model the relationship between visual images taken by an autonomous underwater vehicle during a survey and remotely sens… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
31
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 33 publications
(31 citation statements)
references
References 42 publications
0
31
0
Order By: Relevance
“…Typically, the approach involves learning a multilayer feature representation of each modality individually, followed by a multimodal layer that captures the correlations between the high-level single-modality features [3][4][5] [6]. Srivastava and Salakhutdinov [4] utilise a Deep Boltzmann Machine to model the relationship between visual images and associated text keywords.…”
Section: Related Work a Multimodal Learningmentioning
confidence: 99%
See 3 more Smart Citations
“…Typically, the approach involves learning a multilayer feature representation of each modality individually, followed by a multimodal layer that captures the correlations between the high-level single-modality features [3][4][5] [6]. Srivastava and Salakhutdinov [4] utilise a Deep Boltzmann Machine to model the relationship between visual images and associated text keywords.…”
Section: Related Work a Multimodal Learningmentioning
confidence: 99%
“…In our recent work [3], we build on these techniques, utilising a gated model [7] for the multimodal layer. The gated model can be framed as a 'mixture' of feature learners, which allows the model to predict, for a given bathymetric feature, the different types of visual features that may be observed.…”
Section: Related Work a Multimodal Learningmentioning
confidence: 99%
See 2 more Smart Citations
“…The current vehicle telematics systems produce big volumes of multimodal data [2], which is predominantly stored in cloud-based datacenters and processed by traditional analytics tools. Consequently, vehicle management is becoming a highly data-centric business.…”
Section: Introductionmentioning
confidence: 99%