2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017
DOI: 10.1109/cvpr.2017.440
|View full text |Cite
|
Sign up to set email alerts
|

Predicting Ground-Level Scene Layout from Aerial Imagery

Abstract: We introduce a novel strategy for learning to extract semantically meaningful features from aerial imagery. Instead of manually labeling the aerial imagery, we propose to predict (noisy) semantic features automatically extracted from co-located ground imagery. Our network architecture takes an aerial image as input, extracts features using a convolutional neural network, and then applies an adaptive transformation to map these features into the ground-level perspective. We use an end-to-end learning approach t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

1
264
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 197 publications
(276 citation statements)
references
References 30 publications
1
264
0
Order By: Relevance
“…This is a more challenging task since different views have little or no overlap. To tackle this problem, Zhai et al [48] try to generate panoramic ground-level images from aerial images of the same location by using a convolutional neural network. Krishna and Ali [34] propose a X-Fork and a X-Seq GAN-based structure to address the aerial to street view image translation task using an extra semantic segmentation map.…”
Section: Related Workmentioning
confidence: 99%
“…This is a more challenging task since different views have little or no overlap. To tackle this problem, Zhai et al [48] try to generate panoramic ground-level images from aerial images of the same location by using a convolutional neural network. Krishna and Ali [34] propose a X-Fork and a X-Seq GAN-based structure to address the aerial to street view image translation task using an extra semantic segmentation map.…”
Section: Related Workmentioning
confidence: 99%
“…Top matches (top 1 -top 5 from left to right) Figure 10: Qualitative Results on CVUSA dataset [46] for aerial-to-ground image matching. Images with green borders are the groundtruth panoramas for the corresponding query images.…”
Section: Synthesized Ground Panoramamentioning
confidence: 99%
“…Regmi and Mubarak Shah Center for Research in Computer Vision, University of Central Florida krishna.regmi7@gmail.com, shah@crcv.ucf.edu T-SNE[27] visualization of aerial and ground image features obtained using the two-stream baseline (left) and the proposed feature fusion method (right) for 100 test images on the CVUSA dataset[46].…”
mentioning
confidence: 99%
“…Other methods geolocalize ground level images using different techniques to match features to corresponding ground-level images database (Lin et al, 2013;Müller-Budack et al, 2018). For example, ground-level imagery can be combined with aerial imagery to accom-plish a pixel-accurate semantic segmentation of the scene, or road respectively (Zhai et al, 2017;Máttyus et al, 2016). Lefèvre et al (2017) detects changes by with siamese CNNs through comparing aerial imagery and street view panoramas warped to aerial image geometry.…”
Section: Related Workmentioning
confidence: 99%