2020
DOI: 10.48550/arxiv.2003.00824
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Multi-Scale Representation Learning for Spatial Feature Distributions using Grid Cells

Gengchen Mai,
Krzysztof Janowicz,
Bo Yan
et al.

Abstract: Unsupervised text encoding models have recently fueled substantial progress in Natural Language Processing (NLP). The key idea is to use neural networks to convert words in texts to vector space representations (embeddings) based on word positions in a sentence and their contexts, which are suitable for end-to-end training of downstream tasks. We see a strikingly similar situation in spatial analysis, which focuses on incorporating both absolute positions and spatial contexts of geographic objects such as Poin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(14 citation statements)
references
References 17 publications
0
10
0
Order By: Relevance
“…Different backbones were used in related work for machine learning of geospatial vector data. In the case of [2], a multilayer perceptron (MLP) was used, [28] used an attention-based network (a Transformer), and [29] used a graph convolutional neural network (GCNN) to learn a feature representation of geo-object geometries. While GCNN have gained in popularity within the domain of GIScience, they come with two disadvantages.…”
Section: Lulc Classifications With Geospatial Vector Datamentioning
confidence: 99%
“…Different backbones were used in related work for machine learning of geospatial vector data. In the case of [2], a multilayer perceptron (MLP) was used, [28] used an attention-based network (a Transformer), and [29] used a graph convolutional neural network (GCNN) to learn a feature representation of geo-object geometries. While GCNN have gained in popularity within the domain of GIScience, they come with two disadvantages.…”
Section: Lulc Classifications With Geospatial Vector Datamentioning
confidence: 99%
“…Then we summarize in fine-grained works that incorporate [12,18,26]. (b) Concatenation: Concatenate the image features with the multimodal features in a channel-by-channel fashion [29,35,37,38]. (c) Addition: Add both predictions from the last layer of the image and extra information for a joint prediction [8].…”
Section: Related Workmentioning
confidence: 99%
“…Methods using Additional Information: Besides visual information, researchers have included additional information to improve classification performance. Many existing works [29,31,35,37,38] combine the image feature with the additional multimodal feature directly through channelwise concatenation. [37] is the first to introduce multimodal features, e.g., images, ages, and dates, extracted from MLP backbone network by concatenating them together to make a joint prediction.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The training labels are derived from large-scale geotagged documents such as tweets, check-ins, and images that are available from social sharing platforms. Mai et al [11] proposed Space2vec using an encoder-decoder framework to encode the absolute positions and spatial relationships of places based on POI information. But for large-scale location embedding studies, labels for all locations are expensive to get and the attributes are difficult to define because some locations may have multiple labels and unpopular locations have no label.…”
Section: Related Workmentioning
confidence: 99%