Proceedings of the 30th International Conference on Advances in Geographic Information Systems 2022
DOI: 10.1145/3557915.3561043
|View full text |Cite
|
Sign up to set email alerts
|

Towards a foundation model for geospatial artificial intelligence (vision paper)

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 28 publications
(9 citation statements)
references
References 5 publications
0
6
0
Order By: Relevance
“…(2) To improve model generalizability across space and time, instead of using the FSTL method as Li et al (2022) did, can we directly learn a hypernetwork to simulate how the model's parameters change based on the location and time with meta‐learning method (Tenzer et al, 2022)? (3) Given the increasing popularity of foundation models (FMs) in the natural language and vision communities, such as GPT‐3 (Brown et al, 2020), CLIP (Radford et al, 2021), PaLM (Wei et al, 2022), and DALL∙E2 (Ramesh et al, 2022), could we build a FM for GeoAI which, after pretraining, can be easily adapted to multiple symbolic GeoAI and subsymbolic GeoAI tasks, involving the use of different data modalities (Mai et al, 2022a)? (4) How can we address important ethical aspects, such as better accounting for and mitigating issues of bias, fairness, and transparency (Shin & Basiri, 2022; Zheng & Sieber, 2022), how to reduce the environmental footprint of model training, and how to better connect to communities studying ethics of technology (Goodchild et al, 2022).…”
Section: Conclusion and Next Stepsmentioning
confidence: 99%
“…(2) To improve model generalizability across space and time, instead of using the FSTL method as Li et al (2022) did, can we directly learn a hypernetwork to simulate how the model's parameters change based on the location and time with meta‐learning method (Tenzer et al, 2022)? (3) Given the increasing popularity of foundation models (FMs) in the natural language and vision communities, such as GPT‐3 (Brown et al, 2020), CLIP (Radford et al, 2021), PaLM (Wei et al, 2022), and DALL∙E2 (Ramesh et al, 2022), could we build a FM for GeoAI which, after pretraining, can be easily adapted to multiple symbolic GeoAI and subsymbolic GeoAI tasks, involving the use of different data modalities (Mai et al, 2022a)? (4) How can we address important ethical aspects, such as better accounting for and mitigating issues of bias, fairness, and transparency (Shin & Basiri, 2022; Zheng & Sieber, 2022), how to reduce the environmental footprint of model training, and how to better connect to communities studying ethics of technology (Goodchild et al, 2022).…”
Section: Conclusion and Next Stepsmentioning
confidence: 99%
“…This success in multimodality is significant for the next generation of GeoAI models that could also be pre-trained with geo-data ranging from location descriptions to remote sensing and street-level images, and from vector data to cartographic maps. However, such models would still suffer from a lack of geographic diversity when learning latent spatial representations in a task-agnostic manner (Mai et al, 2022). Additionally, more and more geo-data could become generated by machines at scale.…”
Section: Introductionmentioning
confidence: 99%
“…LLMs, such as Bidirectional Encoder Representations from Transformers (BERT) (Devlin et al 2019) and Generative Pre-trained Transformer (GPT) models (Brown et al 2020;Ouyang et al 2022), are pre-trained on large-scale textual data (e.g., all text on the Internet) in a task-agnostic manner, and can be adapted to domainspecific tasks via fine tuning, few-shot learning, or sometimes even zero-shot learning. Given their foundational roles in completing various domain-specific tasks, LLMs and other large-scale pretrained models are also called foundation models (Bommasani et al 2021;Mai et al 2022).…”
Section: Introductionmentioning
confidence: 99%