2020
DOI: 10.48550/arxiv.2010.03957
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Transformers for Modeling Physical Systems

Nicholas Geneva,
Nicholas Zabaras

Abstract: Transformers are widely used in neural language processing due to their ability to model longer-term dependencies in text. Although these models achieve state-ofthe-art performance for many language related tasks, their applicability outside of the neural language processing field has been minimal. In this work, we propose the use of transformer models for the prediction of dynamical systems representative of physical phenomena. The use of Koopman based embeddings provide a unique and powerful method for proje… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 20 publications
0
5
0
Order By: Relevance
“…However, applying sequence models to predict high-dimensional systems remains a challenge due to their high memory overhead. Dimensionality reduction techniques, such as CNN autoencoders [33,32,26,22,29,16,11,27], POD [44,48,5,31,18,8,47,10], or Koopman operators [24,9,14] can be used to construct a lowdimensional latent space. The auto-regressive sequence model then operates on these linear (POD modes) or nonlinear (CNNs) latents.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…However, applying sequence models to predict high-dimensional systems remains a challenge due to their high memory overhead. Dimensionality reduction techniques, such as CNN autoencoders [33,32,26,22,29,16,11,27], POD [44,48,5,31,18,8,47,10], or Koopman operators [24,9,14] can be used to construct a lowdimensional latent space. The auto-regressive sequence model then operates on these linear (POD modes) or nonlinear (CNNs) latents.…”
Section: Related Workmentioning
confidence: 99%
“…The two trajectories from two different Reynolds numbers are hard to distinguish, which makes it challenging for the learned model to capture parameter variations. Although CNN-based embedding methods [14,50,29] are also nonlinear, they cannot handle irregular geometries with unstructured meshes due to the limitation of classic convolution operations. Using pivotal nodes combined with GNN learning, the proposed model is both flexible in dealing data with irregular meshes and effective in capturing state transitions in the system.…”
Section: Datasetsmentioning
confidence: 99%
“…[16] proposed a Spatio-temporal transformer for 3D human motion modeling by learning the evolution of skeleton joints embeddings through space and time. Also, [23] proposed the use of transformer models for the prediction of dynamical systems representative of physical phenomena. Recently, [24] and [25] applied a Spatio-temporal transformer for video action recognition.…”
Section: Related Workmentioning
confidence: 99%
“…In the past, convolutional neural network architectures have been used to capture the spatial characteristics of complex systems [7]. A temporal dimension to these complex systems was introduced by firstly using an auto-regressive network [5,8] and later by using recurrent neural networks(RNN) where the time dependencies are maintained through the introduction of additional parameters [9,10]. A network that combines the convolutional network and RNN is ConvLSTM [1].…”
Section: Introductionmentioning
confidence: 99%