2020
DOI: 10.48550/arxiv.2001.02908
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Spatial-Temporal Transformer Networks for Traffic Flow Forecasting

Mingxing Xu,
Wenrui Dai,
Chunmiao Liu
et al.
Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
102
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 72 publications
(102 citation statements)
references
References 0 publications
0
102
0
Order By: Relevance
“…They have to stack more layers to learn dependency if the distance between two regions in the graph is long, and it ends up with an inefficient and oversmoothing model [31]. In recent years, Transformer [10] has been applied for traffic forecasting [11]- [14]. The canonical Transformer can be seen as a special graph neural network with a complete graph, in which any pair of regions are connected.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…They have to stack more layers to learn dependency if the distance between two regions in the graph is long, and it ends up with an inefficient and oversmoothing model [31]. In recent years, Transformer [10] has been applied for traffic forecasting [11]- [14]. The canonical Transformer can be seen as a special graph neural network with a complete graph, in which any pair of regions are connected.…”
Section: Related Workmentioning
confidence: 99%
“…commendable, most works consider spatial and temporal dependency separately with independent processing modules [4]- [9], which can hardly capture their joint nature in our current setting. Some recent works apply canonical Transformer [10] to capture region dependency [11]- [14]. While impressive, canonical Transformer limits the training efficiency because it learns a region's embedding as the weighted aggregation of all the other regions based on their computed attention scores.…”
Section: Introductionmentioning
confidence: 99%
“…Similarly, Guo et al [27] proposed GATCN which deals with the spatial feature by graph attention network and the temporal feature by temporal convolutional network. Xu et al [28] proposed a spatial-temporal transformer network (STTN) composed of graph neural networks and transformer layers [29] to dynamically model various scales of spatial dependencies and capture long-range temporal dependencies. Lu et al [30] presented a temporal directed attributed graph to model complex traffic flow and then employed a message-passing mechanism and several variants of LSTMs to model spatial-temporal dependencies.…”
Section: Related Workmentioning
confidence: 99%
“…The power of transformer network is due to its self-attention mechanism, which can model the complex correlations between objects. Recently, some works utilized transformer based methods to extract the spatiotemporal dependencies in traffic tasks, such as trajectory prediction [33], traffic flow forecasting [32] and achieved state-of-the-art results. To our best knowledge, there is no work utilizing transformer based methods to solve the dynamic spatial correlations in OD Matrix prediction.…”
Section: 31mentioning
confidence: 99%