2023
DOI: 10.1109/tpami.2022.3165153
|View full text |Cite
|
Sign up to set email alerts
|

PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
261
0
3

Year Published

2023
2023
2023
2023

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 296 publications
(404 citation statements)
references
References 28 publications
0
261
0
3
Order By: Relevance
“…From the perspective of models, Shi et al (2015) proposed Convolutional Long Short-Term Memory (Con-vLSTM), a spatiotemporal-forecasting neural network model, as a first attempt for DL-based precipitation nowcasting. This work is then followed by a number of studies (Shi et al, 2017;Sønderby et al, 2020;Wang et al, 2017). However, since the structure of ConvLSTM is relatively fixed, current ConvLSTM-based nowcasting methods hardly have specially designed structure for multiple input information.…”
mentioning
confidence: 99%
“…From the perspective of models, Shi et al (2015) proposed Convolutional Long Short-Term Memory (Con-vLSTM), a spatiotemporal-forecasting neural network model, as a first attempt for DL-based precipitation nowcasting. This work is then followed by a number of studies (Shi et al, 2017;Sønderby et al, 2020;Wang et al, 2017). However, since the structure of ConvLSTM is relatively fixed, current ConvLSTM-based nowcasting methods hardly have specially designed structure for multiple input information.…”
mentioning
confidence: 99%
“…Similarly, Ballas et al [11] also employed convolutional layers to the Gated Recurrent Units (GRUs) for video prediction. However, Wang et al [14] held the idea that the temporal information and the spatial information should be equally considered, and proposed a spatial module for ConvLSTMs (ST-LSTM) to help model the spatial representation for each frame. Then, they further proposed the Casual LSTM [15] to help increase the model depth in the temporal domain and Gradient Highway Unit to alleviate the gradient propagation difficulties in deep predictive models.…”
Section: Related Workmentioning
confidence: 99%
“…Moving MNIST Method SSIM/frame ↑ MSE/frame ↓ ConvLSTM (NeurIPS2015) [8] 0.707 103.3 FRNN (ECCV2018) [9] 0.819 68.4 VPN (ICML2017) [48] 0.870 70.0 PredRNN (NeurIPS2017) [14] 0.869 56.8 PredRNN++ (ICML2018) [15] 0.898 46.5 MIM (CVPR2019) [16] 0.910 44.2 E3D-LSTM (ICLR2019) [21] 0.910 41.3 CrevNet (ICLR2020) [17] 0.928 38.5 MAU (NeurIPS2021) [39] 0.931 29.5 STAU 0.939 27.1…”
Section: Moving Mnistmentioning
confidence: 99%
See 1 more Smart Citation
“…A straightforward deep learning solution to visual control problems is to learn action-conditioned video prediction models [38,14,8,53] and then perform Monte-Carlo importance sampling and optimization algorithms, such as the cross-entropy methods, over available behaviors [15,12,29]. Hot topics in video prediction mainly includes long-term and high-fidelity future frames generation [44,43,51,5,52,50,54,41,40,36,56,28,2], dynamics uncertainty modeling [1,10,48,31,7,16,55], object-centric scene decomposition [47,27,18,58,3], and space-time disentanglement [49,27,19,6]. The corresponding technical improvements mainly involve the use of more effective neural architectures, novel probabilistic modeling methods, and specific forms of video representation.…”
Section: Related Workmentioning
confidence: 99%