2022
DOI: 10.48550/arxiv.2203.14186
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

RSTT: Real-time Spatial Temporal Transformer for Space-Time Video Super-Resolution

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
6
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(6 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…SwinIR [21] used Swin Transformer to handle the image restoration task and proposed residual Swin Transformer blocks. RSTT [35] built a spatial-temporal transformer that naturally incorporates the spatial and temporal super-resolution modules into a single model.…”
Section: Vision Transformermentioning
confidence: 99%
See 2 more Smart Citations
“…SwinIR [21] used Swin Transformer to handle the image restoration task and proposed residual Swin Transformer blocks. RSTT [35] built a spatial-temporal transformer that naturally incorporates the spatial and temporal super-resolution modules into a single model.…”
Section: Vision Transformermentioning
confidence: 99%
“…Referring to [35] , we designed a hierarchical U-shaped Transformer named Real-time Video Frame Interpolation Transformer (RVFIT), which spatially expands input video sequences while considering temporal fluency without dividing the model into temporal and spatial super-resolution modules. This design is superior to previous CNN-based frame interpolation methods because of its parallelism in structure, which can accelerate the inference process based on guaranteed performance.…”
Section: Network Overviewmentioning
confidence: 99%
See 1 more Smart Citation
“…SwinIR [39] using Swin Transformer to handle the image restoration task and proposed residual Swin Transformer blocks. RSTT [40] built a spatial-temporal transformer that naturally incorporates the spatial and temporal super-resolution modules into a single model.…”
Section: Video Transformermentioning
confidence: 99%
“…It also achieves promising results in restoration tasks [11,38,80,43,4,37,18,20,5,89,46,72]. In particular, for video restoration, Cao et al [4] propose the first transformer model for video SR, while Liang et al [37] propose an unified framework for video SR, deblurring and denoising.…”
Section: Vision Transformermentioning
confidence: 99%