2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00209
|View full text |Cite
|
Sign up to set email alerts
|

VideoINR: Learning Video Implicit Neural Representation for Continuous Space-Time Super-Resolution

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 46 publications
(21 citation statements)
references
References 37 publications
0
21
0
Order By: Relevance
“…Some unsupervised VFI models have been proposed in recent years, for example, Reda et al [28] developed a unsupervised VFI model based on cycle consistency. A more recent optical flow-based CNN model, VideoINR [24], successfully utilizes the implicit neural representation for continuous VFI.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Some unsupervised VFI models have been proposed in recent years, for example, Reda et al [28] developed a unsupervised VFI model based on cycle consistency. A more recent optical flow-based CNN model, VideoINR [24], successfully utilizes the implicit neural representation for continuous VFI.…”
Section: Related Workmentioning
confidence: 99%
“…Most of the previous models can only predict video frames with a fixed frame rate, which is defined by the training dataset. The only exceptions are the aforementioned Vid-ODE [16] and VideoINR [24]. VideoINR can only perform VFI and it does not satisfy the properties of NPs.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Note that the FVD scores do not degrade even at higher spatial resolutions. Additionally, we compare INR-V with existing SOTA superresolution techniques Chen et al (2022) in Table 7 on 2048 videos randomly sampled from the RainbowJelly dataset. As can be seen, INR-V performs comparably with methods on the task of superresolution.…”
Section: A5 Inferring At Multiple Resolutions and Multiple Lengthsmentioning
confidence: 99%
“…Specifically, TMNet [4] implements a temporal modulation network to interpolate arbitrary intermediate frame(s), but this kernel-based motion estimation method often results in serious temporal inconsistencies when dealing with large motions. The work most relevant to our approach is USTVSR [5] and VideoINR [6]. Both these two methods can modulate the input video to arbitrary resolution and frame rate, but they only consider the LR input within two neighboring video frames, and information from a long distance is ignored, which severely limiting their performance.…”
Section: Introductionmentioning
confidence: 99%