2019
DOI: 10.1109/access.2019.2940510
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Frame Pyramid Refinement Network for Video Frame Interpolation

Abstract: Video frame interpolation aims at synthesizing new video frames in-between existing frames to generate higher frame rate video. Current methods usually use two adjacent frames to generate intermediate frames, but sometimes fail to handle challenges like large motion, occlusion, and motion blur. This paper proposes a multi-frame pyramid refinement network to effectively use spatio-temporal information contained in multiple frames (more than two). There are three technical contributions in the proposed network. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(8 citation statements)
references
References 44 publications
0
8
0
Order By: Relevance
“…Most state-of-the-art optical flow models use deep learning [25], which suggest that CNNs can understand motion information between frames. In order to get better results, many researchers merge optical flow estimation and video interpolation frame in a single model [2], [6], [7], [26]- [28]. Liu et al [6] designs a deep network with a voxel flow layer to synthesize video frames by flowing pixel values from input video volume.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Most state-of-the-art optical flow models use deep learning [25], which suggest that CNNs can understand motion information between frames. In order to get better results, many researchers merge optical flow estimation and video interpolation frame in a single model [2], [6], [7], [26]- [28]. Liu et al [6] designs a deep network with a voxel flow layer to synthesize video frames by flowing pixel values from input video volume.…”
Section: Related Workmentioning
confidence: 99%
“…Niklaus and Liu [27] apply pixel-wise contextual information extracted by a pre-trained network to estimated bidirectional flow, and uses a frame synthesis network to produce the interpolated frame in a context-aware fashion. Zhang et al [7] uses a 3D U-Net feature extractor to excavate spatio-temporal context and rebuild texture, and a coarse-to-fine architecture to improve optical flows estimation. Li et al [28] proposes a lightweight network to estimate optical flow at feature level and introduce a new sobolev loss achieve better results.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…However, objects often follow complex, non-linear trajectories. To this end, researchers recently focused on leveraging information from more than two neighboring frames [7,8,26,50,53].…”
Section: Introductionmentioning
confidence: 99%
“…Many recently published papers focus on addressing the motion analysis. For example, FI-NET [2] computes optical flow at feature level instead of image level to make the motion estimation more accurate; [3] learns the latent motion features instead of learning the optical flow as the motion feature; [4] and [5] learn from 4 input images instead of 2 images and add some techniques like long short term memory (LSTM) and Multi-Frame Pyramid Refinement to predict the motions.…”
Section: Introductionmentioning
confidence: 99%