2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.01246
|View full text |Cite
|
Sign up to set email alerts
|

Zoom-In-To-Check: Boosting Video Interpolation via Instance-Level Discrimination

Abstract: We propose a light-weight video frame interpolation algorithm. Our key innovation is an instance-level supervision that allows information to be learned from the highresolution version of similar objects. Our experiment shows that the proposed method can generate state-of-the-art results across different datasets, with fractional computation resources (time and memory) of competing methods.Given two image frames, a cascade network creates an intermediate frame with 1) a flow-warping module that computes coarse… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
30
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
2

Relationship

1
8

Authors

Journals

citations
Cited by 31 publications
(30 citation statements)
references
References 30 publications
0
30
0
Order By: Relevance
“…Sun et al (Sun et al 2019) select and weight synthetic pixels which are similar with real ones for learning semantic segmentation. Yuan et al (Yuan et al 2019) introduce an instance-level adversarial loss for video frame interpolation problem. Shen et al (Shen et al 2019) propose an instanceaware image-to-image translation framework.…”
Section: Related Workmentioning
confidence: 99%
“…Sun et al (Sun et al 2019) select and weight synthetic pixels which are similar with real ones for learning semantic segmentation. Yuan et al (Yuan et al 2019) introduce an instance-level adversarial loss for video frame interpolation problem. Shen et al (Shen et al 2019) propose an instanceaware image-to-image translation framework.…”
Section: Related Workmentioning
confidence: 99%
“…DAIN (Bao et al 2019) used PWC-Net (Sun et al 2018) and a depth network (Chen et al 2016) to explicitly detect the occlusion. Other interpolation methods such as CtxSyn (Niklaus and Liu 2018) and Zoom-In-to-Check (Yuan et al 2019) warped not only input frames, but also their deep corresponding features. Despite the so far mentioned approaches have been shown effective, the performance can be limited provided that the optical flow or occlusion masks were less accurate.…”
Section: Flow Based Methodsmentioning
confidence: 99%
“…Compared with Sepconv (Niklaus, Mai, and Liu 2017b), our method considers more relevant pixels far away from local grid (black rectangle) with a much smaller kernel size and performs better. estimating flow information together with occlusion masks or visibility maps with deep convolutional neural networks (CNNs) (Jiang et al 2018;Bao et al 2019;2018a;Liu et al 2017;van Amersfoort et al 2017;Liu et al 2019;Xue et al 2019;Peleg et al 2019;Yuan et al 2019;Hannemose et al 2019).…”
Section: Introductionmentioning
confidence: 99%
“…Toward more sophisticated motion modeling, several novel approaches [19,65,3,45], as well as higher-order representations, are proposed. Quadratic [62,23] and cubic [8] flows are estimated from multiple input frames.…”
Section: Related Workmentioning
confidence: 99%