2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.00953
|View full text |Cite
|
Sign up to set email alerts
|

Joint Inductive and Transductive Learning for Video Object Segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
30
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 77 publications
(30 citation statements)
references
References 33 publications
0
30
0
Order By: Relevance
“…(i) Automatic video object segmentation (VOS) segments objects that are visually and/or motion salient in an image sequence [28], [29]. (ii) Semi-automatic VOS relies on an initial labelled frame and subsequently tracks and segments the initialized objects throughout the sequence [30], [31], [32]. (iii) Semantic VS is concerned with segmenting a finite set of semantic categories that are learned during training [20], [33].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…(i) Automatic video object segmentation (VOS) segments objects that are visually and/or motion salient in an image sequence [28], [29]. (ii) Semi-automatic VOS relies on an initial labelled frame and subsequently tracks and segments the initialized objects throughout the sequence [30], [31], [32]. (iii) Semantic VS is concerned with segmenting a finite set of semantic categories that are learned during training [20], [33].…”
Section: Related Workmentioning
confidence: 99%
“…Most closely related to the present work are the following examples. Semi-automatic VOS approaches have used unlabeled frames transductively to enforce temporal continuity [31], [32]. Earlier work on video semantic segmentation applied representation warping to fuse features from consecutive frames to ensure temporal consistency of the predictions in an inductive setting [33].…”
Section: Related Workmentioning
confidence: 99%
“…To address the context limitation, recent state-of-the-art methods use more past frames as feature memory [36,13,64,21,28,58,16]. Particularly, Space-Time Memory (STM) [36] is popular and has been extended by many follow-up works [43,8,18,54,50,31,9,44,33]. Among these extensions, we use STCN [9] as our working memory backbone as it is simple and effective.…”
Section: Related Workmentioning
confidence: 99%
“…AOT [60] is a recent work that extends the attention mechanism to transformers but does not solve the GPU memory explosion problem. Some methods [33,14] Fig. 2.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation