2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00096
|View full text |Cite
|
Sign up to set email alerts
|

SCOPS: Self-Supervised Co-Part Segmentation

Abstract: Figure 1: Robustness to variations. Sample part segmentation obtained by SCOPS on different types of image collections: (left) unaligned faces from CelebA [29], (middle) birds from CUB [44] and (right) horses from PASCAL VOC [11] dataset images, showing that SCOPS can be robust to appearance, viewpoint and pose variations. AbstractParts provide a good intermediate representation of objects that is robust with respect to the camera, pose and appearance variations. Existing works on part segmentation is dominate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
159
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 127 publications
(160 citation statements)
references
References 47 publications
1
159
0
Order By: Relevance
“…Existing datasets [43,88] are relatively small in size, and only provide sparse correspondence ground truths since manually annotating dense ones is prohibitive. In light of this challenge, weakly supervised semantic correspondence are proposed to learn correspondence without correspondence ground truths [25][26][27][28][29][30]. In addition, existing benchmarks and methods have predominantly focused on "objectcentric" scenarios where each image is occupied by a major object.…”
Section: Finding Correspondencementioning
confidence: 99%
See 1 more Smart Citation
“…Existing datasets [43,88] are relatively small in size, and only provide sparse correspondence ground truths since manually annotating dense ones is prohibitive. In light of this challenge, weakly supervised semantic correspondence are proposed to learn correspondence without correspondence ground truths [25][26][27][28][29][30]. In addition, existing benchmarks and methods have predominantly focused on "objectcentric" scenarios where each image is occupied by a major object.…”
Section: Finding Correspondencementioning
confidence: 99%
“…Even though the advantage of learning correspondences and instance segmentation jointly is clear, many state of the art methods do not make use of this approach due to the lack of large scale datasets with both masks and correspondences. To overcome this challenge, weakly supervised methods have been recently introduced to relax the need for costly supervision in both tasks [25][26][27][28][29][30][46][47][48][49]. Our work is aligned with these efforts as we aim to address instance segmentation and semantic correspondence jointly with inexpensive bounding box supervision.…”
Section: Introductionmentioning
confidence: 99%
“…Mask Concentration Loss: In order to promote compactness for object masks, we use a geometric concentration loss as in [35]. The tuple (x, y) in Eq.…”
Section: Lossmentioning
confidence: 99%
“…Learning discriminative image representation in an unsupervised/ self-supervised manner has attracted increasing interest (Agrawal, Carreira, and Malik 2015;Doersch, Gupta, and Efros 2015;Xie et al 2021), for it gets rid of the costly manually-labeled data and achieves promising performance on many down-stream tasks (Larsson et al 2019;Hung et al 2019;Doersch and Zisserman 2017). These methods generally design pretext tasks and learn the representation from the label generated by the tasks, such as rotation predicting (Komodakis and Gidaris 2018), jigsaw (Noroozi and Favaro 2016;Kim et al 2018), in-painting (Pathak et al 2016), colorization (Zhang, Isola, and Efros 2016;Larsson, Maire, and Shakhnarovich 2017) and clustering (Noroozi et al 2018;Caron et al 2018).…”
Section: Introductionmentioning
confidence: 99%