2017 IEEE International Conference on Computer Vision (ICCV) 2017
DOI: 10.1109/iccv.2017.382
|View full text |Cite
|
Sign up to set email alerts
|

Two-Phase Learning for Weakly Supervised Object Localization

Abstract: Weakly supervised semantic segmentation and localization have a problem of focusing only on the most important parts of an image since they use only image-level annotations. In this paper, we solve this problem fundamentally via two-phase learning. Our networks are trained in two steps. In the first step, a conventional fully convolutional network (FCN) is trained to find the most discriminative parts of an image. In the second step, the activations on the most salient parts are suppressed by inference conditi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
108
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 134 publications
(112 citation statements)
references
References 42 publications
0
108
0
Order By: Relevance
“…Feature-level processing can be used to expand the regions activated by a CAM. Adversarial complementary learning [35] and two-phase learning [14] use a classifier to identify the discriminative parts of an object and erase them based on features. A second classifier then is trained to find the complementary parts of the object from those erased features.…”
Section: Feature-level Processingmentioning
confidence: 99%
“…Feature-level processing can be used to expand the regions activated by a CAM. Adversarial complementary learning [35] and two-phase learning [14] use a classifier to identify the discriminative parts of an object and erase them based on features. A second classifier then is trained to find the complementary parts of the object from those erased features.…”
Section: Feature-level Processingmentioning
confidence: 99%
“…Our method achieves mIoU values of 63.9 [45] 49.8 51.2 TransferNet CVPR '16 [11] 52.1 51.2 AISI ECCV '18 [16] 61.3 62. [33] 52.8 53.7 TPL ICCV '17 [22] 53.1 53.8 AE_PSL CVPR '17 [44] 55.0 55.7 DCSP BMVC '17 [2] 58.6 59.2 MEFF CVPR '18 [9] -55.6 GAIN CVPR '18 [26] 55.3 56.8 MCOF CVPR '18 [43] 56.2 57.6 AffinityNet CVPR '18 [1] 58.4 60.5 DSRG CVPR '18 [17] 59.0 60.4 MDC CVPR '18 [46] 60.4 60.8 SeeNet NIPS '18 [15] 61.1 60.7 FickleNet CVPR '19 [24] 61.2 61.9 Ours 63.9 65.0 and 65.0 for PASCAL VOC 2012 validation and test images respectively, which is 94.4% of that of DeepLab [3], trained with fully annotated data, which achieved an mIoU of 67.6 on validation images. Our method is 3.1% better on test images than the best method which uses only image-level annotations for supervision.…”
Section: Results On Image Segmentationmentioning
confidence: 99%
“…That means, there is no parameter overheads even when applied to multiple feature maps at the same time. Furthermore, with ADL, the most discriminative region can be identified and erased efficiently, without auxiliary classifiers [17,59], re-training [49], or additional forward-backward propagation [20].…”
Section: Adl: Attention-based Dropout Layermentioning
confidence: 99%