2016
DOI: 10.48550/arxiv.1603.08695
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning to Refine Object Segments

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(21 citation statements)
references
References 0 publications
0
21
0
Order By: Relevance
“…MCNs [7] category predicted pixels via the result of object detection. In DeepMask [16] and SharpMask [17] two branches for segmentation and object score are engaged. Different form DeepMask, InstanceFCN [18] exploits local coherence rather than high-dimensional features to confirm instances.…”
Section: A Instance Segmentationmentioning
confidence: 99%
“…MCNs [7] category predicted pixels via the result of object detection. In DeepMask [16] and SharpMask [17] two branches for segmentation and object score are engaged. Different form DeepMask, InstanceFCN [18] exploits local coherence rather than high-dimensional features to confirm instances.…”
Section: A Instance Segmentationmentioning
confidence: 99%
“…Each extra layer has 3 × 3 kernels and 64 channels which play a role in dimension reduction. Inspired by [40], we integrate a "refinement module" R to invert the effect of each pooling layer and double the resolution of its input feature map if necessary. As shown in Figure 3, the refinement stream consists of five stacked refinement modules, each of which corresponds to one pooling layer in the backbone network.…”
Section: Refined Vgg Networkmentioning
confidence: 99%
“…In this paper, we develop a pixel-wise and end-to-end contour detection system, Top-Down Convolutional Encoder-Decoder Network (TD-CEDN), which is inspired by the success of Fully Convolutional Networks (FCN) [23], HED, Encoder-Decoder networks [24], [25], [13] and the bottomup/top-down architecture [26]. Being fully convolutional, the developed TD-CEDN can operate on an arbitrary image size and the encoder-decoder network emphasizes its symmetric structure which is similar to the SegNet [25] and DeconvNet [24] but not the same, as shown in Fig.…”
Section: Raw Imagementioning
confidence: 99%
“…SegNet [25] used the max pooling indices to upsample (without learning) the feature maps and convolved with a trainable decoder network. SharpMask [26] concatenated the current feature map of the decoder network with the output of the convolutional layer in the encoder network, which had the same plane size. Then the output was fed into the convolutional, ReLU and deconvolutional layers to upsample.…”
Section: B Refined Modulementioning
confidence: 99%
See 1 more Smart Citation