Proceedings of the 28th ACM International Conference on Multimedia 2020
DOI: 10.1145/3394171.3413750
|View full text |Cite
|
Sign up to set email alerts
|

Deep Structural Contour Detection

Abstract: Object contour detection is the fundamental and preprocessing step for multimedia applications such as icon generation, object segmentation, and tracking. The quality of contour prediction is of great importance in these applications since it affects the subsequent process. In this work, we aim to develop a high-performance contour detection system. We first propose a novel yet very effective loss function for contour detection. The proposed loss function is capable of penalizing the distance of contour-struct… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
31
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 50 publications
(32 citation statements)
references
References 47 publications
0
31
0
Order By: Relevance
“…LPCB [7] explains why CNNs tend to produce blurry edges and proposes a solution to make CNNs directly predict crisp boundaries without post-processing. DSCD [6] proposes a novel loss function and three different scales of dilated convolution blocks to increase the ability of high-level feature extraction. BANet [10] presents a novel Bidirectional encoder-decoder network for accurate boundary detection.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…LPCB [7] explains why CNNs tend to produce blurry edges and proposes a solution to make CNNs directly predict crisp boundaries without post-processing. DSCD [6] proposes a novel loss function and three different scales of dilated convolution blocks to increase the ability of high-level feature extraction. BANet [10] presents a novel Bidirectional encoder-decoder network for accurate boundary detection.…”
Section: Related Workmentioning
confidence: 99%
“…We follow the prior works [6,7,23,36,37] to use F-score as the evaluation metric. The F-score (also known as F-measure, F1-score) can be obtained by 2 ร— ๐‘ƒ ร— ๐‘…/(๐‘ƒ + ๐‘…) where ๐‘ƒ = ๐‘‡ ๐‘ƒ/(๐‘‡ ๐‘ƒ + ๐น ๐‘ƒ), P, TP, FP denotes the Precision, True Positive and False Positive, respectively; ๐‘… = ๐‘‡ ๐‘ƒ/(๐‘‡ ๐‘ƒ + ๐น ๐‘ ), R and FN denotes the Recall and False Negative.…”
Section: Evaluation Metricsmentioning
confidence: 99%
See 2 more Smart Citations
“…2 (a). Under this inpainting framework, paired training data can be easily obtained by randomly generating masks as in general image inpainting [15,19,22,24,30,32,35,35,37,39,44,45,48,50,51,53,56,59] and extracting edges/contours in the masked regions as surrogate sketches using edge detection algorithms [1,5,6,13,21,25,26,38,43,47,49]. Then the problem can be solved by training an inpainting model to predict the original image given the masked image, mask, and sketch as input.…”
Section: Introductionmentioning
confidence: 99%