2018 IEEE Intelligent Vehicles Symposium (IV) 2018
DOI: 10.1109/ivs.2018.8500621
|View full text |Cite
|
Sign up to set email alerts
|

Box2Pix: Single-Shot Instance Segmentation by Assigning Pixels to Object Boxes

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
48
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 71 publications
(49 citation statements)
references
References 22 publications
0
48
0
Order By: Relevance
“…the 3D car center, which provides stronger supervision and helps learn more robust networks. Previously, inducing relative position of object instances has also been shown to be effective in instance segmentation [58,33]. Formally, let…”
Section: A Direct Approachmentioning
confidence: 99%
“…the 3D car center, which provides stronger supervision and helps learn more robust networks. Previously, inducing relative position of object instances has also been shown to be effective in instance segmentation [58,33]. Formally, let…”
Section: A Direct Approachmentioning
confidence: 99%
“…Object masks are then obtained in postprocessing by assigning pixels to cluster centers in this space. The Box2Pix method [23] instead uses supervised learning to map pixels to a specific target space, namely the 2D offsets from the given pixel towards its instance's center, which are found through a bounding box detection branch.…”
Section: Related Workmentioning
confidence: 99%
“…Instance Stixels expand the idea of Semantic Stixels by additionally training a CNN to output a 2D estimation of the position of the instance center for each pixel. This estimation is predicted in image coordinates, as proposed in the Box2Pix method [23]. More specifically, the CNN predicts 2D offsets Ω p ∈ R 2 (i.e.…”
Section: B Instance Stixelsmentioning
confidence: 99%
See 1 more Smart Citation
“…Inspired by recent works [16,22,33] in general semantic instance segmentation, we aim to design a segmentation-based Single-shot Arbitrarily-Shaped Text detector (SAST), which integrates both the high-level object knowledge and low-level pixel information in a single shot and detects scene text of arbitrary shapes with high accuracy and efficiency. Employing a FCN [27] model, various geometric properties of text regions, including text center line (TCL), text border offset (TBO), text center offset (TCO), and text vertex offset (TVO), are designed to learn simultaneously under a multi-task learning formulation.…”
Section: Introductionmentioning
confidence: 99%