2020
DOI: 10.1007/978-3-030-58598-3_3
|View full text |Cite
|
Sign up to set email alerts
|

Many-Shot from Low-Shot: Learning to Annotate Using Mixed Supervision for Object Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 25 publications
0
3
0
Order By: Relevance
“…The paper also proposed criteria for gathering future datasets. Biffi et al discussed the challenges of object detection relying on time-consuming manual annotations [146] They introduced an online annotation module (OAM) that generates reliable annotations from weakly labeled images. This OAM can enhance the performance of Fast(er) R-CNN, improving mAP by 17% and AP50 by 9% on PASCAL VOC 2007 and MS-COCO benchmarks, outperforming other methods using mixed supervision.…”
Section: A Annotating Training Datamentioning
confidence: 99%
“…The paper also proposed criteria for gathering future datasets. Biffi et al discussed the challenges of object detection relying on time-consuming manual annotations [146] They introduced an online annotation module (OAM) that generates reliable annotations from weakly labeled images. This OAM can enhance the performance of Fast(er) R-CNN, improving mAP by 17% and AP50 by 9% on PASCAL VOC 2007 and MS-COCO benchmarks, outperforming other methods using mixed supervision.…”
Section: A Annotating Training Datamentioning
confidence: 99%
“…This, however, brings new challenges and solutions in DL from multi-modal data, which leads to flows of efforts to address. In parallel, more and more recent works consider leveraging the vast volume of both nature images in CV and remote sensing images to perform so-called X-shot learning [20][21][22][23][24][25], harnessing the Object detection methods date back to early and traditional methods that rely on hand-crafted and distinctive features, such as SIFT [6] and HOG [7], and match objects of interest in the images based on object examples (template image). Then, bounding boxes are extracted to describe the successful detections.…”
Section: Introductionmentioning
confidence: 99%
“…This, however, brings new challenges and solutions in DL from multi-modal data, which leads to flows of efforts to address. In parallel, more and more recent works consider leveraging the vast volume of both nature images in CV and remote sensing images to perform so-called X-shot learning [20][21][22][23][24][25], harnessing the power of pre-trained and fine-tuned networks to boost object detection performances [25,26]. This was further aided by using the nowadays well-developed language models to realize automated object detection in a much larger label space, i.e., fine-grained object detection.…”
Section: Introductionmentioning
confidence: 99%