2022
DOI: 10.48550/arxiv.2210.16901
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Foreign Object Debris Detection for Airport Pavement Images based on Self-supervised Localization and Vision Transformer

Abstract: Supervised object detection methods provide subpar performance when applied to Foreign Object Debris (FOD) detection because FOD could be arbitrary objects according to the Federal Aviation Administration (FAA) specification. Current supervised object detection algorithms require datasets that contain annotated examples of every to-be-detected object. While a large and expensive dataset could be developed to include common FOD examples, it is infeasible to collect all possible FOD examples in the dataset repre… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 9 publications
0
1
0
Order By: Relevance
“…14 The principle behind Transformer and its variants is the self-attention mechanism under a general encoder-decoder framework. Munyer et al 15 present a self-supervised foreign object detection and lolcalization using vision Transformer (ViT) backbone. 16 Although Transformer models have been investigated in foreign object detection and localization due to it's capability to model intricate relationships among image patches or pixels across the image, its performance heavily rely on huge amount of training data.…”
Section: Related Workmentioning
confidence: 99%
“…14 The principle behind Transformer and its variants is the self-attention mechanism under a general encoder-decoder framework. Munyer et al 15 present a self-supervised foreign object detection and lolcalization using vision Transformer (ViT) backbone. 16 Although Transformer models have been investigated in foreign object detection and localization due to it's capability to model intricate relationships among image patches or pixels across the image, its performance heavily rely on huge amount of training data.…”
Section: Related Workmentioning
confidence: 99%