2022
DOI: 10.3390/machines10111083
|View full text |Cite
|
Sign up to set email alerts
|

Swin Transformer Combined with Convolution Neural Network for Surface Defect Detection

Abstract: Surface defect detection aims to classify and locate a certain defect that exists in the target surface area. It is an important part of industrial quality inspection. Most of the research on surface defect detection are currently based on convolutional neural networks (CNNs), which are more concerned with local information and lack global perception. Thus, CNNs are unable to effectively extract the defect features. In this paper, a defect detection method based on the Swin transformer is proposed. The structu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 32 publications
0
3
0
Order By: Relevance
“…With Transformer, features can be adaptively aggregated from a global view using self-attention mechanisms rather than convolution operations, which extract features from a local fixed field [ 31 ]. The form of standard self-attention operation can be formulated (see Equations (1) and (2)) as follows [ 32 ]: where X is the input feature map; are the linear transformation functions; is the channel dimension; are abbreviations of query, key, and weight value, respectively. We can calculate the dot products of the query with all keys, divide by , and apply a SoftMax function to obtain the weights on the values.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…With Transformer, features can be adaptively aggregated from a global view using self-attention mechanisms rather than convolution operations, which extract features from a local fixed field [ 31 ]. The form of standard self-attention operation can be formulated (see Equations (1) and (2)) as follows [ 32 ]: where X is the input feature map; are the linear transformation functions; is the channel dimension; are abbreviations of query, key, and weight value, respectively. We can calculate the dot products of the query with all keys, divide by , and apply a SoftMax function to obtain the weights on the values.…”
Section: Methodsmentioning
confidence: 99%
“…Here, we need one more projection there that the cost will be hwc 2 . Hence, MSA cost is calculated below in total (see Equation (3)) [ 32 ]: …”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation