2022
DOI: 10.1177/14759217221089571
|View full text |Cite
|
Sign up to set email alerts
|

Feature pyramid network with self-guided attention refinement module for crack segmentation

Abstract: Automated pavement crack segmentation is challenging due to the random shape of cracks, complex background textures and the presence of miscellaneous objects. In this paper, we implemented a Self-Guided Attention Refinement module and incorporated it on top of a Feature Pyramid Network (FPN) to model long-range contextual information. The module uses multi-scale features integrated from different layers in the FPN to refine the features at each layer of the FPN using a self-attention mechanism. The module enab… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(5 citation statements)
references
References 73 publications
0
5
0
Order By: Relevance
“…More importantly, MT-GAN-CS achieves a high F 1 Score of 0.8 at CR = 128 even the used tasks are not similar to each other, such a performance level is comparable or superior to existing crack segmentation methods. [55][56][57] Furthermore, Figure 12 confirms that the reconstruction accuracy of MT-GAN-CS is basically the same as that of ST-GAN-CS, but its time cost to reconstruct one image block is only about a quarter of the cost by ST-GAN-CS (see Appendix D). Therefore, we conclude that MT-GAN-CS can speed up reconstruction without sacrificing the accuracy of recovered crack regions.…”
Section: Multitask Recovery For Image Blocks With Diverse Cracksmentioning
confidence: 58%
“…More importantly, MT-GAN-CS achieves a high F 1 Score of 0.8 at CR = 128 even the used tasks are not similar to each other, such a performance level is comparable or superior to existing crack segmentation methods. [55][56][57] Furthermore, Figure 12 confirms that the reconstruction accuracy of MT-GAN-CS is basically the same as that of ST-GAN-CS, but its time cost to reconstruct one image block is only about a quarter of the cost by ST-GAN-CS (see Appendix D). Therefore, we conclude that MT-GAN-CS can speed up reconstruction without sacrificing the accuracy of recovered crack regions.…”
Section: Multitask Recovery For Image Blocks With Diverse Cracksmentioning
confidence: 58%
“…Ong et al. (2023) proposed a multiscale encoder–decoder architecture with the embedment of self‐guided attention refinement modules, effectively suppressing background noise while enhancing network representation of crack details. Wu et al.…”
Section: Related Workmentioning
confidence: 99%
“…The authors have specifically noticed that among the previously discussed seven HR representation methods, multiscale sampling enhances the representation of small targets by leveraging multiscale feature information, while physical cascading operations achieve refined representation of randomly shaped targets by incorporating physical constraints. Notably, these two methods both exhibit relatively low model complexity, thus deeply mitigating the dependence on high GPU memory for HR image inference (Cheng et al., 2020; Ong et al., 2023). Consequently, this study aims to integrate these two methods to bridge the research gap in fine‐grained representation of HR crack images.…”
Section: Introductionmentioning
confidence: 99%
“…Inspired by NL, Wan et al (2021) designed CrackResAttentionNet for pavement crack detection based on the encoder-decoder network, where two self-attention-based attention modules were added after different encoder layers to aggregate long-range context information. Ong et al (2023) used selfattention to refine each feature pyramid network (FPN) layer so that the deep and shallow layers of the FPN could enhance crack information and reduce noise impact, respectively. However, the large computational cost of the self-attention mechanism severely limits the detection speed.…”
Section: Introductionmentioning
confidence: 99%
“…Ong et al. (2023) used self‐attention to refine each feature pyramid network (FPN) layer so that the deep and shallow layers of the FPN could enhance crack information and reduce noise impact, respectively. However, the large computational cost of the self‐attention mechanism severely limits the detection speed.…”
Section: Introductionmentioning
confidence: 99%