2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.01272
|View full text |Cite
|
Sign up to set email alerts
|

PCAN: 3D Attention Map Learning Using Contextual Information for Point Cloud Based Retrieval

Abstract: Point cloud based retrieval for place recognition is an emerging problem in vision field. The main challenge is how to find an efficient way to encode the local features into a discriminative global descriptor. In this paper, we propose a Point Contextual Attention Network (PCAN), which can predict the significance of each local point feature based on point context. Our network makes it possible to pay more attention to the task-relevent features when aggregating local features. Experiments on various benchmar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
134
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 201 publications
(143 citation statements)
references
References 40 publications
0
134
0
Order By: Relevance
“…Inspired by the self-attention idea in natural language processing [40], recent works connect the self-attention mechanism with contextual information mining to improve scene understanding tasks such as image recognition [41], semantic segmentation [11] and point cloud recognition [42]. As to 3D point data processing, the work in [14] proposes to utilize the attention network to capture the contextual information in 3D points. Specifically, it presents a point contextual attention network to encode local features into a global descriptor for point cloud based retrieval.…”
Section: Contextual Informationmentioning
confidence: 99%
See 1 more Smart Citation
“…Inspired by the self-attention idea in natural language processing [40], recent works connect the self-attention mechanism with contextual information mining to improve scene understanding tasks such as image recognition [41], semantic segmentation [11] and point cloud recognition [42]. As to 3D point data processing, the work in [14] proposes to utilize the attention network to capture the contextual information in 3D points. Specifically, it presents a point contextual attention network to encode local features into a global descriptor for point cloud based retrieval.…”
Section: Contextual Informationmentioning
confidence: 99%
“…To model the contextual information, three sub-modules are proposed in the framework, i.e., patch-to-patch context (PPC) module, object-to-object context (OOC) module and the global scene context (GSC) module. In particular, similar to [14], we use the self-attention mechanism to model the relationships between elements in both PPC and OOC modules. These two sub-modules aim at adaptively encoding contextual information at the patch and object levels, respectively.…”
Section: Introductionmentioning
confidence: 99%
“…However, the effects of local area and dynamic noise are not taken into account in PointNetVLAD. PCAN [ZX19] proposes a Point Contextual Attention Network to PointNetVLAD, thus making the model pay more attention to more task-related areas. Through this special attention mechanism, their model can extract local features to some extent.…”
Section: Scene Recognition Methodsmentioning
confidence: 99%
“…Our models are compared with recent state-of-the-art methods: PointNetVALD [AUHL18], PCAN [ZX19], DAGC [SLH * 20] and LPD-Net [LZS * 19]. Performance is evaluated by average recall at top 1% and average recall at top 1.…”
Section: Comparison With State-of-the-art Methodsmentioning
confidence: 99%
See 1 more Smart Citation