2022
DOI: 10.1109/lgrs.2021.3061422
|View full text |Cite
|
Sign up to set email alerts
|

Point Transformer for Shape Classification and Retrieval of Urban Roof Point Clouds

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 14 publications
0
2
0
Order By: Relevance
“…There are also some methods [7,30,31] that use selfattention and transformer layers, which have revolutionised the study of machine translation and natural language processing, to directly process 3D points and make progress on point cloud processing tasks. However, they cannot be directly applied to the point-wise class prediction of large-scale point clouds because the self-attention layer will produce L 2 time complexity and memory usage.…”
Section: Point-based Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…There are also some methods [7,30,31] that use selfattention and transformer layers, which have revolutionised the study of machine translation and natural language processing, to directly process 3D points and make progress on point cloud processing tasks. However, they cannot be directly applied to the point-wise class prediction of large-scale point clouds because the self-attention layer will produce L 2 time complexity and memory usage.…”
Section: Point-based Methodsmentioning
confidence: 99%
“…Recently, there are some specially designed deep neural networks used to process 3D point clouds. These methods can be categorised into the following categories: (1) pointbased methods [1][2][3][4][5][6][7][8][9][10] that directly operate 3D points and output semantic information. (2) Voxel-based methods [11][12][13][14][15][16][17][18] that voxelise point clouds into 3D grids and then use 3D CNNs to process these 3D grids.…”
Section: Introductionmentioning
confidence: 99%
“…By using the multi-head attention mechanism, each head can solve for a set of q, k, and v matrices. These matrices are then individually concatenated to obtain the matrices Q, K, and V. Following the practice of Point Transformer [42,43], we use subtraction instead of dot multiplication for the interaction between Q and K. Therefore, the attention score y is calculated according to the scaled dot-product attention:…”
Section: Multiscale Feature Extraction and Fusion (Mef) Unitmentioning
confidence: 99%
“…Li et al proposed the SO-Net model [6], which can systematically adjust the receptive field overlap to perform hierarchical feature extraction by performing point-to-node KNN search on SOM. The Transformer-based method [18,23,14,31] is an algorithm that has emerged in the past two years. Transformer is based on self-attention (SA), which was initially used for natural language processing, and then gradually applied to computer vision due to 2023/3/52 its strong feature representation ability.…”
Section: Introductionmentioning
confidence: 99%