2021
DOI: 10.1007/s41095-021-0229-5
|View full text |Cite
|
Sign up to set email alerts
|

PCT: Point cloud transformer

Abstract: The irregular domain and lack of ordering make it challenging to design deep neural networks for point cloud processing. This paper presents a novel framework named Point Cloud Transformer (PCT) for point cloud learning. PCT is based on Transformer, which achieves huge success in natural language processing and displays great potential in image processing. It is inherently permutation invariant for processing a sequence of points, making it well-suited for point cloud learning. To better capture local context … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
296
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 951 publications
(399 citation statements)
references
References 20 publications
(26 reference statements)
0
296
0
Order By: Relevance
“…whereas LBR represents three operations: linear operation, batch processing, and the ReLU activation function. F a = A(Q, K, V) and F in − F a are similar to a discrete Laplace operator [29]. Our experiments showed that self-attention was more beneficial for subsequent registration tasks when replaced by offset-attention.…”
Section: Transformermentioning
confidence: 73%
See 2 more Smart Citations
“…whereas LBR represents three operations: linear operation, batch processing, and the ReLU activation function. F a = A(Q, K, V) and F in − F a are similar to a discrete Laplace operator [29]. Our experiments showed that self-attention was more beneficial for subsequent registration tasks when replaced by offset-attention.…”
Section: Transformermentioning
confidence: 73%
“…Attention mechanisms use relative importance to focus on different parts of the input sequence, highlighting the relationship between inputs enabling the capture of context and high-order dependencies. Inspired by the point cloud processing Transformer proposed in recent years [29,30], we define Q, K, and V as the query, key, and value matrices, respectively, generated by the linear transformation of input characteristics.𝐹 𝑖𝑛 ∈ ℝ 𝑑 The function () A  describes the mapping of N queries 𝑄 ∈ ℝ 𝑁×𝑑 𝑘 and N key values to 𝐾 ∈ ℝ 𝑁 𝑘 ×𝑑 𝑘 and 𝑉 ∈ ℝ 𝑁 𝑘 ×𝑑𝑣 to the output [31]. The attention weight is calculated by the matrix dot product 𝑄𝐾 𝑇 ∈ ℝ 𝑁×𝑑 :…”
Section: Transformermentioning
confidence: 99%
See 1 more Smart Citation
“…CN [50] proposes a novel channel normalization scheme to balance information from different layers to benefit the model in integrating multilayer structure information. PCT [51] is based on Transformer, and the fundamental idea is introducing the inherent order invariance of Transformer and learning feature through the attention mechanism.…”
Section: Point-based Methodsmentioning
confidence: 99%
“…This mechanism is actually well suited to dealing with data like point clouds. PCT [31] enhances input embedding by supporting farthest point sampling and nearest neighbor search. It applies transformer to point clouds and achieves good results.…”
Section: Introductionmentioning
confidence: 99%