2022
DOI: 10.1609/aaai.v36i1.19957
|View full text |Cite
|
Sign up to set email alerts
|

Contrastive Learning from Extremely Augmented Skeleton Sequences for Self-Supervised Action Recognition

Abstract: In recent years, self-supervised representation learning for skeleton-based action recognition has been developed with the advance of contrastive learning methods. The existing contrastive learning methods use normal augmentations to construct similar positive samples, which limits the ability to explore novel movement patterns. In this paper, to make better use of the movement patterns introduced by extreme augmentations, a Contrastive Learning framework utilizing Abundant Information Mining for self-supervis… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
53
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 60 publications
(53 citation statements)
references
References 30 publications
0
53
0
Order By: Relevance
“…The SoTA methods used in the comparative experiment were (1) the base SGN [ 1 ], (2) ST-GCN [ 4 ], (3) a shift graph convolutional network (Shift-GCN) [ 12 ], (4) an Info-GCN for representation learning for the human skeleton [ 25 ], (5) disentangling and unifying graph convolutions (MS-G3D) [ 26 ], (6) a decoupled spatial-temporal attention network (DSTA-Net) [ 27 ], (7) 3S-aimCLR based on contrastive learning from extremely augmented skeleton sequences [ 28 ], (8) rich activated GCN (RA-GCNv2) [ 29 ], and (9) the proposed SGN-SHA.…”
Section: Resultsmentioning
confidence: 99%
“…The SoTA methods used in the comparative experiment were (1) the base SGN [ 1 ], (2) ST-GCN [ 4 ], (3) a shift graph convolutional network (Shift-GCN) [ 12 ], (4) an Info-GCN for representation learning for the human skeleton [ 25 ], (5) disentangling and unifying graph convolutions (MS-G3D) [ 26 ], (6) a decoupled spatial-temporal attention network (DSTA-Net) [ 27 ], (7) 3S-aimCLR based on contrastive learning from extremely augmented skeleton sequences [ 28 ], (8) rich activated GCN (RA-GCNv2) [ 29 ], and (9) the proposed SGN-SHA.…”
Section: Resultsmentioning
confidence: 99%
“…In the field of skeleton-based action recognition, prior works (Li et al, 2021;Mao et al, 2022;Guo et al, 2022) proposed to apply contrastive learning in the pre-training stage by roughly following the frameworks mentioned above. CrossCLR (Li et al, 2021) mined positive pairs in the data space and explored the cross-modal distribution relationships.…”
Section: Contrastive Learningmentioning
confidence: 99%
“…Further, CMD (Mao et al, 2022) transferred the cross-modal knowledge in a distillation manner. And AimCLR (Guo et al, 2022) used extreme augmentations to improve the representation universality.…”
Section: Contrastive Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…Wang et al [48] proposed the contrast-reconstruction representation learning network to capture postures and motion dynamics simultaneously. In [49], Guo et al utilized the abundant information mining strategy to make better use of the movement patterns. In [50], [51], it is suggested that contrasting congruent and incongruent views of graphs with mutual information maximization can help encode rich representations.…”
Section: Self-supervised Learningmentioning
confidence: 99%