2018
DOI: 10.1016/j.neucom.2018.08.042
|View full text |Cite
|
Sign up to set email alerts
|

3D separable convolutional neural network for dynamic hand gesture recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
31
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 63 publications
(31 citation statements)
references
References 12 publications
0
31
0
Order By: Relevance
“…A 3D CNN model allows features to be extracted in both spatial and temporal dimensions by performing 3D convolution. This enables motion information encoded in various frames to be captured [20]. Traditional classification approaches have also been utilised.…”
Section: Related Workmentioning
confidence: 99%
“…A 3D CNN model allows features to be extracted in both spatial and temporal dimensions by performing 3D convolution. This enables motion information encoded in various frames to be captured [20]. Traditional classification approaches have also been utilised.…”
Section: Related Workmentioning
confidence: 99%
“…Convex hull Open CV function was used to detect a number of defects (concavities) on hand and stored into the defect array. Based upon a number of defects, finger count was determined [20]. Kinect V2 to recognize different human hand gestures.…”
Section: Real-time Hand Gesture Recognition Using Kinect V2mentioning
confidence: 99%
“…For data augmentation, they use reverse ordering of the frames and horizontal mirroring, partial elastic deformation, and temporal elastic deformation. 3D convolution process is decomposed into two parts: depth-wise and point-wise [20]. However, due to the decomposition, the network is deepened, which lead to gradient dispersion and renders the network difficult to train.…”
Section: Real-time Hand Gesture Recognition Using 3d Convolutional Nementioning
confidence: 99%
“…Multiple modality inputs are used to further improve the performance of these CNN-based models. Depth [ 10 , 14 ] and optical flow [ 11 , 15 ] are the most common complements to the RGB images. However, the depth modality needs an extra sensor to capture accurate information from the user’s hands.…”
Section: Introductionmentioning
confidence: 99%