2020
DOI: 10.1049/iet-ipr.2019.0985
|View full text |Cite
|
Sign up to set email alerts
|

Efficient inception V2 based deep convolutional neural network for real‐time hand action recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
22
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 38 publications
(26 citation statements)
references
References 25 publications
0
22
0
Order By: Relevance
“…Ablation experiments ( Supplementary Table 2) showed that the most important factors for model performance were the usage of dice loss and the addition of squeeze blocks. We experimented with other architectural blocks, namely inception [14] and residuals [15], that have been employed in similar computer vision tasks [16] but were not able to observe any significant improvements. We found that having a constant number of filters (64) across every layer of the network performs better than the typical increase/decrease in filters for every downsampling/upsampling step, respectively.…”
mentioning
confidence: 99%
“…Ablation experiments ( Supplementary Table 2) showed that the most important factors for model performance were the usage of dice loss and the addition of squeeze blocks. We experimented with other architectural blocks, namely inception [14] and residuals [15], that have been employed in similar computer vision tasks [16] but were not able to observe any significant improvements. We found that having a constant number of filters (64) across every layer of the network performs better than the typical increase/decrease in filters for every downsampling/upsampling step, respectively.…”
mentioning
confidence: 99%
“…1. YOLO-V2 overcomes the challenges faced by the other recognition systems like Single Shot Detector (SSD) [7], and the Faster Region based Convolutional Neural Network (Faster-RCNN) [6]. YOLO-V2 is the modified version of conventional YOLO architecture.…”
Section: Methodsmentioning
confidence: 99%
“…YOLO-V2 model utilizes DarkNet-19 CNN architecture as a backbone for extracting the feature vectors from the image. The YOLO-V2 CNN model has been trained and evaluated with the assistance of two benchmark datasets including the NUS Hand posture-II (NUSHP-II) dataset [19], Senz 3D hand dataset (SENZ-3D) [20] and the custom designed dataset (MITI-HD) [7]. Figure 2 illustrates the flow diagram that describes the collection of data samples and the technique used for pre-processing the hand gesture data samples.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations