2020 International Conference on Information and Communication Technology Convergence (ICTC) 2020
DOI: 10.1109/ictc49870.2020.9289551
|View full text |Cite
|
Sign up to set email alerts
|

Robust Keypoint Normalization Method for Korean Sign Language Translation using Transformer

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(13 citation statements)
references
References 2 publications
0
13
0
Order By: Relevance
“…Human pose estimation systems are used to extract features in seven works (37%) [38,34,9,49,37,85]. The estimated poses can be the sole inputs to the translation model [38,37,49], or augment other spatial or spatiotemporal features [34,9,85].…”
Section: Sign Language Representationsmentioning
confidence: 99%
See 3 more Smart Citations
“…Human pose estimation systems are used to extract features in seven works (37%) [38,34,9,49,37,85]. The estimated poses can be the sole inputs to the translation model [38,37,49], or augment other spatial or spatiotemporal features [34,9,85].…”
Section: Sign Language Representationsmentioning
confidence: 99%
“…As one paper may discuss several tasks, the total count is higher than the amount of papers. 1, 57, 51] and transformers also in 12 papers [80,82,46,84,9,11,38,49,37,44,18,81]. Within the RNN based models, several attention schemes are used: no attention, Luong attention [42] and Bahdanau attention [3].…”
Section: Sign Language Translation Modelsmentioning
confidence: 99%
See 2 more Smart Citations
“…In their latest work, Camgoz et al in [ 97 ], adopted additional modalities and a cross-modal attention to synchronize the different streams and model both inter- and intra-contextual information. Kim et al in [ 98 ], used a deep neural network for human keypoint extraction that were fed to a transformer encoder-decoder network, while the keypoints were normalized based on the neck location. A comparison of existing methods for SLT that are evaluated on the Phoenix-2014-T dataset, is shown in Table 4 .…”
Section: Sign Language Recognitionmentioning
confidence: 99%