2018 IEEE International Conference on Imaging Systems and Techniques (IST) 2018
DOI: 10.1109/ist.2018.8577085
|View full text |Cite
|
Sign up to set email alerts
|

A Deep Learning Approach for Analyzing Video and Skeletal Features in Sign Language Recognition

Abstract: Sign language recognition (SLR) refers to the classification of signs with a specific meaning performed by the deaf and/or hearing-impaired people in their everyday communication. In this work, we propose a deep learning based framework, in which we examine and analyze the contribution of video (image and optical flow) and skeletal (body, hand and face) features in the challenging task of isolated SLR, in which each signed video corresponds to a single word. Moreover, we employ various fusion schemes in order … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
30
0
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
7
3

Relationship

3
7

Authors

Journals

citations
Cited by 43 publications
(32 citation statements)
references
References 23 publications
0
30
0
2
Order By: Relevance
“…These details make the problem complicated, as it is not essentially a full repetitive set of actions that mean the same thing, but a constructed discourse [25] with verbs and actions, as well as facial expressions, including happiness, deception, disappointment, and other impressions that can be driven in synchronization with the hand movements. Authors of [26], have recapitulated, in a very organized way, all the methods categorizing sign language over a full pipeline methodology recognition process, from the sensor acquisition to processing, and then to recognition methods.…”
Section: Methods Used In Arslmentioning
confidence: 99%
“…These details make the problem complicated, as it is not essentially a full repetitive set of actions that mean the same thing, but a constructed discourse [25] with verbs and actions, as well as facial expressions, including happiness, deception, disappointment, and other impressions that can be driven in synchronization with the hand movements. Authors of [26], have recapitulated, in a very organized way, all the methods categorizing sign language over a full pipeline methodology recognition process, from the sensor acquisition to processing, and then to recognition methods.…”
Section: Methods Used In Arslmentioning
confidence: 99%
“…Input observations are skeleton joint information, depth and RGB images, using a Gauss-Bernoulli deep belief network suitable for input forms to process bone dynamics, and applying convolutional neural networks to adjust and fuse batch depth and RGB images. Konstantinidis et al [16] propose a sign language recognition RNN networks based on RGB, skeleton data, and facial expression features. The data fusion schemes are analyzed.…”
Section: A Rgb-dmentioning
confidence: 99%
“…This is mainly because sign languages feature thousands of signs, sometimes differing only by subtle changes in hand motion, shape, or position and involving significant finger overlaps and occlusions [ 2 ]. SLR tasks are divided into Isolated Sign Language Recognition (ISLR) [ 3 , 4 , 5 ] and Continuous Sign Language Recognition (CSLR) [ 6 , 7 , 8 ]. The CSLR task focuses on recognizing sequences of glosses from videos without predefined annotation boundaries, and it is more challenging compared to ISLR [ 9 ], in which the temporal boundaries of glosses in the videos are predefined.…”
Section: Introductionmentioning
confidence: 99%