2018
DOI: 10.1007/s11042-018-6565-5
|View full text |Cite
|
Sign up to set email alerts
|

On the role of multimodal learning in the recognition of sign language

Abstract: Sign Language Recognition (SLR) has become one of the most important research areas in the field of human computer interaction. SLR systems are meant to automatically translate sign language into text or speech, in order to reduce the communicational gap between deaf and hearing people. The aim of this paper is to exploit multimodal learning techniques for an accurate SLR, making use of data provided by Kinect and Leap Motion. In this regard, single-modality approaches as well as different multimodal methods, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 22 publications
(8 citation statements)
references
References 22 publications
0
5
0
1
Order By: Relevance
“…The large-scale dataset-based investigation could be future work to improve recognition accuracy. [92] isoGD, SBU, NATOPS, SKIG RGB, Depth, Dynamic Rastgoo et al [3] RKS-PERSIANSIGN, NYU RGB, Dynamic Köpüklü et al [93] EgoGesture, NVIDIA benchmarks Lim et al [94] RWTH-BOSTON-50, ASLLVD Chen et al [95] DHG-14/28 Dataset, SHREC'17 Track Dataset Ferreira et al [96] Real video samples RGB, Depth, Static Gomez-Donoso et al [97] STB Spurr et al [98] NYU, STB, MSRA, ICVL Kazakos et al [99] NYU Li et al [100] B2RGB-SH, STB RGB, Static Mueller et al [101] EgoDexter, Dexter, STB Victor [102] Egohands Baek et al [103] BigHand2.2M, MSRA, ICVL, NYU Depth, Static Moon et al [104] MSRA, ICVL, NYU Ge et al [105] MSRA, ICVL, NYU Ge and et. al [106] MSRA, NYU Dibra et al [107] ICVL, NYU Sinha et al [108] NYU Zimmermann and Brox [109] Dexter, STB 3D, RGB Marin-Jimenez et al [110] UBC3V, ITOP 3D, Depth Deng et al [111] NYU Oberweger et al [112] MSRA Oberweger et al [113] NYU Rastgoo et al [114] Massey 2012, ASL Fingerspelling A, SL Surrey 2D, Depth, RGB Duan et al [115] RGBD-HuDaAct, isoGD Chen et al [116] NYU, ICVL, MSRA 2D, Depth Dadashzadeh et al [117] OUHANDS Wang et al [118] Human3.6M Yuan et al [119] BigHand2.2M, MSRA, ICVL, NYU Guo et al [120] ITOP, MSRA, ICVL, NYU Fang and Lei [121] ICVL, NYU Madadi et al [122] MSRA, NYU Wang et al [123] isoGD Haque et al [124] EVAL, ITOP Tagliasacchi et al [125] Real video samples Rastgoo et al…”
Section: A Manual Slrmentioning
confidence: 99%
“…The large-scale dataset-based investigation could be future work to improve recognition accuracy. [92] isoGD, SBU, NATOPS, SKIG RGB, Depth, Dynamic Rastgoo et al [3] RKS-PERSIANSIGN, NYU RGB, Dynamic Köpüklü et al [93] EgoGesture, NVIDIA benchmarks Lim et al [94] RWTH-BOSTON-50, ASLLVD Chen et al [95] DHG-14/28 Dataset, SHREC'17 Track Dataset Ferreira et al [96] Real video samples RGB, Depth, Static Gomez-Donoso et al [97] STB Spurr et al [98] NYU, STB, MSRA, ICVL Kazakos et al [99] NYU Li et al [100] B2RGB-SH, STB RGB, Static Mueller et al [101] EgoDexter, Dexter, STB Victor [102] Egohands Baek et al [103] BigHand2.2M, MSRA, ICVL, NYU Depth, Static Moon et al [104] MSRA, ICVL, NYU Ge et al [105] MSRA, ICVL, NYU Ge and et. al [106] MSRA, NYU Dibra et al [107] ICVL, NYU Sinha et al [108] NYU Zimmermann and Brox [109] Dexter, STB 3D, RGB Marin-Jimenez et al [110] UBC3V, ITOP 3D, Depth Deng et al [111] NYU Oberweger et al [112] MSRA Oberweger et al [113] NYU Rastgoo et al [114] Massey 2012, ASL Fingerspelling A, SL Surrey 2D, Depth, RGB Duan et al [115] RGBD-HuDaAct, isoGD Chen et al [116] NYU, ICVL, MSRA 2D, Depth Dadashzadeh et al [117] OUHANDS Wang et al [118] Human3.6M Yuan et al [119] BigHand2.2M, MSRA, ICVL, NYU Guo et al [120] ITOP, MSRA, ICVL, NYU Fang and Lei [121] ICVL, NYU Madadi et al [122] MSRA, NYU Wang et al [123] isoGD Haque et al [124] EVAL, ITOP Tagliasacchi et al [125] Real video samples Rastgoo et al…”
Section: A Manual Slrmentioning
confidence: 99%
“…Kumar et al [1] used Kinect and Leap Motion sensors for data acquisition and achieved 96.33% accuracy on a 50-sign Indian sign language dataset when they used both of the data modalities. Ferreira et al [24] developed a multimodal SLR system to recognize 10 motionless signs in the American sign language. In their study, they used RGB, depth and 3D skeletal data obtained by Leap Motion.…”
Section: Related Workmentioning
confidence: 99%
“…In order to extract the manual signs from the noisy background of the images, the automatic hand detection algorithm [28] is used as a pre-processing step. The images are then cropped, resized to the average sign size of the training set, and normalized to be in the range [−1, 1].…”
Section: A Implementation Detailsmentioning
confidence: 99%
“…The images are then cropped, resized to the average sign size of the training set, and normalized to be in the range [−1, 1]. Throughout this section, the proposed model is compared with state-of-the-art methods for each dataset [15], [16], [24], [27], [28]. Nevertheless, to further attest the robustness of the proposed model, two different baselines are also implemented: 1) (Baseline 1) A CNN trained from scratch with ℓ -2 regularization.…”
Section: A Implementation Detailsmentioning
confidence: 99%