2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN) 2018
DOI: 10.1109/roman.2018.8525717
|View full text |Cite
|
Sign up to set email alerts
|

Data-driven development of Virtual Sign Language Communication Agents

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(9 citation statements)
references
References 19 publications
0
9
0
Order By: Relevance
“…This observation should be very interesting and valuable in the future and help to reduce the overall number of necessary sign tracking features. Moreover, our results considerably outperform results reported for an end-to-end approach utilizing a reduced number of sentence collections of the same data set [ 31 ]. As such, the staged approach appears meaningful for small, sparse and insufficient data sets with high NME content, which are otherwise hard to learn.…”
Section: Discussionmentioning
confidence: 44%
See 1 more Smart Citation
“…This observation should be very interesting and valuable in the future and help to reduce the overall number of necessary sign tracking features. Moreover, our results considerably outperform results reported for an end-to-end approach utilizing a reduced number of sentence collections of the same data set [ 31 ]. As such, the staged approach appears meaningful for small, sparse and insufficient data sets with high NME content, which are otherwise hard to learn.…”
Section: Discussionmentioning
confidence: 44%
“…Previous efforts to implement a bidirectional sequence-to-sequence system from joint position data [ 31 ] show that it is very difficult to learn a working system for the given data set due to its high variability induced by the presence of linguistic features. For this reason, we perform our following analysis under a two-stage system with temporal segmentation.…”
Section: Cslr System Pipelinementioning
confidence: 99%
“…A common approach to sign language representation is the use of 3D avatars that with a high degree of accuracy and realism can reproduce facial expressions and body/hand movements in a way that represent signs understandable by deaf or hearing-impaired people. Balayn et al in [ 99 ], developed a virtual communication agent for sign language to recognize Japanese sign language sentences from video recordings and synthesize sign language animations. Their system adopted a deep LSTM encoder-decoder network to translate sign language videos to spoken text, while a separate encoder-decoder network used as input the sign language glosses and extracted specific encodings, which were then used to synthesize the avatar motion.…”
Section: Sign Language Representationmentioning
confidence: 99%
“…Other examples of normalization can be observed in the experiments by [67], [70], [71].Balayn et al [67] normalized Japanese sign language (JSL) motion sentences and used them as inputs and outputs for Seq2Seq models. Konstantinidis et al [70] normalized hand positions, which were used as inputs for the classifier together with cropped hand regions.…”
Section: ) Normalization and Filteringmentioning
confidence: 99%
“…Koller et al [63] utilized 1024-dimensional feature maps and applied PCA to reduce the dimensionality to 200. Another experiment by [67] used PCA to select data streams exhibiting a high variance represented by approximately 492 dimensions. The use of PCA on Kinect data has also helped to reduce cases of overfitting.…”
Section: Principal Component Analysismentioning
confidence: 99%