2013
DOI: 10.1007/978-3-642-38628-2_10
|View full text |Cite
|
Sign up to set email alerts
|

Modality Combination Techniques for Continuous Sign Language Recognition

Abstract: Abstract. Sign languages comprise parallel aspects and use several modalities to form a sign but so far it is not clear how to best combine these modalities in the context of statistical sign language recognition. We investigate early combination of features, late fusion of decisions, as well as synchronous combination on the hidden Markov model state level, and asynchronous combination on the gloss level. This is done for five modalities on two publicly available benchmark databases consisting of challenging … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2013
2013
2021
2021

Publication Types

Select...
3
3
3

Relationship

4
5

Authors

Journals

citations
Cited by 47 publications
(37 citation statements)
references
References 15 publications
(12 reference statements)
0
21
0
Order By: Relevance
“…Real-world translation typically requires continuous sign language recognition [105,41], where a continuous stream of signing is deciphered. Continuous recognition is a significantly more challenging and realistic problem than recognizing individual signs, confounded by epenthesis effects (insertion of extra features into signs), co-articulation (the ending of one sign affecting the start of the next), and spontaneous sign production (which may include slang, non-uniform speed, etc.…”
Section: Recognition and Computer Visionmentioning
confidence: 99%
“…Real-world translation typically requires continuous sign language recognition [105,41], where a continuous stream of signing is deciphered. Continuous recognition is a significantly more challenging and realistic problem than recognizing individual signs, confounded by epenthesis effects (insertion of extra features into signs), co-articulation (the ending of one sign affecting the start of the next), and spontaneous sign production (which may include slang, non-uniform speed, etc.…”
Section: Recognition and Computer Visionmentioning
confidence: 99%
“…Brand proposes an algorithm ("N-head dynamic programming") that reaches O(T (N M ) 2 ) complexity [21]. The asynchronous HMM requires both streams to share the same topology [23]. This is imposed due to the assumption that there is a single underlying hidden process that has multiple distinct probabilities to emit an observation on all or just on a single stream.…”
Section: Related Workmentioning
confidence: 99%
“…However, they all deal with signs in isolation (not in a continuous sentence sequence), which essentially turns the multi-stream HMM with synchronisation at the end of signs into a standard late fusion approach and is therefore not comparable to the approach analysed in this paper. In 2013, Forster et al [23] compared different modality combination techniques to recognise continuous DGS with a vocabulary of up to 455 signs. They use multistream HMMs with synchronisation constraints at the wordlevel.…”
Section: Related Workmentioning
confidence: 99%
“…, s T defined by the symbol sequences in ψ. For an efficient implementation, following [11], we assume a first order Markov dependency and maximum approximation:…”
Section: Sequential Time-decodingmentioning
confidence: 99%