2020
DOI: 10.1109/access.2020.3028072
|View full text |Cite
|
Sign up to set email alerts
|

AUTSL: A Large Scale Multi-Modal Turkish Sign Language Dataset and Baseline Methods

Abstract: Sign language recognition is a challenging problem where signs are identified by simultaneous local and global articulations of multiple sources, i.e. hand shape and orientation, hand movements, body posture, and facial expressions. Solving this problem computationally for a large vocabulary of signs in real life settings is still a challenge, even with the state-of-the-art models. In this study, we present a new largescale multi-modal Turkish Sign Language dataset (AUTSL) with a benchmark and provide baseline… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
44
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 118 publications
(75 citation statements)
references
References 48 publications
0
44
0
1
Order By: Relevance
“…Then, they built a temporal attention-based model for classification. In our previous work [5], we proposed a baseline model for a new large-scale isolated Turkish Sign Language (AUTSL) dataset. We integrated a Feature Pooling Module and a temporal attention model to focus on more relevant spatio-temporal parts of the videos.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Then, they built a temporal attention-based model for classification. In our previous work [5], we proposed a baseline model for a new large-scale isolated Turkish Sign Language (AUTSL) dataset. We integrated a Feature Pooling Module and a temporal attention model to focus on more relevant spatio-temporal parts of the videos.…”
Section: Related Workmentioning
confidence: 99%
“…We evaluate our proposed models on two very recently shared large-scale isolated sign language datasets: AUTSL [5] and BosphorusSign22k [6].…”
Section: A Datasets and Preprocessingmentioning
confidence: 99%
See 1 more Smart Citation
“…Contrary to the previous methods that use a single Kinect sensor, this work additionally employs a machine vision camera, along with a television screen, for sign demonstration. Sincan et al in [ 16 ], captured isolated Turkish sign language glosses using Kinect sensors with a large variety of indoor and outdoor backgrounds, revealing the importance of capturing videos with various backgrounds. Adaloglou et al in [ 17 ], created a large sign language dataset with RealSense D435 sensor that records both RGB and depth information.…”
Section: Sign Language Capturingmentioning
confidence: 99%
“…Примечание: ЗР -задача распознавания; ПО -количество предметных областей; Р -разметка; УЗ -устройство захвата (камера, сенсор); ЦК -цветная камера; н/д -нет данных.ИНФОРМАЦИОННОУПРАВЛЯЮЩИЕ СИСТЕМЫ № 6, 202112 …”
unclassified