Interspeech 2019 2019
DOI: 10.21437/interspeech.2019-2110
|View full text |Cite
|
Sign up to set email alerts
|

Using Attention Networks and Adversarial Augmentation for Styrian Dialect Continuous Sleepiness and Baby Sound Recognition

Abstract: In this study, we present extensive attention-based networks with data augmentation methods to participate in the IN-TERSPEECH 2019 ComPareE Challenge, specifically the three Sub-challenges: Styrian Dialect Recognition, Continuous Sleepiness Regression, and Baby Sound Classification. For Styrian Dialect Sub-challenge, these dialects are classified into Northern Styrian (NorthernS), Urban Sytrian (UrbanS), and Eastern Styrian (EasternS). Our proposed model achieves an UAR 49.5% on the test set, which is 2.5% hi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(10 citation statements)
references
References 21 publications
0
10
0
Order By: Relevance
“…(Note that this score was improved to .387 by training ensembles of classifiers [20].) In [21], the authors employ attention networks and adversarial augmentation, in the end, their best results (.369 of CC on test) are achieved by a fusion of neural network models. In [22], a .367 of CC was obtained by an early fusion of the learnt representations from attention and sequence to sequence autoencoders.…”
Section: Resultsmentioning
confidence: 99%
“…(Note that this score was improved to .387 by training ensembles of classifiers [20].) In [21], the authors employ attention networks and adversarial augmentation, in the end, their best results (.369 of CC on test) are achieved by a fusion of neural network models. In [22], a .367 of CC was obtained by an early fusion of the learnt representations from attention and sequence to sequence autoencoders.…”
Section: Resultsmentioning
confidence: 99%
“…However, notice that the automated method does lead to much better performance than Zooniverse classifications for Crying, and to some improvements in Canonical and Junk. For instance, the team who won the challenge in 2019 improved UAR by about 2%, primarily through gains in the laughing class obtained by adding training data (Yeh et al, 2019). That state of the art was challenged by Kaya, Verkholyak, Markitantov, and Karpov (2020), who obtained a UAR of 61% on the same data as the challenge, thanks to improvements in all of the classes but for Laughing.…”
Section: Discussionmentioning
confidence: 99%
“…However, notice that the automated method does lead to much better performance than Zooniverse classifications for Crying, and to some improvements in Canonical and Junk. For instance, the team who won the challenge in 2019 improved UAR by about 2%, primarily through gains in the laughing class obtained by adding training data (Yeh et al, 2019). That state of the art was challenged by Kaya et al (2020), who obtained a UAR of 61% on the same data as the challenge, thanks to improvements in all of the classes but for Laughing.…”
Section: Further Research Directionsmentioning
confidence: 99%