2022
DOI: 10.3389/fcomp.2022.837269
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating the Impact of Voice Activity Detection on Speech Emotion Recognition for Autistic Children

Abstract: Individuals with autism are known to face challenges with emotion regulation, and express their affective states in a variety of ways. With this in mind, an increasing amount of research on automatic affect recognition from speech and other modalities has recently been presented to assist and provide support, as well as to improve understanding of autistic individuals' behaviours. As well as the emotion expressed from the voice, for autistic children the dynamics of verbal speech can be inconsistent and vary g… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 9 publications
(12 citation statements)
references
References 26 publications
0
10
0
Order By: Relevance
“…As suggested by Milling et al (2022), the performance of VAD systems can have an impact on the estimation of affective states based on audio recordings. This hypothesis is supported by the clear differences in valence and arousal predictions in Table 5, when comparing the average absolute difference between VAD-based SER predictions and the non-VAD-based SER predictions, in particular for the arousal case.…”
Section: Discussionmentioning
confidence: 99%
See 4 more Smart Citations
“…As suggested by Milling et al (2022), the performance of VAD systems can have an impact on the estimation of affective states based on audio recordings. This hypothesis is supported by the clear differences in valence and arousal predictions in Table 5, when comparing the average absolute difference between VAD-based SER predictions and the non-VAD-based SER predictions, in particular for the arousal case.…”
Section: Discussionmentioning
confidence: 99%
“…The model was trained on audio data from the British study arm of the project containing audio data from 84 robot-supported intervention sessions of 25 English-speaking children with ASD from the United Kingdom, having a total duration of more than 17 hours. For the experiments reported by Milling et al (2022), the data was partitioned into train, development, and test partitions in a speaker-independent way. In total, this data contains more than 2 hours of annotated child vocalisations (partially overlapping with other vocalisations).…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations