2016
DOI: 10.1016/j.engappai.2016.02.018
|View full text |Cite
|
Sign up to set email alerts
|

Multistage data selection-based unsupervised speaker adaptation for personalized speech emotion recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
14
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
7
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 21 publications
(14 citation statements)
references
References 28 publications
0
14
0
Order By: Relevance
“…First, pre‐processing the Naxi speech signal to be identified, then extracting its feature parameters, entering the next model to match, and comparing the obtained template with the template stored in the GMM model library. The highest matching probability is the recognition result [11]. GMM speaker recognition system process was shown in Fig.…”
Section: Speech Recognition Methodsmentioning
confidence: 99%
“…First, pre‐processing the Naxi speech signal to be identified, then extracting its feature parameters, entering the next model to match, and comparing the obtained template with the template stored in the GMM model library. The highest matching probability is the recognition result [11]. GMM speaker recognition system process was shown in Fig.…”
Section: Speech Recognition Methodsmentioning
confidence: 99%
“…Their system was evaluated on Arabic Emirati-Accented and SUSAS datasets and they obtained an average recognition rate of 83.97% and 86.67%, respectively. Kim and Park [17] proposed a multistage data selection method for speech emotion recognition from previous voice data accumulated on personal devices. Multistage data selection is conducted using log likelihood distance based measure and a universal background model [17] and obtained an average recognition rate of 83.9%.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Kim and Park [17] proposed a multistage data selection method for speech emotion recognition from previous voice data accumulated on personal devices. Multistage data selection is conducted using log likelihood distance based measure and a universal background model [17] and obtained an average recognition rate of 83.9%. Literature shows that many of the researches obtained higher accuracy by using hybrid classifier models.…”
Section: Literature Reviewmentioning
confidence: 99%
“…To ignite the interactions between smart devices and their owners, automatic speaker recognition (ASR) plays an important role to determine the speaker identity based on a short piece of audio. Moreover, the capability of ASR comes with a wide range of applications, such as biometric authentication [23], forensics [10], and personalized services in electronics [13]. In particular, the text-independent ASR with only acoustic information is the most general and non-trial task, which can be used in everyday situations.…”
Section: Introductionmentioning
confidence: 99%