2021
DOI: 10.1109/lsp.2021.3086395
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Corpus Speech Emotion Recognition Based on Few-Shot Learning and Domain Adaptation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 30 publications
(16 citation statements)
references
References 35 publications
0
12
0
Order By: Relevance
“…Classes Accuracy GAN [29] eGeMAPS [27]+EMO-DB 2 66% (UAR) FLUDA [30] IS10 [31]+IEMOCAP 4 50% (UA) VAE+LSTM [13] LogMel+IEMOCAP 4 56.08% (UA) AE+LSTM [13] LogMel+IEMOCAP 4 55.42% (UA) Stacked-AE+BLSTM-RNN [12] COVAREP+IEMOCAP [ on a similar concept as that in this paper and their performance in Tab. 1.…”
Section: Methods Features+datasetmentioning
confidence: 93%
“…Classes Accuracy GAN [29] eGeMAPS [27]+EMO-DB 2 66% (UAR) FLUDA [30] IS10 [31]+IEMOCAP 4 50% (UA) VAE+LSTM [13] LogMel+IEMOCAP 4 56.08% (UA) AE+LSTM [13] LogMel+IEMOCAP 4 55.42% (UA) Stacked-AE+BLSTM-RNN [12] COVAREP+IEMOCAP [ on a similar concept as that in this paper and their performance in Tab. 1.…”
Section: Methods Features+datasetmentioning
confidence: 93%
“…Features+Dataset classes Accuracy GAN [18] eGeMAPS [10]+EMO-DB 2 66% (UAR) FLUDA [1] IS10 [27]+IEMOCAP(+) 4 50% (UA) VAE+LSTM [20] LogMel+IEMOCAP 4 56.08% (UA) AE+LSTM [20] LogMel+IEMOCAP 4 55.42% (UA) Stacked-AE+BLSTM-RNN [13] COVAREP+IEMOCAP [6] 4 50.26% (UA) DAE+Linear-SVM (baseline) eGeMAPS+IEMOCAP 4 52.09% (UA)…”
Section: Methodsmentioning
confidence: 99%
“…Compared with existing state-of-the-art approaches, our proposed CTA-RNN architecture can significantly improve the performance of SER in both within-corpus and cross-corpus experiments. Different from previous works [18][26] [27], the unlabeled target datasets were not available in advance for the cross-corpus experiments described in this paper. The excellent robustness of our approach mainly benefits from the large-scale ASR datasets for pre-training.…”
Section: Fusion Of Asr Embeddingsmentioning
confidence: 99%