2021 12th International Symposium on Chinese Spoken Language Processing (ISCSLP) 2021
DOI: 10.1109/iscslp49672.2021.9362119
|View full text |Cite
|
Sign up to set email alerts
|

Speech Emotion Recognition Based on Acoustic Segment Model

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
2
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 23 publications
0
4
0
Order By: Relevance
“…A deep learning-based language model outperforms standard methods when applied to the Bag of Words model. Siyuan Zheng et al (2021) [17] To address this problem, they present in this work an acoustic segment model (ASM)-based technique for speech emotion recognition (SER).This research suggests a brand-new SER paradigm based on ASM. Topic models like LSA in the field of information retrieval can process the relationship between a document and a word.…”
Section: General Architecture Of Speech Emotion Detectionmentioning
confidence: 99%
“…A deep learning-based language model outperforms standard methods when applied to the Bag of Words model. Siyuan Zheng et al (2021) [17] To address this problem, they present in this work an acoustic segment model (ASM)-based technique for speech emotion recognition (SER).This research suggests a brand-new SER paradigm based on ASM. Topic models like LSA in the field of information retrieval can process the relationship between a document and a word.…”
Section: General Architecture Of Speech Emotion Detectionmentioning
confidence: 99%
“…Besides, studies were using CNN in combination with LSTM [17][18][19][20]. CNN, DCNN, and multi-channel CNN models were used in [21][22][23][24]. A combination of CNN and RNN models to get the CRNN model was used in [25].…”
Section: Related Workmentioning
confidence: 99%
“…For the feature parameters that have been used for emotion recognition, some studies combine the features of speech and textual data. Those are the studies in [13,15,21,24,27,28]. There are a large number of studies that have used a spectrogram, a Mel-spectrogram, or a combination of a spectrogram and a MFCC as feature parameters [10,14,17,18,22,23,25,[29][30][31][32][33][34].…”
Section: Related Workmentioning
confidence: 99%
“…With the power of technologies, we have today; new novel ways have been introduced to interpret emotions. The speech emotion recognition system is useful in psychiatric diagnosis, lie detection, call center conversations, customer voice review, and voice messages [1].…”
Section: Introductionmentioning
confidence: 99%