2016
DOI: 10.1109/tmm.2016.2557721
|View full text |Cite
|
Sign up to set email alerts
|

Sparse Kernel Reduced-Rank Regression for Bimodal Emotion Recognition From Facial Expression and Speech

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
42
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 90 publications
(43 citation statements)
references
References 50 publications
1
42
0
Order By: Relevance
“…It is clear from these results that a SVM learns better from higher-dimensional feature sets such as the ComParE and the STC sets, which is also a consistent phenomenon observed in [5]. Yan et al [29] recently published a baseline result on the eNTERFACE'05 corpus using the PC feature set. They trained a SVM classifier on the PC feature set with a speaker-dependent five-fold cross validation evaluation strategy as one of their baseline models.…”
Section: Resultssupporting
confidence: 63%
See 2 more Smart Citations
“…It is clear from these results that a SVM learns better from higher-dimensional feature sets such as the ComParE and the STC sets, which is also a consistent phenomenon observed in [5]. Yan et al [29] recently published a baseline result on the eNTERFACE'05 corpus using the PC feature set. They trained a SVM classifier on the PC feature set with a speaker-dependent five-fold cross validation evaluation strategy as one of their baseline models.…”
Section: Resultssupporting
confidence: 63%
“…For the ease to compare models, Eyben et al [5] summarized the performances by a SVM trained on the INTERSPEECH challenge feature sets over several public corpora. Yan et al [29] recently proposed a sparse kernel reduced-rank regression (SKRRR) for bimodal emotion recognition from facial expressions and speech, which has achieved one of the state-of-the-art performances on the eNTERFACE'05 [30] corpus.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…where M ∈ R p×N denotes the facial image matrix, ψ(M) ∈ R p φ ×N denotes the mapped facial image matrix, E ∈ R p×c is the projection matrix of the LPP method, B ∈ R p ϕ ×c is the projection matrix of ψ(M), p and p ϕ denote the dimension of the facial image matrix and the mapped facial image matrix respectively, N and c denote the number of the facial image and projection vectors respectively [1], [14],…”
Section: Regression-based Robust Locality Preserving Projectionsmentioning
confidence: 99%
“…where M L and M S is denoted as the low-rank term and the sparse term of the facial image matrix M respectively [3], [4], [8]- [10], α MS is the sparse parameter of the sparse term M S [1], [14]. The same to the RPCA [9], [10] and the RR [3], [4] method, the formula of (3) also can be written as the regression modal of (4) arg min…”
Section: Regression-based Robust Locality Preserving Projectionsmentioning
confidence: 99%