ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019
DOI: 10.1109/icassp.2019.8682571
|View full text |Cite
|
Sign up to set email alerts
|

Formant-gaps Features for Speaker Verification Using Whispered Speech

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(2 citation statements)
references
References 29 publications
0
2
0
Order By: Relevance
“…Current mainstream WSR methods can be broadly classified into three categories. The first class of methods focuses on feature extraction [14][15][16][17][18], such as the Mel-scale Frequency Cepstral Coefficient (MFCC), Weighted Instantaneous Frequency, Auditory-inspired Amplitude Modulation Features (AAMFs), Formant-gaps (FoGs) features, and Weighted Modified Linear frequency cepstrum coefficients. Among them, MFCC has been used with GMM and has been used for the speaker recognition task.…”
Section: Introductionmentioning
confidence: 99%
“…Current mainstream WSR methods can be broadly classified into three categories. The first class of methods focuses on feature extraction [14][15][16][17][18], such as the Mel-scale Frequency Cepstral Coefficient (MFCC), Weighted Instantaneous Frequency, Auditory-inspired Amplitude Modulation Features (AAMFs), Formant-gaps (FoGs) features, and Weighted Modified Linear frequency cepstrum coefficients. Among them, MFCC has been used with GMM and has been used for the speaker recognition task.…”
Section: Introductionmentioning
confidence: 99%
“…Speakers often use whispered speech in their day-to-day life. With the recent advancements in speech-enabled and virtual assistant devices [7], robust methods for whispered speech detection [8], recognition, and whispered speechbased speaker verification [9,10] had been developed. This motivated us to consider whispered speech together with neutral speech for recording device classification.…”
Section: Introductionmentioning
confidence: 99%