2015
DOI: 10.1016/j.engappai.2015.01.010
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised rapid speaker adaptation based on selective eigenvoice merging for user-specific voice interaction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
4

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 31 publications
0
2
0
Order By: Relevance
“…Biometrics may use physical or behavioral characteristics for identification purposes, and different alternatives have been explored over the years: fingerprint [3,4,5], hand geometry [6,7], palmprint [8], voice [9,10], face [11,12,13], and handwritten signature [14]. Among those, face stands out for its acceptability and recognition cost, turning out to be one of the best option for a wide range of applications, from low-security uses (e.g., social media and smartphone access control) to high-security applications (e.g., border control and video surveillance in critical places).…”
Section: Introductionmentioning
confidence: 99%
“…Biometrics may use physical or behavioral characteristics for identification purposes, and different alternatives have been explored over the years: fingerprint [3,4,5], hand geometry [6,7], palmprint [8], voice [9,10], face [11,12,13], and handwritten signature [14]. Among those, face stands out for its acceptability and recognition cost, turning out to be one of the best option for a wide range of applications, from low-security uses (e.g., social media and smartphone access control) to high-security applications (e.g., border control and video surveillance in critical places).…”
Section: Introductionmentioning
confidence: 99%
“…Finally, the SA model represents a model transformed from the SI model according to speaker adaptation procedures. The adaptation only requires a relatively small amount of data (called adaptation data) obtained from the user (called the target speaker), but produces the usercharacterized acoustic model, nearly achieving the performance of the SD model (Matsui and Furui, 1998;Choi et al, 2015).…”
Section: Introductionmentioning
confidence: 99%