2019
DOI: 10.48550/arxiv.1904.04589
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Ensemble Models for Spoofing Detection in Automatic Speaker Verification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(11 citation statements)
references
References 0 publications
0
11
0
Order By: Relevance
“…Confirming the observations made in [9] and [23], the input features are not normalized, but simply scaled down to be in the range from −1 to 1.…”
Section: Audio Preprocessing and Feature Extractionmentioning
confidence: 80%
See 2 more Smart Citations
“…Confirming the observations made in [9] and [23], the input features are not normalized, but simply scaled down to be in the range from −1 to 1.…”
Section: Audio Preprocessing and Feature Extractionmentioning
confidence: 80%
“…Based on existing literature (e.g. [23]), it can be explained that the beginning and tailing silence cues can lead to better performance. Considering these findings and our practical application, we decided to use 8.5s input length and to do cutting and padding at the end of the audio from now on (so that we do not rely on voice activity detection in practical applications).…”
Section: Experimental Setup and Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…There is some previous work that examines the impact of silence in ASVspoof. For example, [8] find that the length of silence in the PA part of ASVspoof 2019 differs between bonafide and spoof data. Additionally, for the ASVspoof 2017 replay detection challenge, [14] observe that the presence of certain attributes such as 'clicks', trailing and leading silence in the audio may hint at the target label.…”
Section: Related Workmentioning
confidence: 97%
“…The recurrent ASVspoof contest challenges researchers to find suitable machine learning algorithms for this task. There is a large body of related work [10,11,12,13], which uses machine learning to identify artifacts in audio waveforms that may indicate a deepfake. Such artifacts include noisy glitch, phase mismatch, reverberation, or loss of intelligibility [14,15], but also artefacts that humans cannot perceive.…”
Section: Introductionmentioning
confidence: 99%