2021
DOI: 10.1016/j.csl.2021.101199
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial attack and defense strategies for deep speaker recognition systems

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
63
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 53 publications
(64 citation statements)
references
References 6 publications
1
63
0
Order By: Relevance
“…This category shows more resistance to voice synthesis and conversion spoofing attacks, but it is valuable to human impersonation. Employing this feature in speaker recognition systems provides robustness to adversarial attacks to which most deep speaker recognition systems are prone to [ 40 ]. One can argue that a speaker recognition system that employs short-term spectral and prosodic features is more robust than those systems that employ only one type of these features.…”
Section: Related Workmentioning
confidence: 99%
“…This category shows more resistance to voice synthesis and conversion spoofing attacks, but it is valuable to human impersonation. Employing this feature in speaker recognition systems provides robustness to adversarial attacks to which most deep speaker recognition systems are prone to [ 40 ]. One can argue that a speaker recognition system that employs short-term spectral and prosodic features is more robust than those systems that employ only one type of these features.…”
Section: Related Workmentioning
confidence: 99%
“…f (•) is a well-trained classifier. The authors [78] provide a wide range of investigation related to the topic and pointed out such a strategy can largely infect the speaker recognition methods. However, adversarial training can considerably defend such perturbation attacks [79].…”
Section: Figure 16mentioning
confidence: 99%
“…• Tamper proof: Speaker recognition systems should be tamper-proof against adversarial attacks and voice mimicry. In some cases, adversarial attacks require extensive investigation over the available architectures resulting in hard to generate [78]. However, voice mimicry does not significantly affect speaker recognition models [126].…”
Section: Metamentioning
confidence: 99%
“…As a result of the rapid development of modern speech processing technology, people can easily edit "false" audio that cannot be easily distinguished by hearing, which makes it more difficult for people to distinguish the true and false audio in social networks. Therefore, the application of machine learning methods to mine hidden information in acoustic signals to identify authenticity and monitor risks [4,5] has become a research hotspot in the field of acoustic signal processing. The special complex noise and sudden acoustic events in ambient background sounds are some of the important factors to judge audio authenticity.…”
Section: Introductionmentioning
confidence: 99%