2018 IEEE Security and Privacy Workshops (SPW) 2018
DOI: 10.1109/spw.2018.00009
|View full text |Cite
|
Sign up to set email alerts
|

Audio Adversarial Examples: Targeted Attacks on Speech-to-Text

Abstract: We construct targeted audio adversarial examples on automatic speech recognition. Given any audio waveform, we can produce another that is over 99.9% similar, but transcribes as any phrase we choose (recognizing up to 50 characters per second of audio). We apply our white-box iterative optimization-based attack to Mozilla's implementation DeepSpeech end-to-end, and show it has a 100% success rate. The feasibility of this attack introduce a new domain to study adversarial examples. 1

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

12
809
1

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 817 publications
(862 citation statements)
references
References 41 publications
12
809
1
Order By: Relevance
“…One of the most successful white-box attacking methods is C&W [9]. This method uses Connectionist Temporal Classification (CTC) loss function [26] for perturbation optimization.…”
Section: Attacking Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…One of the most successful white-box attacking methods is C&W [9]. This method uses Connectionist Temporal Classification (CTC) loss function [26] for perturbation optimization.…”
Section: Attacking Methodsmentioning
confidence: 99%
“…Earlier adversarial attacks were applied on machine learning models of image domain [4,5,6,7,8] and then these attacking methods have been spread out onto other domains, e.g. speech signals [9,10,11,12,13]. The adversary adds a very small optimized perturbation, which is not detectable by human, to a legitimate input and generates an adversarial example that results the learning model to return a wrong output.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Thus, simply playing the pre-generated universal perturbation nearby the victim speaker becomes possible for launching adversarial attacks. For showing the possibility of launching real-time attacks, we compare the attack launching time of using the conventional individual targeted attack method [6] and our proposed universal attack for a given audio signal. Particularly, the conventional targeted attack requires at least 15s to deploy, measured on a Tesla V100 GPU with 32GB memory, while our proposed universal method only takes an average of 0.015s, which results in a 100× speedup.…”
Section: Attack Evaluationmentioning
confidence: 99%
“…Carlini & Wagner, 2018). While we would not expect this to be a widespread problem in typical online experimental settings, researchers are also beginning to devise strategies for counteracting adversarial examples (Madry et al, 2017).…”
Section: Speech-to-text Engines As a Driver For Scalable Online Verbamentioning
confidence: 99%