Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.540
|View full text |Cite
|
Sign up to set email alerts
|

Word-level Textual Adversarial Attacking as Combinatorial Optimization

Abstract: Adversarial attacks are carried out to reveal the vulnerability of deep neural networks. Textual adversarial attacking is challenging because text is discrete and a small perturbation can bring significant change to the original input. Word-level attacking, which can be regarded as a combinatorial optimization problem, is a well-studied class of textual attack methods. However, existing word-level attack models are far from perfect, largely because unsuitable search space reduction methods and inefficient opti… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
275
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 260 publications
(293 citation statements)
references
References 45 publications
1
275
0
Order By: Relevance
“…Unlike most work on textual adversarial examples, Morpheus produces its adversaries by exploiting the morphology of the text. Zang et al [ 165 ] suggested applying word substitutions using the minimum semantic units, called sememes. The assumption was that the sememes of a word are indicative of the word’s meaning and, therefore, words with the same sememes should be good substitutes for each another.…”
Section: Different Scopes Of Machine Learning Interpretability: a mentioning
confidence: 99%
“…Unlike most work on textual adversarial examples, Morpheus produces its adversaries by exploiting the morphology of the text. Zang et al [ 165 ] suggested applying word substitutions using the minimum semantic units, called sememes. The assumption was that the sememes of a word are indicative of the word’s meaning and, therefore, words with the same sememes should be good substitutes for each another.…”
Section: Different Scopes Of Machine Learning Interpretability: a mentioning
confidence: 99%
“…Queries Cache hits Alzantot et al (2018) 1029 736 Zang et al (2020) 3745 3080 Table 1: "Queries" stands for average number of queries to victim model to attack one sample, while "cache hits" represents the average number of times a query has resulted in a hit to the model output cache. Each cache hit saves a query to the model, so more cache hits indicates a higher performance boost due to caching.…”
Section: Attackmentioning
confidence: 99%
“…In some cases, this high-level caching can cause a significant performance increase. We experimented with attacking 100 samples for BERT-base model (Devlin et al, 2018) trained on SST-2 dataset (Socher et al, 2013) using methods proposed by Alzantot et al (2018) and Zang et al (2020). Table 1 shows that in both cases, significant number of queries to the victim model result in hits to the model output cache, helping us save time by avoiding unnecessary computations.…”
Section: Attackmentioning
confidence: 99%
See 1 more Smart Citation
“…Researchers have proposed numerous methods to generate adversarial texts, which can be divided into char-level [12], word-level [13], sentence-level [14], and multi-level (i.e., a mixture of the previous three methods) [15,16]. They modify the characters, words, and sentences in the inputs, respectively.…”
Section: Introductionmentioning
confidence: 99%