2021
DOI: 10.48550/arxiv.2111.01528
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

HydraText: Multi-objective Optimization for Adversarial Textual Attack

Abstract: The field of adversarial textual attack has significantly grown over the last years, where the commonly considered objective is to craft adversarial examples that can successfully fool the target models. However, the imperceptibility of attacks, which is also an essential objective, is often left out by previous studies. In this work, we advocate considering both objectives at the same time, and propose a novel multi-optimization approach (dubbed HydraText) with provable performance guarantee to achieve succes… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 41 publications
0
2
0
Order By: Relevance
“…A decision maker can then select a final solution based on their requirements. Compared with traditional optimization methods for multiple objectives (e.g., weighted-sum [37]), Pareto optimization algorithms are designed via different meta-heuristics without any trade-off parameters [62,36].…”
Section: Related Workmentioning
confidence: 99%
“…A decision maker can then select a final solution based on their requirements. Compared with traditional optimization methods for multiple objectives (e.g., weighted-sum [37]), Pareto optimization algorithms are designed via different meta-heuristics without any trade-off parameters [62,36].…”
Section: Related Workmentioning
confidence: 99%
“…Deep neural networks (DNNs) have achieved significant progress in wide applications, such as image classification [9], face recognition [27], object detection [28], speech recognition [14] and machine translation [3]. Despite their success, deep learning models have exhibited vulnerability to adversarial attacks [12,20,21,36]. Crafted by adding some small perturbations to benign inputs, adversarial examples (AEs) can fool DNNs into making wrong predictions, which is a critical threat especially for some security-sensitive scenarios such as autonomous driving [34].…”
Section: Introductionmentioning
confidence: 99%