2021
DOI: 10.48550/arxiv.2106.09898
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Bad Characters: Imperceptible NLP Attacks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
6
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(7 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…Privacy of text is more acutely relevant for most use cases, and prior work [4] suggests that stronger forms of training data recovery are possible on these sorts of models. Coupling that work with the increased complexity in the adversarial example space and recent breakthroughs [2] in adversarial examples on text classifiers make this realm a natural extension of the current work. Additionally, future work might consider how certain defenses against membership inference and adversarial examplesperform when paired with Bayesianism.…”
Section: Discussionmentioning
confidence: 98%
“…Privacy of text is more acutely relevant for most use cases, and prior work [4] suggests that stronger forms of training data recovery are possible on these sorts of models. Coupling that work with the increased complexity in the adversarial example space and recent breakthroughs [2] in adversarial examples on text classifiers make this realm a natural extension of the current work. Additionally, future work might consider how certain defenses against membership inference and adversarial examplesperform when paired with Bayesianism.…”
Section: Discussionmentioning
confidence: 98%
“…Since the original publication of the content in this chapter in 2019, a number of works have followed-up, either by adapting the evaluation framework to other tasks, e.g. semantic parsing in Huang et al (2021), or by building upon it for designing more imperceptible adversarial perturbations, for instance using the proposed evaluation metrics as rewards for reinforcement learning based perturbation generation (Zou et al, 2020), or pushing further the concept of indistinguishable perturbations with encoding specific character substitutions (Boucher et al, 2021).…”
Section: Discussionmentioning
confidence: 99%
“…Boutros et al [4] extended the sponge examples attack so it could be applied on FPGA devices. In [3], the authors presented a method for creating sponge examples that preserve the original input's visual appearance. Cina et al [6] proposed sponge poisoning, a technique that performs sponge attacks during training time, resulting in a poisoned model with decreased performance.…”
Section: Availability-based Attacksmentioning
confidence: 99%
“…Shumailov et al [20] presented sponge examples, which are perturbed inputs designed to increase the energy consumed by natural language processing (NLP) and computer vision models, when deployed on hardware accelerators, by increasing the number of active neurons during classification. Following this work, other studies have proposed sponge-like attacks, mainly targeting image classification models [4,3,6,10].…”
mentioning
confidence: 99%