2018
DOI: 10.48550/arxiv.1806.09030
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On Adversarial Examples for Character-Level Neural Machine Translation

Abstract: Evaluating on adversarial examples has become a standard procedure to measure robustness of deep learning models. Due to the difficulty of creating white-box adversarial examples for discrete text input, most analyses of the robustness of NLP models have been done through blackbox adversarial examples. We investigate adversarial examples for character-level neural machine translation (NMT), and contrast black-box adversaries with a novel white-box adversary, which employs differentiable string-edit operations … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
24
0
5

Year Published

2019
2019
2022
2022

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 71 publications
(29 citation statements)
references
References 15 publications
0
24
0
5
Order By: Relevance
“…This work is related to work on adversarial attacks in the text domain, which can be roughly divided into the following categories. One category is adversarial misspelling, which tries to evade the classifier by some "human-imperceptible" misspelling on certain selected characters [14], [17], [67]. The core idea is to design a strategy to identify the important positions and afterwards some standard character-level operations like insertion, deletion, substitution and swap can be applied.…”
Section: Related Workmentioning
confidence: 99%
“…This work is related to work on adversarial attacks in the text domain, which can be roughly divided into the following categories. One category is adversarial misspelling, which tries to evade the classifier by some "human-imperceptible" misspelling on certain selected characters [14], [17], [67]. The core idea is to design a strategy to identify the important positions and afterwards some standard character-level operations like insertion, deletion, substitution and swap can be applied.…”
Section: Related Workmentioning
confidence: 99%
“…Typically, a target word is replaced by an equivalent one chosen from a space of possible replacements. Such a space was identified by Ebrahimi et al [83] and referred to as the embedding space. In [83], the authors consider neural machine translation (NMT) and propose elementary modifications to achieve word-level adversarial attacks.…”
Section: Optimizing Adversarial Attacksmentioning
confidence: 99%
“…Such a space was identified by Ebrahimi et al [83] and referred to as the embedding space. In [83], the authors consider neural machine translation (NMT) and propose elementary modifications to achieve word-level adversarial attacks. In this regard, the authors extend the HotFLip algorithm proposed in [56] adding, removing, or replacing a word in the translation input.…”
Section: Optimizing Adversarial Attacksmentioning
confidence: 99%
See 1 more Smart Citation
“…Instead of designing error features, recent researchers adopt ideas from adversarial learning (Goodfellow, Shlens, and Szegedy, 2014) to generate adversarial samples to mine NLP system pitfalls (Cheng et al, 2018a;Ebrahimi, Lowd, and Dou, 2018;Zhao, Dua, and Singh, 2017). Adversarial samples are minor perturbed inputs which keep the semantic meaning of the input, yet yield degraded outputs.…”
Section: Introductionmentioning
confidence: 99%