2021
DOI: 10.48550/arxiv.2107.08909
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MEGEX: Data-Free Model Extraction Attack against Gradient-Based Explainable AI

Takayuki Miura,
Satoshi Hasegawa,
Toshiki Shibahara

Abstract: The advance of explainable artificial intelligence, which provides reasons for its predictions, is expected to accelerate the use of deep neural networks in the real world like Machine Learning as a Service (MLaaS) that returns predictions on queried data with the trained model. Deep neural networks deployed in MLaaS face the threat of model extraction attacks. A model extraction attack is an attack to violate intellectual property and privacy in which an adversary steals trained models in a cloud using only t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
9
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(13 citation statements)
references
References 12 publications
(32 reference statements)
0
9
0
Order By: Relevance
“…Unlike adversarial attacks [14,19,42], which try to undermine the performance and credibility of the target model, privacy attacks aim to violate the target model's privacy by abusing its permissions. Model stealing attack [43,67,72,88], which steals various components of a black-box machine learning(ML) model(e.g. hyperparameters [75], architecture [46]), is one of the most common privacy attacks.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Unlike adversarial attacks [14,19,42], which try to undermine the performance and credibility of the target model, privacy attacks aim to violate the target model's privacy by abusing its permissions. Model stealing attack [43,67,72,88], which steals various components of a black-box machine learning(ML) model(e.g. hyperparameters [75], architecture [46]), is one of the most common privacy attacks.…”
Section: Related Workmentioning
confidence: 99%
“…For instance, Zhang et al [44] proposes to use randomized recommendation lists to resist membership inference attacks on recommender systems. Model stealing attack [14,20,32,33] aims to steal internal information of the target model, including hyperparameters [34], architecture [23], etc. Model stealing attacks can also be used to realize functional stealing attacks [11,14,24], which means building a clone model to imitate the predictions of the target model.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Miura et al [84] proposed a data-free model extraction (DFME) attack called MEGEX. The objective of that study was to clone a model without the initial dataset using both the prediction and explanation of the results.…”
Section: Privacymentioning
confidence: 99%
“…for each feature 𝑖. A high partial differential value indicates that a pixel significantly affects the prediction, and analysing the map these values (so-called gradient map) can explain a model's decision-making [125]. Shrikumar et al [168] suggest enhancing numerical explanations by using the input feature value multiplied by the gradient, 𝜙 𝑖 (𝑥) = 𝑥 𝑖 × 𝜕𝑓 𝜕𝑥 𝑖 (𝑥).…”
mentioning
confidence: 99%