2022
DOI: 10.1609/aaai.v36i2.20036
|View full text |Cite
|
Sign up to set email alerts
|

Defending against Model Stealing via Verifying Embedded External Features

Abstract: Obtaining a well-trained model involves expensive data collection and training procedures, therefore the model is a valuable intellectual property. Recent studies revealed that adversaries can `steal' deployed models even when they have no training samples and can not get access to the model parameters or structures. Currently, there were some defense methods to alleviate this threat, mostly by increasing the cost of model stealing. In this paper, we explore the defense from another angle by verifying whether … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 30 publications
(18 citation statements)
references
References 25 publications
0
10
0
Order By: Relevance
“…6. [92] propose to embed external features into the host model by embedding a few images modified via style transfer algorithm. A binary meta-classifier is also trained on the gradients of model weights (i.e., both host model and a benign model trained with clean data) using transformed images to extract the embedded external features.…”
Section: B Watermark-based Solutions To Ip Protection In Aigcmentioning
confidence: 99%
“…6. [92] propose to embed external features into the host model by embedding a few images modified via style transfer algorithm. A binary meta-classifier is also trained on the gradients of model weights (i.e., both host model and a benign model trained with clean data) using transformed images to extract the embedded external features.…”
Section: B Watermark-based Solutions To Ip Protection In Aigcmentioning
confidence: 99%
“…For example, defenders may round the probability vectors [4], introduce noise to the output vectors which will result in a high loss in the processes of model stealing [7], or only return the most confident label instead of the whole output vector [5]. However, these defenses may significantly reduce the performance of victim models and may even be bypassed by adaptive attacks [10], [11], [12].…”
Section: Active Defensesmentioning
confidence: 99%
“…For example, defenders can introduce randomness or perturbations in the victim models [4], [7], [8] or watermark the victim model via (targeted) backdoor attacks or data poisoning [9], [10], [11]. However, existing active defenses may lead to poor performance of the victim model and could even be bypassed by advanced adaptive attacks [10], [11], [12]; the verification-based methods target only limited simple stealing scenarios (e.g., direct copy or fine-tuning) and have minor effects in defending against more complicated model stealing. Besides, these methods also introduce some stealthy latent short-cuts (e.g., hidden backdoors) in the victim model, which could be maliciously used.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The FSL approach involves using a large auxiliary set of labeled data from disjoint classes to acquire transferable knowledge or representations that can help in the few-shot tasks. Recently, the security implications of FSL have been brought to the forefront of the community (Li et al 2022a;Guan et al 2022), such as the challenge of training a robust few-shot model against adversarial attacks (Li et al 2019b;Jia et al 2020;Huang et al 2021Huang et al , 2023.…”
Section: Introductionmentioning
confidence: 99%