Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval 2022
DOI: 10.1145/3477495.3531820
|View full text |Cite
|
Sign up to set email alerts
|

Unlearning Protected User Attributes in Recommendations with Adversarial Training

Abstract: Collaborative filtering algorithms capture underlying consumption patterns, including the ones specific to particular demographics or protected information of users, e. g., gender, race, and location. These encoded biases can influence the decision of a recommendation system (RS) towards further separation of the contents provided to various demographic subgroups, and raise privacy concerns regarding the disclosure of users' protected attributes. In this work, we investigate the possibility and challenges of r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 16 publications
(10 citation statements)
references
References 26 publications
0
10
0
Order By: Relevance
“…It is well-known that existing LMs trained on English text encode societal biases (Bolukbasi et al, 2016;Caliskan et al, 2017;Rekabsaz et al, 2021b) and stereotypes and using them in downstream tasks might lead to unfair treatment of various social groups (Zerveas et al, 2022;Krieg et al, 2022;Ganhör et al, 2022;Rekabsaz et al, 2021a;Melchiorre et al, 2021;Rekabsaz and Schedl, 2020;Elazar and Goldberg, 2018). Since we propose a method to transfer the English LMs to new languages, it is highly probable that the existing biases are also transferred to the target LMs.…”
Section: Risksmentioning
confidence: 98%
“…It is well-known that existing LMs trained on English text encode societal biases (Bolukbasi et al, 2016;Caliskan et al, 2017;Rekabsaz et al, 2021b) and stereotypes and using them in downstream tasks might lead to unfair treatment of various social groups (Zerveas et al, 2022;Krieg et al, 2022;Ganhör et al, 2022;Rekabsaz et al, 2021a;Melchiorre et al, 2021;Rekabsaz and Schedl, 2020;Elazar and Goldberg, 2018). Since we propose a method to transfer the English LMs to new languages, it is highly probable that the existing biases are also transferred to the target LMs.…”
Section: Risksmentioning
confidence: 98%
“…Compared with in-training setting (InT-AU), the post-training setting (PoT-AU) is more challenging. Firstly, PoT-AU allows no interference with the training process, which means InT-AU methods, i.e., adding network block [24], and adversarial training [18], are not applicable. Secondly, even though PoT-AU cuts down the connection with the training process, directly manipulating user embedding by adding artificially-designed noise, e.g., differential privacy [1], is inappropriate, because i) it will inevitably degrade recommendation performance, and ii) its unlearning ability is not promising, as the functional mechanism of attacking models, including complex machine learning models, is not well-understood.…”
Section: Motivationmentioning
confidence: 99%
“…Experiments are conducted on three publicly accessible datasets that contain both input data, i.e., user-item interactions, and user attributes, i.e., gender. Following [18], the provided gender information of the users are limited to females and males.…”
Section: Experimental Settingsmentioning
confidence: 99%
See 2 more Smart Citations