2020
DOI: 10.2478/popets-2020-0024
|View full text |Cite
|
Sign up to set email alerts
|

Secure and Scalable Document Similarity on Distributed Databases: Differential Privacy to the Rescue

Abstract: Privacy-preserving collaborative data analysis enables richer models than what each party can learn with their own data. Secure Multi-Party Computation (MPC) offers a robust cryptographic approach to this problem, and in fact several protocols have been proposed for various data analysis and machine learning tasks. In this work, we focus on secure similarity computation between text documents, and the application to k-nearest neighbors (k-NN) classification. Due to its non-parametric nature, k-NN presents scal… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 41 publications
0
4
0
Order By: Relevance
“…Securely computing kNNC is another closely related field when data is stored in different local devices. Majority of the frameworks that ensure privacy for kNNC often use some sort of secure multi-party computation (SMC) protocols (Zhan, Chang, and Matwin 2008;Xiong, Chitti, and Liu 2006;Qi and Atallah 2008;Schoppmann et al 2020;Shaul, Feldman, and Rus 2020;Chen et al 2020).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Securely computing kNNC is another closely related field when data is stored in different local devices. Majority of the frameworks that ensure privacy for kNNC often use some sort of secure multi-party computation (SMC) protocols (Zhan, Chang, and Matwin 2008;Xiong, Chitti, and Liu 2006;Qi and Atallah 2008;Schoppmann et al 2020;Shaul, Feldman, and Rus 2020;Chen et al 2020).…”
Section: Related Workmentioning
confidence: 99%
“…In most secure kNNC settings considered in the literature, the goal is to keep the training data secure from the party making the test query (Qi and Atallah 2008;Shaul, Feldman, and Rus 2018;Wu et al 2019) and it is not clear how those approaches extend to the multiparty federated setting where the per-party data (train or test) should remain localized. Of particular relevance is Schoppmann et al (2020) which proposed a scheme to compute a secure inner-product between any test point and all training points (distributed across parties) and then perform a secure top-k protocol to perform kNNC. This procedure explicitly computes the neighbors for a test point, which involves n secure similarity computations for each test point (on top of the secure top-k protocol).…”
Section: Related Workmentioning
confidence: 99%
“…Recently, Schoppmann et al [21] have proposed a secure scheme for classifying collaborated documents, where a k-nearest neighbor classifier is applied to compute the secure similarity of textual documents.…”
Section: Related Workmentioning
confidence: 99%
“…A previous version ofSchoppmann et al (2020) was titled "Private Nearest Neighbors Classification in Federated Databases" (https://eprint.iacr.org/eprint-bin/versions.pl?entry=2018/289) but has since been changed as the focus of the paper was shifted.…”
mentioning
confidence: 99%