Federated Learning (FL) is a promising paradigm, where the local users collaboratively learn models by repeatedly sharing information while the data is kept distributing on these users. FL has been considered in multiple access channels (FL-MAC), which is a hot issue. Even though FL-MAC has many advantages, it is still possible to leak privacy to a third party during the whole training process. To avoid privacy leakage, we propose to add Rényi differential privacy (RDP) into FL-MAC. At the same time, to maximize the convergent rate of users under the constraints of transmission rate and privacy, the quantization stochastic gradient descent (QSGD) is performed by users. We also illustrate our results on MNIST, and the illustration demonstrate that our scheme can improve the model accuracy with a little loss of communication efficiency.
The emerging of shuffle model has attracted considerable attention of scientists owing to his unique properties in solving the privacy problems in federated learning, specifically the trade off problem between privacy and utility in central and local model. Where, the central model relies on a trusted server which collects users’ raw data and then perturbs it. While in the local model all users perturb their data locally then they send their perturbed data to server. Both models have pron and con. The server in central model enjoys with high accuracy but the users suffer from insufficient privacy in contrast, the local model which provides sufficient privacy at users’ side but the server suffers from limited accuracy. Shuffle model has advanced property of hide position of input messages by perturbing it with perturbation π. Therefore, the scientists considered on adding shuffle model between users and servers to make the server untrusted where the users communicate with the server through the shuffle and boosting the privacy by adding perturbation π for users’ messages without increasing the noise level. Consequently, the usage of modified technique differential privacy federated learning with shuffle model will explores the gap between privacy and accuracy in both models. So this new model attracted many researchers in recent work. In this review, we initiate the analytic learning of a shuffled model for distributed differentially private mechanisms. We focused on the role of shuffle model for solving the problem between privacy and accuracy by summarizing the recent researches about shuffle model and its practical results. Furthermore, we present two types of shuffle, single shuffle and m shuffles with the statistical analysis for each one in boosting the privacy amplification of users with the same level of accuracy by reasoning the practical results of recent papers.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.