Abstract:The detection of outliers in data, while preserving the privacy of individual agents who contributed to the data set, is an increasingly important task when monitoring and controlling large-scale systems. In this paper, we use an algorithm based on the sparse vector technique to perform differentially private outlier detection in multivariate Gaussian signals. Specifically, we derive analytical expressions to quantify the trade-off between detection accuracy and privacy. We validate our analytical results thro… Show more
“…It provides each individual agent with the guarantee that the output of the considered query will not be significantly altered by whether or not they contribute their data, or what value they contribute. Differential privacy can be achieved through input perturbation, output perturbation [11], [17] or the sparse vector technique [18], [19]. In this paper, we consider input perturbation, which has the advantage that each individual agent can perturb its data before sending it to a data aggregator, thereby eliminating the need to trust the aggregator.…”
Section: Introductionmentioning
confidence: 99%
“…scalar quantities are considered, we consider multivariate signals provided by individual agents who may be correlated. Unlike [19] where the sparse vector technique is used, we consider scenarios in which individual agents do not necessarily trust the data aggregator, and therefore we design an input perturbation architecture to guarantee differential privacy for the agents' data. Using the squared Mahalanobis distance, we derive analytical formulas for the trade-offs between the accuracy of detection and privacy level.…”
The detection of outliers has become increasingly important for the control and monitoring of large-scale networked systems such as transportation and smart grids. Data from these systems, such as location traces or power consumption, are collected from individual agents, and are often privacy-sensitive. Furthermore, the networked nature of these systems results in the data of different individuals being correlated with each other. In this paper, we use the concept of differential privacy to design a privacy-preserving algorithm for outlier detection in correlated data. We determine analytic formulas to evaluate the performance of the proposed differentially private algorithm, and we analyze the trade-off between privacy level and detection accuracy. We illustrate our methodology using an example based on outlier detection in household electricity usage data.
“…It provides each individual agent with the guarantee that the output of the considered query will not be significantly altered by whether or not they contribute their data, or what value they contribute. Differential privacy can be achieved through input perturbation, output perturbation [11], [17] or the sparse vector technique [18], [19]. In this paper, we consider input perturbation, which has the advantage that each individual agent can perturb its data before sending it to a data aggregator, thereby eliminating the need to trust the aggregator.…”
Section: Introductionmentioning
confidence: 99%
“…scalar quantities are considered, we consider multivariate signals provided by individual agents who may be correlated. Unlike [19] where the sparse vector technique is used, we consider scenarios in which individual agents do not necessarily trust the data aggregator, and therefore we design an input perturbation architecture to guarantee differential privacy for the agents' data. Using the squared Mahalanobis distance, we derive analytical formulas for the trade-offs between the accuracy of detection and privacy level.…”
The detection of outliers has become increasingly important for the control and monitoring of large-scale networked systems such as transportation and smart grids. Data from these systems, such as location traces or power consumption, are collected from individual agents, and are often privacy-sensitive. Furthermore, the networked nature of these systems results in the data of different individuals being correlated with each other. In this paper, we use the concept of differential privacy to design a privacy-preserving algorithm for outlier detection in correlated data. We determine analytic formulas to evaluate the performance of the proposed differentially private algorithm, and we analyze the trade-off between privacy level and detection accuracy. We illustrate our methodology using an example based on outlier detection in household electricity usage data.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.