Abstract-t-Closeness is a privacy model recently defined for data anonymization. A data set is said to satisfy t-closeness if, for each group of records sharing a combination of key attributes, the distance between the distribution of a confidential attribute in the group and the distribution of the attribute in the entire data set is no more than a threshold t. Here, we define a privacy measure in terms of information theory, similar to t-closeness. Then, we use the tools of that theory to show that our privacy measure can be achieved by the postrandomization method (PRAM) for masking in the discrete case, and by a form of noise addition in the general case.
We address the problem of designing optimal quantizers for distributed source coding. The generality of our formulation includes both the symmetric and asymmetric scenarios, together with a number of coding schemes, such as the ideal coding achieving a rate equal to the joint conditional entropy of the quantized sources given the side information. We show the optimality conditions quantizers must satisfy, and generalize the Lloyd algorithm for its design. Experimental results are shown for the Gaussian scalar asymmetric case.
Abstract-We present a mathematical formulation for the optimization of query forgery for private information retrieval, in the sense that the privacy risk is minimized for a given traffic and processing overhead. The privacy risk is measured as an information-theoretic divergence between the user's query distribution and the population's, which includes the entropy of the user's distribution as a special case. We carefully justify and interpret our privacy criterion from diverse perspectives. Our formulation poses a mathematically tractable problem that bears substantial resemblance with rate-distortion theory.Index Terms-Entropy, Kullback-Leibler divergence, privacy risk, private information retrieval, query forgery.
In recent years, the general interest in routing for vehicular ad hoc networks (VANETs) has increased notably. Many proposals have been presented to improve the behavior of the routing decisions in these very changeable networks. In this paper, we propose a new routing protocol for VANETs that uses four different metrics. which are the distance to destination, the vehicles' density, the vehicles' trajectory and the available bandwidth, making use of the information retrieved by the sensors of the vehicle, in order to make forwarding decisions, minimizing packet losses and packet delay. Through simulation, we compare our proposal to other protocols, such as AODV (Ad hoc On-Demand Distance Vector), GPSR (Greedy Perimeter Stateless Routing), I-GPSR (Improvement GPSR) and to our previous proposal, GBSR-B (Greedy Buffer Stateless Routing Building-aware). Besides, we present a performance evaluation of the individual importance of each metric to make forwarding decisions. Experimental results show that our proposed forwarding decision outperforms existing solutions in terms of packet delivery.
Abstract:Recommendation systems are information-filtering systems that tailor information to users on the basis of knowledge about their preferences. The ability of these systems to profile users is what enables such intelligent functionality, but at the same time, it is the source of serious privacy concerns. In this paper we investigate a privacy-enhancing technology that aims at hindering an attacker in its efforts to accurately profile users based on the items they rate. Our approach capitalizes on the combination of two perturbative mechanisms-the forgery and the suppression of ratings. While this technique enhances user privacy to a certain extent, it inevitably comes at the cost of a loss in data utility, namely a degradation of the recommendation's accuracy. In short, it poses a trade-off between privacy and utility. The theoretical analysis of such trade-off is the object of this work. We measure privacy as the Kullback-Leibler divergence between the user's and the population's item distributions, and quantify utility as the proportion of ratings users consent to forge and eliminate. Equipped with these quantitative measures, we find a closed-form solution to the problem of optimal forgery and suppression of ratings, an optimization problem that includes, as a particular case, the maximization of the entropy of the perturbed profile. We characterize the optimal trade-off surface among privacy, forgery rate and suppression rate, * Some parts of this paper (a reduced version of Sections 1 and 2) were presented at the International Workshop on Data Privacy Management, Leuven, Belgium, September 2011 [1]. The formulation of the trade-off between privacy and utility (Section 3), the current theoretical analysis (Section 4), the experimental work (Section 5), the conclusions (Section 6) and the proofs given in the Appendices are all new work.Entropy 2014, 16 1587 and experimentally evaluate how our approach could contribute to privacy protection in a real-world recommendation system.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.