Abstract:The popularity of distributed file systems continues to grow. Reasons they are preferred over traditional centralized file systems include fault tolerance, availability, scalability and performance. In addition, Peer-to-Peer (P2P) system concepts and scalable functions are being incorporated into the domain of file systems. This survey paper explores the design paradigms and important issues that relate to such systems and discusses the various research activities in the field of Distributed Peer-to-Peer file … Show more
“…(However, despite a large churn rate, these studies also show that the total number of peers in the network is relatively stable.) P2P algorithms have been proposed for a wide variety of tasks such as data storage and retrieval [45,19,18,14,26], collaborative filtering [11], spam detection [15], data mining [17], worm detection and suppression [38,59], privacy protection of archived data [23], and recently, for cloud computing services as well [57,8]. However, all algorithms proposed for these problems have no theoretical guarantees of being able to work in a dynamically changing network with a very high churn rate, which can be as much as linear (in the network size) per round.…”
We study robust and efficient distributed algorithms for searching, storing, and maintaining data in dynamic Peer-to-Peer (P2P) networks. P2P networks are highly dynamic networks that experience heavy node churn (i.e., nodes join and leave the network continuously over time). Our goal is to guarantee, despite high node churn rate, that a large number of nodes in the network can store, retrieve, and maintain a large number of data items. Our main contributions are fast randomized distributed algorithms that guarantee the above with high probability even under high adversarial churn. In particular, we present the following main results:1. A randomized distributed search algorithm that with high probability guarantees that searches from as many as n − o(n) nodes (n is the stable network size) succeed in O(log n)-rounds despite O(n/ log 1+δ n) churn, for any small constant δ > 0, per round. We assume that the churn is controlled by an oblivious adversary (that has complete knowledge and control of what nodes join and leave and at what time and has unlimited computational power, but is oblivious to the random choices made by the algorithm).2. A storage and maintenance algorithm that guarantees, with high probability, data items can be efficiently stored (with only Θ(log n) copies of each data item) and maintained in a dynamic P2P network with churn rate up to O(n/ log 1+δ n) per round. Our search algorithm together with our storage and maintenance algorithm guarantees that as many as n − o(n) nodes can efficiently store, maintain, and search even under O(n/ log 1+δ n) churn per round. Our algorithms require only polylogarithmic in n bits to be processed and sent (per round) by each node.To the best of our knowledge, our algorithms are the first-known, fully-distributed storage and search algorithms that provably work under highly dynamic settings (i.e., high churn rates per step). Furthermore, they are localized (i.e., do not require any global topological knowledge) and scalable. A technical contribution of this paper, which may be of independent interest, is
“…(However, despite a large churn rate, these studies also show that the total number of peers in the network is relatively stable.) P2P algorithms have been proposed for a wide variety of tasks such as data storage and retrieval [45,19,18,14,26], collaborative filtering [11], spam detection [15], data mining [17], worm detection and suppression [38,59], privacy protection of archived data [23], and recently, for cloud computing services as well [57,8]. However, all algorithms proposed for these problems have no theoretical guarantees of being able to work in a dynamically changing network with a very high churn rate, which can be as much as linear (in the network size) per round.…”
We study robust and efficient distributed algorithms for searching, storing, and maintaining data in dynamic Peer-to-Peer (P2P) networks. P2P networks are highly dynamic networks that experience heavy node churn (i.e., nodes join and leave the network continuously over time). Our goal is to guarantee, despite high node churn rate, that a large number of nodes in the network can store, retrieve, and maintain a large number of data items. Our main contributions are fast randomized distributed algorithms that guarantee the above with high probability even under high adversarial churn. In particular, we present the following main results:1. A randomized distributed search algorithm that with high probability guarantees that searches from as many as n − o(n) nodes (n is the stable network size) succeed in O(log n)-rounds despite O(n/ log 1+δ n) churn, for any small constant δ > 0, per round. We assume that the churn is controlled by an oblivious adversary (that has complete knowledge and control of what nodes join and leave and at what time and has unlimited computational power, but is oblivious to the random choices made by the algorithm).2. A storage and maintenance algorithm that guarantees, with high probability, data items can be efficiently stored (with only Θ(log n) copies of each data item) and maintained in a dynamic P2P network with churn rate up to O(n/ log 1+δ n) per round. Our search algorithm together with our storage and maintenance algorithm guarantees that as many as n − o(n) nodes can efficiently store, maintain, and search even under O(n/ log 1+δ n) churn per round. Our algorithms require only polylogarithmic in n bits to be processed and sent (per round) by each node.To the best of our knowledge, our algorithms are the first-known, fully-distributed storage and search algorithms that provably work under highly dynamic settings (i.e., high churn rates per step). Furthermore, they are localized (i.e., do not require any global topological knowledge) and scalable. A technical contribution of this paper, which may be of independent interest, is
“…Issues on designing a P2P storage system include redundancy management, access control, load balancing and fast resource location. Hasan et al [40] give a survey on current P2P storage systems and analyze their advantages and disadvantages.…”
“…The main architectural elements include the seed, the tracker, and the leech; see [8], [19], [20] for overviews. At the heart of this content distribution scheme are static files called torrents, which provide information about the file to be distributed, such as, its length and name, along with a hashed value (checksum of the file).…”
Section: Object Resolution In Bittorrentmentioning
The established host-centric networking paradigm is challenged due to handicaps related with disconnected operation, mobility, and broken locator/identifier semantics. This paper soberly examines another topic of great interest: distributed information object resolution. After recapping the notion of an information object, we review object resolution in today's Internet which is based on Uniform Resource Identifiers (URIs). We revisit the implications of DNS involvement in URI resolution and discuss how two different types of content distribution networks work with respect to name resolution. Then we evaluate proposals championing the replacement of DNS with alternatives based on distributed hash tables. We present the pros and cons and highlight the importance of latency in resolution. The paper positions these issues in the context of a Network of Information (NetInf) and concludes with open research topics in the area.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.