Abstract. Desktop grids use the free resources in Intranet and Internet environments for large-scale computation and storage. While desktop grids offer a high return on investment, one critical issue is the validation of results returned by participating hosts. Several mechanisms for result validation have been previously proposed. However, the characterization of errors is poorly understood. To study error rates, we implemented and deployed a desktop grid application across several thousand hosts distributed over the Internet. We then analyzed the results to give quantitative and empirical characterization of errors stemming from input or output (I/O) failures. We find that in practice, error rates are widespread across hosts but occur relatively infrequently. Moreover, we find that error rates tend to not be stationary over time nor correlated between hosts. In light of these characterization results, we evaluated state-ofthe-art error detection mechanisms and describe the trade-offs for using each mechanism.
The success of grid computing in open environments like the Internet is highly dependent on the adoption of mechanisms to detect failures and malicious sabotage attempts. It is also required to maintain a trust management system that permits one to distinguish the trustable from the non-trustable participants in a global computation. Without these mechanisms, users with data-critical applications will never rely on desktop grids, and will rather prefer to support higher costs to run their computations in closed and secure computing systems.This paper discusses the topics of sabotage-tolerance and trust management. After reviewing the state-of-the-art, we present two novel techniques: a mechanism for sabotage detection and a protocol for distributed trust management. The proposed techniques are targeted at the paradigm of volunteer-based computing commonly used on desktop grids.
Current infrastructures for Volunteer Computing follow a centralized architecture for data distribution, creating a potential bottleneck when tasks require large input files or the central server has limited bandwidth. In this paper we propose two new data models for Berkeley Open Infrastructure for Network Computing (BOINC): an approach based on the popular BitTorrent protocol; and a Content Delivery Network approach. While the latter remains on a theoretical level, we developed a prototype that adds BitTorrent functionality for task distribution and conducted medium-scale tests of the environment. Our preliminary results indicate that the BitTorrent client had a negligible influence on the BOINC client's computation time. The BOINC server showed an unexpectedly low bandwidth output when seeding the file, as well as spikes on CPU usage. This paper discusses the tests that were performed, how they were evaluated, as well as some improvements that could be made in future research on both approaches.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.