ترغب بنشر مسار تعليمي؟ اضغط هنا

An Adaptive Checkpointing Scheme for Peer-to-Peer Based Volunteer Computing Work Flows

547   0   0.0 ( 0 )
 نشر من قبل Lei Ni
 تاريخ النشر 2007
  مجال البحث الهندسة المعلوماتية
والبحث باللغة English




اسأل ChatGPT حول البحث

Volunteer Computing, sometimes called Public Resource Computing, is an emerging computational model that is very suitable for work-pooled parallel processing. As more complex grid applications make use of work flows in their design and deployment it is reasonable to consider the impact of work flow deployment over a Volunteer Computing infrastructure. In this case, the inter work flow I/O can lead to a significant increase in I/O demands at the work pool server. A possible solution is the use of a Peer-to- Peer based parallel computing architecture to off-load this I/O demand to the workers; where the workers can fulfill some aspects of work flow coordination and I/O checking, etc. However, achieving robustness in such a large scale system is a challenging hurdle towards the decentralized execution of work flows and general parallel processes. To increase robustness, we propose and show the merits of using an adaptive checkpoint scheme that efficiently checkpoints the status of the parallel processes according to the estimation of relevant network and peer parameters. Our scheme uses statistical data observed during runtime to dynamically make checkpoint decisions in a completely de- centralized manner. The results of simulation show support for our proposed approach in terms of reduced required runtime.



قيم البحث

اقرأ أيضاً

Scalability and efficient global search in unstructured peer-to-peer overlays have been extensively studied in the literature. The global search comes at the expense of local interactions between peers. Most of the unstructured peer-to-peer overlays do not provide any performance guarantee. In this work we propose a novel Quality of Service enabled lookup for unstructured peer-to-peer overlays that will allow the users query to traverse only those overlay links which satisfy the given constraints. Additionally, it also improves the scalability by judiciously using the overlay resources. Our approach selectively forwards the queries using QoS metrics like latency, bandwidth, and overlay link status so as to ensure improved performance in a scenario where the degree of peer joins and leaves are high. User is given only those results which can be downloaded with the given constraints. Also, the protocol aims at minimizing the message overhead over the overlay network.
We study three capacity problems in the mobile telephone model, a network abstraction that models the peer-to-peer communication capabilities implemented in most commodity smartphone operating systems. The capacity of a network expresses how much sus tained throughput can be maintained for a set of communication demands, and is therefore a fundamental bound on the usefulness of a network. Because of this importance, wireless network capacity has been active area of research for the last two decades. The three capacity problems that we study differ in the structure of the communication demands. The first problem is pairwise capacity, where the demands are (source, destination) pairs. Pairwise capacity is one of the most classical definitions, as it was analyzed in the seminal paper of Gupta and Kumar on wireless network capacity. The second problem we study is broadcast capacity, in which a single source must deliver packets to all other nodes in the network. Finally, we turn our attention to all-to-all capacity, in which all nodes must deliver packets to all other nodes. In all three of these problems we characterize the optimal achievable throughput for any given network, and design algorithms which asymptotically match this performance. We also study these problems in networks generated randomly by a process introduced by Gupta and Kumar, and fully characterize their achievable throughput. Interestingly, the techniques that we develop for all-to-all capacity also allow us to design a one-shot gossip algorithm that runs within a polylogarithmic factor of optimal in every graph. This largely resolves an open question from previous work on the one-shot gossip problem in this model.
Scalability and security problems of the centralized architecture models in cyberphysical systems have great potential to be solved by novel blockchain based distributed models.A decentralized energy trading system takes advantage of various sources and effectively coordinates the energy to ensure optimal utilization of the available resources. It achieves that goal by managing physical, social and business infrastructures using technologies such as Internet of Things (IoT), cloud computing and network systems. Addressing the importance of blockchain-enabled energy trading in the context of cyberphysical systems, this article provides a thorough overview of the P2P energy trading and the utilization of blockchain to enhance the efficiency and the overall performance including the degree of decentralization, scalability and the security of the systems. Three blockchain based energy trading models have been proposed to overcome the technical challenges and market barriers for better adoption of this disruptive technology.
A multiagent based model for a system of cooperative agents aiming at growth is proposed. This is based on a set of generalized Verhulst-Lotka-Volterra differential equations. In this study, strong cooperation is allowed among agents having similar s izes, and weak cooperation if agent have markedly different sizes, thus establishing a peer-to-peer modulated interaction scheme. A rigorous analysis of the stable configurations is presented first examining the fixed points of the system, next determining their stability as a function of the model parameters. It is found that the agents are self-organizing into clusters. Furthermore, it is demonstrated that, depending on parameter values, multiple stable configurations can coexist. It occurs that only one of them always emerges with probability close to one, because its associated attractor dominates over the rest. This is shown through numerical integrations and simulations,after analytic developments. In contrast to the competitive case, agents are able to increase their capacity beyond the no-interaction case limit. In other words, when some collaborative partnership among a relatively small number of partners takes place, all agents act in good faith prioritizing the common good, whence receiving a mutual benefit allowing them to surpass their capacity.
Random hashing is a standard method to balance loads among nodes in Peer-to-Peer networks. However, hashing destroys locality properties of object keys, the critical properties to many applications, more specifically, those that require range searchi ng. To preserve a key order while keeping loads balanced, Ganesan, Bawa and Garcia-Molina proposed a load-balancing algorithm that supports both object insertion and deletion that guarantees a ratio of 4.237 between the maximum and minimum loads among nodes in the network using constant amortized costs. However, their algorithm is not straightforward to implement in real networks because it is recursive. Their algorithm mostly uses local operations with global max-min load information. In this work, we present a simple non-recursive algorithm using essentially the same primitive operations as in Ganesan {em et al.}s work. We prove that for insertion and deletion, our algorithm guarantees a constant max-min load ratio of 7.464 with constant amortized costs.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا