Do you want to publish a course? Click here

A continuous mathematical model of fault tolerance mechanisms for parallel applications

نموذج رياضي مستمر للتسامح مع الأعطال في التطبيقات المتوازية

1271   1   12   0 ( 0 )
 Publication date 2016
and research's language is العربية
 Created by Shamra Editor




Ask ChatGPT about the research

In this paper, we introduce a continuous mathematical model to optimize the compromise between the overhead of fault tolerance mechanism and the faults impacts in the environment of execution. The fault tolerance mechanism considered in this research is a coordinated checkpoint/recovery mechanism and the study based on stochastic model of different performance critics of parallel application on parallel and distributed environment.

References used
Feitelson D.G,2005-The supercomputer industry in light of the top500 data Computing in Science Engineering,7(1):42-47
Oldeld R.A.and all, 2007-Modeling the impact of checkpoints on next-generation systems, 24th IEEE conference on mass storage systems and technologies, pages30–46
Cappello F., Geist A., Gropp B., Kale L., Kramer B. and Snir M., 2009-Toward exascale resilience, International Journal of High Performance Computing Applications, 23(4) :374
rate research

Read More

In this paper we present a study on the time cost added to the grid computing as a result of the use of a coordinated checkpoint / recovery fault tolerance protocol, we aim to find a mathematical model which determined the suitable time to save t he checkpoints for application, to achieve a minimum finish time of parallel application in grid computing with faults and fault tolerance protocols, we have find this model by serial modeling to the goal errors, execution environment and the chosen fault tolerance protocol all that by Kolmogorov differential equations.
In this research, We introduce two probabilistic mechanisms to certificate parallel applications on distribute architecture supposing that there are no oracles on which we depend on certification, in addition to introducing cost model of two mecha nisms and compare them. In this research, we are interested in parallel applications, which are represented by data-flow graph that is built dynamically during the execution and which are executed in a wide distributed heterogeneous and dynamic environment and these applications use the principle of work stealing to distribute the tasks among the processors.
Overlay multicast (Application-Level Multicast (ALM)) constructs a multicast delivery tree among end hosts. Unlike traditional IP multicast where the internal tree nodes are dedicated routers which are relatively stable and do not leave the multicast tree voluntarily, the non-leaf nodes in the overlay tree are free end hosts which can join/leave the overlay at will, or even crash without notification. So, the leaving node can leave suddenly and cannot give its descendants (and the Rendez-vous Point (RP)) the time to prepare the recovering (the reconnection) of the overlay tree, and so there is a need to trigger a rearrangement process in which each one of its descendants should rejoin the overlay tree. In this case, all of its downstream nodes are partitioned from the overlay tree and cannot get the multicast data any more. These dynamic characteristics cause the instability of the overlay tree, which can significantly impact the user. A key challenge in constructing an efficient and resilient ALM protocol is to provide fast data recovery when overlay node failures partition the data delivery paths. In this paper, we analyze the performance of the ALM tree recovery solutions using different metrics.
We introduce an auto adaptive strategy enables to write a parallel algorithm adapts to the number of available resources at allocated parallel environment to execute the parallel program. The parallel applications we are studying which are represe nted by data-flow graph which built dynamically during the execution. The new suggested strategy is based on coupling of a sequential algorithm and a parallel one and relies on the principle of work stealing in the tasks scheduling. We offer a study of the complexity of the adaptive algorithm and analyze its performance on processors and compare it with a performance of a classic parallel algorithm.
The NonUniform Memory Access (NUMA) machines are distributed shared memory systems. In this paper, we extend conventional virtual memory concepts to describe the status of memory in distributed, NUMA machines. We present a mathematical model for v irtual memory systems in centralized systems and NUMA machines. The model will show the status of memory in response to memory references.
comments
Fetching comments Fetching comments
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا