MapReduce is the efficient framework for parallel processing of distributed big data in cluster environment. In such a cluster, task failures can impact on performance of applications. Although MapReduce automatically reschedules the failed tasks, it takes long completion time because it starts from scratch. The checkpointing mechanism is the valuable technique to avoid reexecution of failed tasks in MapReduce. However, defining incorrect checkpoint interval can still decrease the performance of MapReduce applications and job completion time. In this paper, the optimum checkpoint interval is proposed to reduce MapReduce job completion time when failures occur. The proposed system defines checkpoint interval that is based on five parameters:...
Le passage de l'échelle des nouvelles plates-formes de calcul parallèle et distribué soulève de nomb...
Researchers have mentioned that the three most difficult and growing problems in the future of high-...
By leveraging the enormous amount of computational capabilities, scientists today are being able to ...
MapReduce is the efficient framework for parallel processing of distributed big data in cluster envi...
The MapReduce has become popular in big data environment due to its efficient parallel processing. H...
[[abstract]]The computing paradigm of MapReduce has gained extreme popularity in the area of large-s...
High-Performance Computing (HPC) has passed the Petascale mark and is moving forward to Exascale. As...
Due to the growing size of compute clusters, large scale parallel applications increasingly have to ...
International audienceThis work provides an analysis of checkpointing strategies for minimizing expe...
Since the last decade, computing systems turn to large scale parallel platforms composed of thousand...
International audienceOmission failures represent an important source of problems in data-intensive ...
This report provides an introduction to the design of scheduling algorithms to cope with faults on l...
Abstract—HPC community projects that future extreme scale systems will be much less stable than curr...
Abstract — Checkpointing is a typical approach to tolerate failures in today’s supercomputing cluste...
Checkpoint and recovery protocols are commonly used in distributed applications for providing fault ...
Le passage de l'échelle des nouvelles plates-formes de calcul parallèle et distribué soulève de nomb...
Researchers have mentioned that the three most difficult and growing problems in the future of high-...
By leveraging the enormous amount of computational capabilities, scientists today are being able to ...
MapReduce is the efficient framework for parallel processing of distributed big data in cluster envi...
The MapReduce has become popular in big data environment due to its efficient parallel processing. H...
[[abstract]]The computing paradigm of MapReduce has gained extreme popularity in the area of large-s...
High-Performance Computing (HPC) has passed the Petascale mark and is moving forward to Exascale. As...
Due to the growing size of compute clusters, large scale parallel applications increasingly have to ...
International audienceThis work provides an analysis of checkpointing strategies for minimizing expe...
Since the last decade, computing systems turn to large scale parallel platforms composed of thousand...
International audienceOmission failures represent an important source of problems in data-intensive ...
This report provides an introduction to the design of scheduling algorithms to cope with faults on l...
Abstract—HPC community projects that future extreme scale systems will be much less stable than curr...
Abstract — Checkpointing is a typical approach to tolerate failures in today’s supercomputing cluste...
Checkpoint and recovery protocols are commonly used in distributed applications for providing fault ...
Le passage de l'échelle des nouvelles plates-formes de calcul parallèle et distribué soulève de nomb...
Researchers have mentioned that the three most difficult and growing problems in the future of high-...
By leveraging the enormous amount of computational capabilities, scientists today are being able to ...