Task characteristics estimations such as runtime, disk space, and memory consumption, are commonly used by scheduling algorithms and resource provisioning techniques to provide successful and efficient workflow executions. These methods assume that accurate estimations are available, but in pro-duction systems it is hard to compute such estimates with good accuracy. In this work, we first profile three real sci-entific workflows collecting fine-grained information such as process I/O, runtime, memory usage, and CPU utilization. We then propose a method to automatically characterize workflow task needs based on these profiles. Our method es-timates task runtime, disk space, and memory consumption based on the size of tasks input data. It loo...
In this paper we propose a novel method for auto-scaling data-centric workflow tasks. Scaling is ach...
Automation of the execution of computational tasks is at the heart of improving scientific productiv...
2014-09-15Scientific workflows are a means of defining and orchestrating large, complex, multi-stage...
Estimates of task runtime, disk space usage, and memory consumption, are commonly used by scheduling...
With the increasing amount of data available to scientists in disciplines as diverse as bioinformat...
Abstract: Researchers working on the planning, scheduling, and execution of scientific workflows nee...
Many fields of modern science require huge amounts of computation, and workflows are a very popular ...
We present a software tool, the Container Profiler, that measures and records the resource usage of ...
Abstract—Scientific workflows, which capture large compu-tational problems, may be executed on large...
Workload consolidation is a common method to increase resource utilization of the clusters or data c...
Many scientific workflow scheduling algorithms need to be informed about task runtimes a-priori to c...
Scientific workflows are often used to automate large-scale data analysis pipelines on clusters, gri...
Abstract: Evaluation methods for parallel computers often require the availability of relevant workl...
Domain experts are often untrained in big data technologies and this limits their ability to exploit...
High throughput computing (HTC) has aided the scientific community in the analysis of vast amounts o...
In this paper we propose a novel method for auto-scaling data-centric workflow tasks. Scaling is ach...
Automation of the execution of computational tasks is at the heart of improving scientific productiv...
2014-09-15Scientific workflows are a means of defining and orchestrating large, complex, multi-stage...
Estimates of task runtime, disk space usage, and memory consumption, are commonly used by scheduling...
With the increasing amount of data available to scientists in disciplines as diverse as bioinformat...
Abstract: Researchers working on the planning, scheduling, and execution of scientific workflows nee...
Many fields of modern science require huge amounts of computation, and workflows are a very popular ...
We present a software tool, the Container Profiler, that measures and records the resource usage of ...
Abstract—Scientific workflows, which capture large compu-tational problems, may be executed on large...
Workload consolidation is a common method to increase resource utilization of the clusters or data c...
Many scientific workflow scheduling algorithms need to be informed about task runtimes a-priori to c...
Scientific workflows are often used to automate large-scale data analysis pipelines on clusters, gri...
Abstract: Evaluation methods for parallel computers often require the availability of relevant workl...
Domain experts are often untrained in big data technologies and this limits their ability to exploit...
High throughput computing (HTC) has aided the scientific community in the analysis of vast amounts o...
In this paper we propose a novel method for auto-scaling data-centric workflow tasks. Scaling is ach...
Automation of the execution of computational tasks is at the heart of improving scientific productiv...
2014-09-15Scientific workflows are a means of defining and orchestrating large, complex, multi-stage...