Over the last two decades, the field of computational science has seen a dramatic shift towards incorporating high-throughput computation and big-data analysis as fundamental pillars of the scientific discovery process. This has necessitated the development of tools and techniques to deal with the generation, storage and processing of large amounts of data. In this work we present an in-depth look at the workflow engine powering AiiDA, a widely adopted, highly flexible and database-backed informatics infrastructure with an emphasis on data reproducibility. We detail many of the design choices that were made which were informed by several important goals: the ability to scale from running on individual laptops up to high-performance supercom...
The evolution of High-Performance Computing (HPC) platforms enables the design and execution of prog...
Workflows have recently emerged as a paradigm for representing and managing complex distributed scie...
Increasingly, online computer applications rely on large-scale data analyses to offer personalised a...
The ever-growing availability of computing power and the sustained development of advanced computati...
Modern advances in computational technology have facilitated great strides in a wide variety of scie...
This paper introduces FireWorks, a workflow software for running high-throughput calculation workflo...
Background Reproducibility is one of the tenets of the scientific method. Scientific experiments ...
Workflows have been used traditionally as a mean to describe and implement the computing usually par...
Automation of the execution of computational tasks is at the heart of improving scientific productiv...
Progress in science is deeply bound to the effective use of high-performance computing infrastructur...
Dataflow-style workflows offer a simple, high-level programming model for flexible prototyping of sc...
Scientific computing has entered a new era of scale and sharing with the arrival of cyberinfrastruct...
The role of data in modern scientific workflows becomes more and more crucial. The unprecedented amo...
International audienceWorkflows may be defined as abstractions used to model the coherent flow of ac...
Big data and complex analysis workflows (pipelines) are common issues in data driven science such as...
The evolution of High-Performance Computing (HPC) platforms enables the design and execution of prog...
Workflows have recently emerged as a paradigm for representing and managing complex distributed scie...
Increasingly, online computer applications rely on large-scale data analyses to offer personalised a...
The ever-growing availability of computing power and the sustained development of advanced computati...
Modern advances in computational technology have facilitated great strides in a wide variety of scie...
This paper introduces FireWorks, a workflow software for running high-throughput calculation workflo...
Background Reproducibility is one of the tenets of the scientific method. Scientific experiments ...
Workflows have been used traditionally as a mean to describe and implement the computing usually par...
Automation of the execution of computational tasks is at the heart of improving scientific productiv...
Progress in science is deeply bound to the effective use of high-performance computing infrastructur...
Dataflow-style workflows offer a simple, high-level programming model for flexible prototyping of sc...
Scientific computing has entered a new era of scale and sharing with the arrival of cyberinfrastruct...
The role of data in modern scientific workflows becomes more and more crucial. The unprecedented amo...
International audienceWorkflows may be defined as abstractions used to model the coherent flow of ac...
Big data and complex analysis workflows (pipelines) are common issues in data driven science such as...
The evolution of High-Performance Computing (HPC) platforms enables the design and execution of prog...
Workflows have recently emerged as a paradigm for representing and managing complex distributed scie...
Increasingly, online computer applications rely on large-scale data analyses to offer personalised a...