Gathering data provenance at the operating system level is useful for capturing system-wide activity. However, many modern programs are complex and can perform numerous tasks concurrently. Capturing their prove-nance at this level, where processes are treated as sin-gle entities, may lead to the loss of useful intra-process detail. This can, in turn, produce false dependencies in the provenance graph. Using the LLVM compiler frame-work and SPADE provenance infrastructure, we investi-gate adding provenance instrumentation to allow intra-process provenance to be captured automatically. This results in a more accurate representation of the prove-nance relationships and eliminates some false dependen-cies. Since the capture of fine-grained prov...
Automatic provenance collection describes systems that observe processes and data transformations in...
Data processing pipelines that are designed to clean, transform and alter data in preparation for le...
Data processing pipelines that are designed to clean, transform and alter data in preparation for le...
Abstract. Often data processing is not implemented by a workflow sys-tem or an integration applicati...
Often data processing is not implemented by a workflow system or an integration application but is p...
International audienceOften data processing is not implemented by a workflow system or an integratio...
Often data processing is not implemented by a work ow system or an integration application but is pe...
Automatic provenance capture from arbitrary applications is a challenging problem. Different approac...
Scientists can facilitate data intensive applications to study and understand the behavior of a comp...
Data provenance is the history of a digital artifact, from the point of collection to its present<br...
Integrated provenance support promises to be a chief advantage of scientific workflow systems over s...
Workflow provenance typically assumes that each module is a “black-box”, so that each output depends...
Workflow provenance typically assumes that each module is a “black-box”, so that each output depends...
Scientists require provenance information either to validate their model or to investigate the origi...
Data processing pipelines that are designed to clean, transform and alter data in preparation for le...
Automatic provenance collection describes systems that observe processes and data transformations in...
Data processing pipelines that are designed to clean, transform and alter data in preparation for le...
Data processing pipelines that are designed to clean, transform and alter data in preparation for le...
Abstract. Often data processing is not implemented by a workflow sys-tem or an integration applicati...
Often data processing is not implemented by a workflow system or an integration application but is p...
International audienceOften data processing is not implemented by a workflow system or an integratio...
Often data processing is not implemented by a work ow system or an integration application but is pe...
Automatic provenance capture from arbitrary applications is a challenging problem. Different approac...
Scientists can facilitate data intensive applications to study and understand the behavior of a comp...
Data provenance is the history of a digital artifact, from the point of collection to its present<br...
Integrated provenance support promises to be a chief advantage of scientific workflow systems over s...
Workflow provenance typically assumes that each module is a “black-box”, so that each output depends...
Workflow provenance typically assumes that each module is a “black-box”, so that each output depends...
Scientists require provenance information either to validate their model or to investigate the origi...
Data processing pipelines that are designed to clean, transform and alter data in preparation for le...
Automatic provenance collection describes systems that observe processes and data transformations in...
Data processing pipelines that are designed to clean, transform and alter data in preparation for le...
Data processing pipelines that are designed to clean, transform and alter data in preparation for le...