International audienceOne has a large workload that is "divisible"--its constituent work's granularity can be adjusted arbitrarily--and one has access to p remote worker computers that can assist in computing the workload. How can one best utilize the workers? Complicating this question is the fact that each worker is subject to interruptions (of known likelihood) that kill all work in progress on it. One wishes to orchestrate sharing the workload with the workers in a way that maximizes the expected amount of work completed. Strategies are presented for achieving this goal, by balancing the desire to checkpoint often--thereby decreasing the amount of vulnerable work at any point--vs. the desire to avoid the context-switching required to ch...