Failure analysis of distributed scientific workflows executing in the cloud

  • Authors:
  • Taghrid Samak;Dan Gunter;Monte Goode;Ewa Deelman;Gideon Juve;Fabio Silva;Karan Vahi

  • Affiliations:
  • Lawrence Berkeley National Laboratory, Berkeley, CA;Lawrence Berkeley National Laboratory, Berkeley, CA;Lawrence Berkeley National Laboratory, Berkeley, CA;University of Southern California, Marina Del Rey, CA;University of Southern California, Marina Del Rey, CA;University of Southern California, Marina Del Rey, CA;University of Southern California, Marina Del Rey, CA

  • Venue:
  • Proceedings of the 8th International Conference on Network and Service Management
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

This work presents models characterizing failures observed during the execution of large scientific applications on Amazon EC2. Scientific workflows are used as the underlying abstraction for application representations. As scientific workflows scale to hundreds of thousands of distinct tasks, failures due to software and hardware faults become increasingly common. We study job failure models for data collected from 4 scientific applications, by our Stampede framework. In particular, we show that a Naive Bayes classifier can accurately predict the failure probability of jobs. The models allow us to predict job failures for a given execution resource and then use these failure predictions for two higher-level goals: (1) to suggest a better job assignment, and (2) to provide quantitative feedback to the workflow component developer about the robustness of their application codes.