Managing Large-Scale Workflow Execution from Resource Provisioning to Provenance Tracking: The CyberShake Example

  • Authors:
  • Ewa Deelman;Scott Callaghan;Edward Field;Hunter Francoeur;Robert Graves;Nitin Gupta;Vipin Gupta;Thomas H. Jordan;Carl Kesselman;Philip Maechling;John Mehringer;Gaurang Mehta;David Okaya;Karan Vahi;Li Zhao

  • Affiliations:
  • USC Information Sciences Institute, USA;US Geological Survey, USA;University of Southern California, Los Angeles, USA;US Geological Survey, USA;URS Corporation, USA;US Geological Survey, USA;US Geological Survey, USA;US Geological Survey, USA;USC Information Sciences Institute, USA;US Geological Survey, USA;US Geological Survey, USA;USC Information Sciences Institute, USA;US Geological Survey, USA;USC Information Sciences Institute, USA;US Geological Survey, USA

  • Venue:
  • E-SCIENCE '06 Proceedings of the Second IEEE International Conference on e-Science and Grid Computing
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper discusses the process of building an environment where large-scale, complex, scientific analysis can be scheduled onto a heterogeneous collection of computational and storage resources. The example application is the Southern California Earthquake Center (SCEC) CyberShake project, an analysis designed to compute probabilistic seismic hazard curves for sites in the Los Angeles area. We explain which software tools were used to build to the system, describe their functionality and interactions. We show the results of running the CyberShake analysis that included over 250,000 jobs using resources available through SCEC and the TeraGrid.