Failure prediction for HPC systems and applications: Current situation and open issues

  • Authors:
  • Ana Gainaru;Franck Cappello;Marc Snir;William Kramer

  • Affiliations:
  • National Centre for Supercomputing Applications, Urbana, IL, USA, University of Illinois at Urbana-Champaign, Urbana, IL, USA;University of Illinois at Urbana-Champaign, Urbana, IL, USA, INRIA, Rocquencourt, Le Chesnay Cedex, France;University of Illinois at Urbana-Champaign, Urbana, IL, USA, Argonne National Laboratory, Argonne, IL, USA;National Centre for Supercomputing Applications, Urbana, IL, USA

  • Venue:
  • International Journal of High Performance Computing Applications
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

As large-scale systems evolve towards post-petascale computing, it is crucial to focus on providing fault-tolerance strategies that aim to minimize fault's effects on applications. By far the most popular technique is the checkpoint-restart strategy. A complement to this classical approach is failure avoidance, by which the occurrence of a fault is predicted and proactive measures are taken. This requires a reliable prediction system to anticipate failures and their locations. One way of offering prediction is by the analysis of system logs generated during production by large-scale systems. Current research in this field presents a number of limitations that make them unusable for running on real production high-performance computing (HPC) systems. Based on our observations that different failures have different distributions and behaviours, we propose a novel hybrid approach that combines signal analysis with data mining in order to overcome current limitations. We show that by analysing each event according to its specific behaviour, our prediction provides a precision of over 90% and its able to discover about 50% of all failures in a system, result which allows its integration in proactive fault tolerance protocols.