An Infrastructure for Tackling Input-Sensitivity of GPU Program Optimizations

  • Authors:
  • Xipeng Shen;Yixun Liu;Eddy Z. Zhang;Poornima Bhamidipati

  • Affiliations:
  • Computer Science Department, College of William and Mary, Williamsburg, USA;Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, USA 20892-1182;Department of Computer Science, Rutgers, The State University of New Jersey, New Brunswick, USA 08901;Capital One, Williamsburg, USA 23185

  • Venue:
  • International Journal of Parallel Programming
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Graphic processing units (GPU) have become increasingly adopted for the enhancement of computing throughput. However, the development of a high-quality GPU application is challenging, due to the large optimization space and complex unpredictable effects of optimizations on GPU program performance. Many recent efforts have been employing empirical search-based auto-tuners to tackle the problem, but few of them have concentrated on the influence of program inputs on the optimizations. In this paper, based on a set of CUDA and OpenCL kernels, we report some evidences on the importance for auto-tuners to adapt to program input changes, and present a framework, G-ADAPT+, to address the influence by constructing cross-input predictive models for automatically predicting the (near-)optimal configurations for an arbitrary input to a GPU program. G-ADAPT+ is based on source-to-source compilers, specifically, Cetus and ROSE. It supports the optimizations of both CUDA and OpenCL programs.