March 14, 2016
Conference Paper

Critical Points Based Register-Concurrency Autotuning for GPUs

Abstract

The unprecedented prevalence of GPGPU is largely attributed to its abundant on-chip register resources, which allow massively concurrent threads and extremely fast context switch. However, due to internal memory capacity constraints, there is a tradeoff between the per-thread register usage and the overall concurrency. This becomes a design problem in terms of performance tuning, since the performance “sweet spot” which can be significantly affected by these two factors is generally unknown beforehand. In this paper, we propose an effective autotuning solution to quickly and efficiently select the optimal number of registers perthread for delivering the best GPU performance. Experiments on three generations of GPUs (Nvidia Fermi, Kepler and Maxwell) demonstrate that our simple strategy can achieve an average of 10% performance improvement while a max of 50% over the original version without modifying the user program. Additionally, to reduce local cache misses due to register spilling and further improve performance, we explore three optimization schemes (i.e. bypass L1 for global memory access, enlarge local L1 cache and spill into shared memory) and discuss their impact on performance on a Kepler GPU.

Revised: May 13, 2016 | Published: March 14, 2016

Citation

Li A., S. Song, A. Kumar, E. Zhang, D. Chavarría-Miranda, and H. Corporaal. 2016. Critical Points Based Register-Concurrency Autotuning for GPUs. In Proceedings of the Design, Automation and Test in Europe Conference (DATE 2016), March 14-18, 2016, Dresden, Germany, 1273-1278. Piscataway, New Jersey:IEEE. PNNL-SA-114732.