TY - GEN
T1 - GPGPU performance estimation for frequency scaling using cross-benchmarking
AU - Wang, Qiang
AU - Liu, Chengjian
AU - Chu, Xiaowen
N1 - Funding Information:
This research was supported by Hong Kong RGC GRF grant HKBU 12200418. We thank the anonymous reviewers for their constructive comments and suggestions. We would also like to thank NVIDIA AI Technology Centre (NVAITC) for providing the GPUs for some experiments.
PY - 2020/2/23
Y1 - 2020/2/23
N2 - Dynamic Voltage and Frequency Scaling (DVFS) on General-Purpose Graphics Processing Units (GPGPUs) is now becoming one of the most significant techniques to balance computational performance and energy consumption. However, there are still few fast and accurate models for predicting GPU kernel execution time under different core and memory frequency settings, which is important to determine the best frequency configuration for energy saving. Accordingly, a novel GPGPU performance estimation model with both core and memory frequency scaling is herein proposed. We design a cross-benchmarking suite, which simulates kernels with a wide range of instruction distributions. The synthetic kernels generated by this suite can be used for model pre-training or as supplementary training samples. Then we apply two different machine learning algorithms, Support Vector Regression (SVR) and Gradient Boosting Decision Tree (GBDT), to study the correlation between kernel performance counters and kernel performance. The models trained only with our cross-benchmarking suite achieve satisfying accuracy (16%∼22% mean absolute error) on 24 unseen real application kernels. Validated on three modern GPUs with a wide frequency scaling range, by using a collection of 24 real application kernels, the proposed model is able to achieve accurate results (5.1%, 2.8%, 6.5% mean absolute error) for the target GPUs (GTX 980, Titan X Pascal and Tesla P100).
AB - Dynamic Voltage and Frequency Scaling (DVFS) on General-Purpose Graphics Processing Units (GPGPUs) is now becoming one of the most significant techniques to balance computational performance and energy consumption. However, there are still few fast and accurate models for predicting GPU kernel execution time under different core and memory frequency settings, which is important to determine the best frequency configuration for energy saving. Accordingly, a novel GPGPU performance estimation model with both core and memory frequency scaling is herein proposed. We design a cross-benchmarking suite, which simulates kernels with a wide range of instruction distributions. The synthetic kernels generated by this suite can be used for model pre-training or as supplementary training samples. Then we apply two different machine learning algorithms, Support Vector Regression (SVR) and Gradient Boosting Decision Tree (GBDT), to study the correlation between kernel performance counters and kernel performance. The models trained only with our cross-benchmarking suite achieve satisfying accuracy (16%∼22% mean absolute error) on 24 unseen real application kernels. Validated on three modern GPUs with a wide frequency scaling range, by using a collection of 24 real application kernels, the proposed model is able to achieve accurate results (5.1%, 2.8%, 6.5% mean absolute error) for the target GPUs (GTX 980, Titan X Pascal and Tesla P100).
KW - Dynamic Voltage and Frequency Scaling
KW - GPU Performance Modeling
KW - Graphics Processing Units
KW - Machine Learning
UR - http://www.scopus.com/inward/record.url?scp=85082084732&partnerID=8YFLogxK
U2 - 10.1145/3366428.3380767
DO - 10.1145/3366428.3380767
M3 - Conference proceeding
AN - SCOPUS:85082084732
T3 - GPGPU 2020 - Proceedings of the 2020 General Purpose Processing Using GPU
SP - 31
EP - 40
BT - GPGPU 2020 - Proceedings of the 2020 General Purpose Processing Using GPU
PB - Association for Computing Machinery (ACM)
T2 - 13th Annual Workshop on General Purpose Processing using Graphics Processing Unit, GPGPU 2020
Y2 - 23 February 2020 through 23 February 2020
ER -