Authors: Mohammad Zaeed and Tanzima Z. Islam (Texas State University); Younghyun Cho (University of California, Berkeley); and Xiaoye Sherry Li, Hengrui Luo, and Yang Liu (Lawrence Berkeley National Laboratory (LBNL))
Abstract: Autotuning is a widely used method for guiding developers of large-scale applications to achieve high performance. However, autotuners typically employ black-box optimizations to recommend parameter settings at the cost of users missing the opportunity to identify performance bottlenecks. Performance analysis fills that gap and identifies problems and optimization opportunities that can result in better runtime and utilization of hardware resources. This work combines the best of the both worlds by integrating a systematic performance analysis and visualization approach into a publicly available autotuning framework, GPTune, to suggest users which configuration parameters are important to tune, to what value, and how tuning the parameters affect hardware-application interactions. Our experiments demonstrate that a subset of the task parameters impact the execution time of the Hypre application; the memory traffic and page faults cause performance problems in the Plasma-DGEMM routine on Cori-Haswell.
Best Poster Finalist (BP): no
Poster summary: PDF
Back to Poster Archive Listing