Transformations for Energy Efficient Accelerated Chain Matrix Multiplication (TEE-ACM2)
DescriptionGPU matrix chain multiplication serves as a basis for a wide range of scientific domains like computer graphics, physics, and machine learning. While its time performance was studied for years, there has been significantly less effort in optimizing its energy efficiency. GPU power consumption is heavily impacted by the number of data transfers performed. In fact, a data transfer from global memory needs a thousand times more energy than a double precision arithmetic operation. Thus, minimizing data transfers is key for reducing the energy consumption. We present an energy efficient solution for Matrix Chain Multiplication on GPUs that minimizes computation as well as off-chip data transfers. For this, optimizations at three different levels are provided. For a single matrix multiplication, we use a large tile blocking strategy. Then, we extend our approach to three matrices. Finally, we propose a solution for a sequence of matrices.
Event Type
Posters
Research Posters
TimeTuesday, 15 November 20228:30am - 5pm CST
LocationC1-2-3
Registration Categories
TP
XO/EX
Poster view
Back To Top Button