MAPPARAT: A Resource Constrained FPGA-Based Accelerator for Sparse-Dense Matrix Multiplication

M. R. Ashuthosh, Santosh Krishna, Vishvas Sudarshan,Srinivasan Subramaniyan,Madhura Purnaprajna

2022 35th International Conference on VLSI Design and 2022 21st International Conference on Embedded Systems (VLSID)(2022)

引用 0|浏览3
暂无评分
摘要
Matrix Multiplication has gained importance due to its wide usage in Deep Neural Networks. The presence of sparsity in matrices needs special considerations to avoid redundancy in computations and memory accesses. Sparsity becomes relevant in the choice of compression format for storage and memory access. In addition to compression format, the choice of the algorithm also influences the performance of the matrix multiplier. The interplay of algorithm and compression formats results in significant variations in several performance parameters such as execution time, memory, and total energy consumed. This paper presents MAPPARAT, a custom FPGA-based hardware accelerator for $sparse \times dense$ matrix multiplication. Our analysis show that the choice of the compression format is heavily dependent on sparsity of the input matrices. We present two variants of MAPPARAT based on the algorithm used for $sparse \times dense$ matrix multiplication, viz., row-wise and column-wise product. A difference in speedup of $2.5\times$ and a difference in energy consumption by about $2.6\times$ is seen between the two variants. We show that an intelligent choice of algorithm and compression format based on the variations in sparsity, matrix dimensions, and device specifications is necessary for performance acceleration. For identical sparse matrices, a speedup of up to $3.6\times$ is observed, when the dense format is chosen for one of the matrices for sparsity in the range of 30% to 90%. MAPPARAT on a resource-constrained device shows performance efficiency of up to 7 GOPs-per-second-per-watt.
更多
查看译文
关键词
MAPPARAT,resource constrained FPGA-based accelerator,sparse-dense matrix multiplication,memory accesses,compression format,matrix multiplier,compression formats results,custom FPGA-based hardware accelerator,matrix dimensions,performance acceleration,identical sparse matrices,dense format,sparse dense matrix multiplication,deep neural networks,row-wise product,column-wise product,energy consumption,resource-constrained device
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要