FAST: FPGA Acceleration of Neural Networks Training

Alireza Borhani, Mohammad H. Goharinejad,Hamid R. Zarandi

2022 12th International Conference on Computer and Knowledge Engineering (ICCKE)(2022)

引用 0|浏览6
暂无评分
摘要
Training state-of-the-art ANNs is computationally and memory intensive. Thus, implementing the training on embedded devices with limited resources is challenging. In order to address this challenge, we propose FAST, a low-precision method to implement and optimize ANN training on FPGA. FAST first addresses the challenge of implementing the non-polynomial sigmoid activation function by presenting a solution using PNLA methods. Then, it introduces Hardware optimized PReLU (HOPE) activation function, which is specifically devised to reduce the required resources and increase the accuracy of computations on FPGA. We evaluated FAST against the software implementations of ANNs, using training tasks available in the MNIST benchmark. The results show that FAST improves the training speed by 8.6$\times$ and reduces the required memory size by orders of magnitude. It is worthwhile to mention that the method imposes almost no degradation in training accuracy.
更多
查看译文
关键词
Field Programmable Gate Array,Embedded Devices,Artificial Neural Network,Machine Learning,Approximation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要