Accelerating Fully Spectral CNNs with Adaptive Activation Functions on FPGA.

DATE(2021)

引用 1|浏览12
暂无评分
摘要
Computing convolutional layers in frequency domain can largely reduce the computation overhead for training and inference of convolutional neural networks (CNNs). However, existing designs with such an idea require repeated spatial- and frequency-domain transforms due to the absence of nonlinear functions in the frequency domain, as such it makes the benefit less attractive for low-latency inference. This paper presents a fully spectral CNN approach by proposing a novel adaptive Rectified Linear Unit (ReLU) activation in spectral domain. The proposed design maintains the non-linearity in the network while taking into account the hardware efficiency in algorithm level. The spectral model size is further optimized by merging and fusing layers. Then, a customized hardware architecture is proposed to implement the designed spectral network on FPGA device with DSP optimizations for 8-bit fixed point multipliers. Our hardware accelerator is implemented on Intel's Arria 10 device and applied to the MNIST, SVHN, AT&T and CIFAR-10 datasets. Experimental results show a speed improvement of 6x similar to 10x and 4x similar to 5.7x compared to state-of-the-art spatial or FFT-based designs respectively, while achieving similar accuracy across the benchmark datasets.
更多
查看译文
关键词
CIFAR-10 datasets,adaptive rectified linear unit activation,FPGA device,hardware architecture,spectral model size,hardware efficiency,fully spectral CNN approach,low-latency inference,nonlinear functions,frequency-domain transforms,convolutional neural networks,frequency domain,adaptive activation functions,FFT-based designs,hardware accelerator,8-bit fixed point multipliers
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要