高级检索
肖航, 许浩博, 王颖, 李佳骏, 王郁杰, 韩银和. 面向卷积神经网络的高能效比特稀疏加速器设计[J]. 计算机辅助设计与图形学学报, 2023, 35(7): 1122-1131. DOI: 10.3724/SP.J.1089.2023.19478
引用本文: 肖航, 许浩博, 王颖, 李佳骏, 王郁杰, 韩银和. 面向卷积神经网络的高能效比特稀疏加速器设计[J]. 计算机辅助设计与图形学学报, 2023, 35(7): 1122-1131. DOI: 10.3724/SP.J.1089.2023.19478
Xiao Hang, Xu Haobo, Wang Ying, Li Jiajun, Wang Yujie, Han Yinhe. Energy-Efficient Bit-Sparse Accelerator Design for Convolutional Neural Network[J]. Journal of Computer-Aided Design & Computer Graphics, 2023, 35(7): 1122-1131. DOI: 10.3724/SP.J.1089.2023.19478
Citation: Xiao Hang, Xu Haobo, Wang Ying, Li Jiajun, Wang Yujie, Han Yinhe. Energy-Efficient Bit-Sparse Accelerator Design for Convolutional Neural Network[J]. Journal of Computer-Aided Design & Computer Graphics, 2023, 35(7): 1122-1131. DOI: 10.3724/SP.J.1089.2023.19478

面向卷积神经网络的高能效比特稀疏加速器设计

Energy-Efficient Bit-Sparse Accelerator Design for Convolutional Neural Network

  • 摘要: 为解决当前比特稀疏架构的性能瓶颈,提出高能效比特稀疏加速器设计.首先提出一种激活值编码方法和相应的电路来提高卷积神经网络的比特稀疏度,结合比特串行电路实时跳过激活值的零值比特来加速神经网络的计算;然后提出一种列共享同步机制,以解决比特稀疏架构的同步问题,并在较小的面积和功耗开销下大幅提高比特稀疏架构的计算性能.在SMIC 40 nm工艺和1 GHz频率下,评估不同的比特稀疏架构在卷积神经网络上的能效.实验结果表明,与非稀疏加速器VAA和比特稀疏加速器LS-PRA相比,所提出的加速器AS-PRA分别提高了544%和179%的能效.

     

    Abstract: The high energy-efficient bit-sparse accelerator design is proposed to address the performance bottleneck of current bit-sparse architectures. Firstly, a coding method and corresponding circuit are proposed to enhance the bit-sparsity of convolutional neural networks, and employ the bit-serial circuit to eliminate computations of zero bits on the fly and accelerate neural networks. Secondly, a column shared scheme is proposed to address the synchronization issue of bit-sparse architectures for further acceleration with small area and power overhead. Finally, the energy efficiency of different bit-sparse architectures is evaluated with SMIC 40 nm technology at 1 GHz. The experimental results show that the energy efficiency of the proposed accelerator is 544% and 179% higher than dense accelerator (VAA) and bit-sparse accelerator (LS-PRA), respectively.

     

/

返回文章
返回