王巍, 周凯利, 王伊昌, 王广, 杨正琳, 袁军. 卷积神经网络(CNN)算法的FPGA并行结构设计[J]. 微电子学与计算机, 2019, 36(4): 57-62, 66.
引用本文: 王巍, 周凯利, 王伊昌, 王广, 杨正琳, 袁军. 卷积神经网络(CNN)算法的FPGA并行结构设计[J]. 微电子学与计算机, 2019, 36(4): 57-62, 66.
WANG Wei, ZHOU Kai-li, WANG Yi-chang, WANG Guang, YANG Zheng-lin, YUAN Jun. FPGA Parallel Structure Design of Convolutional Neural Network (CNN) Algorithm[J]. Microelectronics & Computer, 2019, 36(4): 57-62, 66.
Citation: WANG Wei, ZHOU Kai-li, WANG Yi-chang, WANG Guang, YANG Zheng-lin, YUAN Jun. FPGA Parallel Structure Design of Convolutional Neural Network (CNN) Algorithm[J]. Microelectronics & Computer, 2019, 36(4): 57-62, 66.

卷积神经网络(CNN)算法的FPGA并行结构设计

FPGA Parallel Structure Design of Convolutional Neural Network (CNN) Algorithm

  • 摘要: 本文进行了CNN算法的FPGA并行结构设计.该设计首先利用CNN的并行计算特征以及循环变换方法, 实现了可高效进行并行流水线的卷积计算电路, 然后利用能够减少存储器访存时间的双缓存技术, 在输入输出部分实现了缓存阵列, 用于提高电路的计算性能(GOPS, 每秒十亿次运算数).同时本文还对激活函数进行了优化设计, 利用查找表和多项式结合的分段拟合方法设计了激活函数(sigmoid)的硬件电路, 以保证近似的激活函数的硬件电路不会使精度下降.实验结果表明:输入时钟为150 MHz时, 整体电路在计算性能上由15.87 GOPS提高到了20.62 GOPS, 并在MNIST数据集上的识别率达到了98.81%.

     

    Abstract: In this paper, the FPGA parallel structure design of CNN algorithm is carried out. The design first uses the parallel computing features of CNN and the cyclic transformation method to realize a convolution calculation circuit that can efficiently perform parallel pipelines. Then, using the double-buffer technology that can reduce the memory access time, a cache array is implemented in the input and output sections to improve the computational performance of the circuit (GOPS, one billion operations per second). At the same time, the activation function is optimized. The hardware circuit of the activation function (sigmoid) is designed by using the segmentation fitting method of lookup table and polynomial to ensure that the hardware circuit of the approximate activation function will not reduce the accuracy. The experimental results show that when the input clock is 150MHz, the overall performance of the circuit is improved from 15.87 GOPS to 20.62 GOPS, and the recognition rate on the MNIST data set reaches 98.81%.

     

/

返回文章
返回