首页 | 本学科首页   官方微博 | 高级检索  
     

基于FPGA的递归神经网络加速器的研究进展
作者姓名:高琛  张帆
作者单位:国家数字交换系统工程技术研究中心,河南 郑州 450002
基金项目:国家自然科学基金资助项目(61572520);国家自然科学基金创新研究群体资助项目(61521003)
摘    要:递归神经网络(RNN)近些年来被越来越多地应用在机器学习领域,尤其是在处理序列学习任务中,相比CNN等神经网络性能更为优异。但是RNN及其变体,如LSTM、GRU等全连接网络的计算及存储复杂性较高,导致其推理计算慢,很难被应用在产品中。一方面,传统的计算平台CPU不适合处理RNN的大规模矩阵运算;另一方面,硬件加速平台GPU的共享内存和全局内存使基于GPU的RNN加速器的功耗比较高。FPGA 由于其并行计算及低功耗的特性,近些年来被越来越多地用来做 RNN 加速器的硬件平台。对近些年基于FPGA的RNN加速器进行了研究,将其中用到的数据优化算法及硬件架构设计技术进行了总结介绍,并进一步提出了未来研究的方向。

关 键 词:递归神经网络  FGPA  加速器  

Survey of FPGA based recurrent neural network accelerator
Authors:Chen GAO  Fan ZHANG
Affiliation:National Digital Switching System Engineering and Technological Research Center,Zhengzhou 450002,China
Abstract:Recurrent neural network(RNN) has been used wildly used in machine learning field in recent years,especially in dealing with sequential learning tasks compared with other neural network like CNN.However,RNN and its variants,such as LSTM,GRU and other fully connected networks,have high computational and storage complexity,which makes its inference calculation slow and difficult to be applied in products.On the one hand,traditional computing platforms such as CPU are not suitable for large-scale matrix operation of RNN.On the other hand,the shared memory and global memory of hardware acceleration platform GPU make the power consumption of GPU-based RNN accelerator higher.More and more research has been done on the RNN accelerator of the FPGA in recent years because of its parallel computing and low power consumption performance.An overview of the researches on RNN accelerator based on FPGA in recent years is given.The optimization algorithm of software level and the architecture design of hardware level used in these accelerator are summarized and some future research directions are proposed.
Keywords:recurrent neural network  FPGA  accelerator  
点击此处可从《》浏览原始摘要信息
点击此处可从《》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号