首页 | 本学科首页   官方微博 | 高级检索  
     

深度学习加速技术研究
引用本文:杨旭瑜,张铮,张为华.深度学习加速技术研究[J].计算机系统应用,2016,25(9):1-9.
作者姓名:杨旭瑜  张铮  张为华
作者单位:复旦大学 软件学院, 上海 201203;复旦大学 上海市数据科学重点实验室, 上海 201203;复旦大学 并行处理研究所, 上海 201203,解放军信息工程大学 数学工程与先进计算国家重点实验室, 郑州 450001,复旦大学 软件学院, 上海 201203;复旦大学 上海市数据科学重点实验室, 上海 201203;复旦大学 并行处理研究所, 上海 201203
基金项目:国家高技术研究发展计划(863)(2012AA010905);国家自然科学基金(61370081)
摘    要:深度学习是近年来机器学习的研究热点,并已广泛应用于不同领域. 但由于训练模型复杂和训练集规模庞大等原因导致的深度学习性能问题已成为其发展的一大阻碍. 近年来计算机硬件的快速发展,尤其是处理器核数的不断增加和整体运算能力的快速提高,给深度学习加速提供了硬件基础,然而其训练算法并行度低和内存开销巨大等问题使得加速研究工作困难重重. 首先介绍了深度学习的背景和训练算法,对当前主要的深度学习加速研究工作进行归纳总结. 在此基础上,对经典的深度学习模型进行性能测试,分析了深度学习及并行算法的性能问题. 最后,对深度学习的未来发展进行了展望.

关 键 词:深度学习  深度学习加速  深度学习性能  分布式学习  硬件加速
收稿时间:2015/12/28 0:00:00
修稿时间:2016/1/27 0:00:00

Research on Deep Learning Acceleration Technique
YANG Xu-Yu,ZHANG Zheng and ZHANG Wei-Hua.Research on Deep Learning Acceleration Technique[J].Computer Systems& Applications,2016,25(9):1-9.
Authors:YANG Xu-Yu  ZHANG Zheng and ZHANG Wei-Hua
Affiliation:Software School, Fudan University, Shanghai 201203, China;Shanghai Key Laboratory of Data Science, Fudan University, Shanghai 201203, China;Parallel Processing Institute, Fudan University, Shanghai 201203, China,State Key Laboratory of Mathematical Engineering and Advanced Computing, PLA Information Engineering University, Zhengzhou 450001, China and Software School, Fudan University, Shanghai 201203, China;Shanghai Key Laboratory of Data Science, Fudan University, Shanghai 201203, China;Parallel Processing Institute, Fudan University, Shanghai 201203, China
Abstract:Deep learning has become a popular research direction recently and has been applied to many fields. But the performance issue caused by the complicated learning model and massive training data has becomes an obstacle of deep learning evolution. As the development of processor technology, the core number and performance of the processor have increased rapidly. However, the acceleration research is restricted mainly due to the low parallelism and high memory consumption of the training algorithm. The paper introduces the background of deep learning and its training algorithm, then summarizes current deep learning acceleration works. Further, it analyzes classic deep models training algorithm to explain the causes of deep learning performance problem. Based on the analysis, it lists the challenges about deep learning evolution and makes proposal to address them.
Keywords:deep learning  deep learning acceleration  deep learning performance  distribution deep learning computing  hardware acceleration
点击此处可从《计算机系统应用》浏览原始摘要信息
点击此处可从《计算机系统应用》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号