首页 | 本学科首页   官方微博 | 高级检索  
     

深度神经网络压缩与加速综述
引用本文:曾焕强,胡浩麟,林向伟,侯军辉,蔡灿辉. 深度神经网络压缩与加速综述[J]. 信号处理, 2022, 38(1): 183-194. DOI: 10.16798/j.issn.1003-0530.2022.01.021
作者姓名:曾焕强  胡浩麟  林向伟  侯军辉  蔡灿辉
作者单位:1.华侨大学工学院, 福建 泉州 362021
基金项目:国家自然科学基金(61871434,61802136);福建省自然科学基金杰出青年项目(2019J06017);厦门市科技重大项目(3502ZCQ20191005);厦门市科技局产学研协同创新项目(3502Z20203033);福建省教改项目(FBJG20180038)。
摘    要:近年来,随着图形处理器性能的飞速提升,深度神经网络取得了巨大的发展成就,在许多人工智能任务中屡创佳绩.然而,主流的深度学习网络模型由于存在计算复杂度高、内存占用较大、耗时长等缺陷,难以部署在计算资源受限的移动设备或时延要求严格的应用中.因此,在不显著影响模型精度的前提下,通过对深度神经网络进行压缩和加速来轻量化模型逐渐...

关 键 词:深度神经网络压缩与加速  深度学习  模型剪枝  知识蒸馏  参数量化
收稿时间:2021-03-03

Deep Neural Network Compression and Acceleration:An Overview
ZENG Huanqiang,HU Haolin,LIN Xiangwei,HOU Junhui,CAI Canhui. Deep Neural Network Compression and Acceleration:An Overview[J]. Signal Processing(China), 2022, 38(1): 183-194. DOI: 10.16798/j.issn.1003-0530.2022.01.021
Authors:ZENG Huanqiang  HU Haolin  LIN Xiangwei  HOU Junhui  CAI Canhui
Affiliation:1.School of Engineering,Huaqiao University, Quanzhou, Fujian 362021, China2.School of Information Science and Engineering,Huaqiao University, Xiamen, Fujian 361021, China3.Department of Computer Science,City University of Hong Kong, Hong Kong 999077, China
Abstract:In recent years, with the rapid improvement of graphic processor unit(GPU) performance, deep neural network (DNN) has made great achievements in many artificial intelligence tasks. However, the mainstream deep learning network model has some defects, such as high computational complexity, large memory consumption and long time-consuming, which makes it difficult to be deployed in mobile devices with limited computing resources or applications with strict delay requirements. Therefore, on the premise of maintaining the accuracy of the model, it gradually attracts a lot of attention from both academia and industry to reduce the weight of the model by compressing and accelerating the DNN. This paper reviews the compression and acceleration techniques of DNNs in recent years. These technologies can be divided into four categories: quantization, model pruning, lightweight convolution kernel design and knowledge distillation. For each technology category, this paper firstly analyzes the development status and existing defects. Then, this paper summarizes the performance evaluation methods of model compression and acceleration. Finally, the challenges in the field of model compression and acceleration, and the possible future research directions are discussed. 
Keywords:DNN compression and acceleration  deep learning  model pruning  knowledge distillation  parameter quantization
本文献已被 维普 等数据库收录!
点击此处可从《信号处理》浏览原始摘要信息
点击此处可从《信号处理》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号