首页 | 本学科首页   官方微博 | 高级检索  
     

基于指数移动平均知识蒸馏的神经网络低比特量化方法
引用本文:吕君环,许柯,王东.基于指数移动平均知识蒸馏的神经网络低比特量化方法[J].模式识别与人工智能,2021,34(12):1143-1151.
作者姓名:吕君环  许柯  王东
作者单位:北京交通大学 信息科学研究所 北京100044;北京交通大学 现代信息科学与网络技术北京市重点实验室 北京100044
基金项目:国家重点研发计划项目(No.2019YFB2204200)、中央高校基本科研业务费项目(No.2020JBM020)、北京市自然科学基金项目(No.4202063)资助
摘    要:目前存储和计算成本严重阻碍深度神经网络应用和推广,而神经网络量化是一种有效的压缩方法.神经网络低比特量化存在的显著困难是量化比特数越低,网络分类精度也越低.为了解决这一问题,文中提出基于指数移动平均知识蒸馏的神经网络低比特量化方法.首先利用少量图像进行自适应初始化,训练激活和权重的量化步长,加快量化网络收敛.再引入指数移动平均(EMA)知识蒸馏的思想,利用EMA对蒸馏损失和任务损失进行归一化,指导量化网络训练.在ImageNet、CIFAR-10数据集上的分类任务表明,文中方法可获得接近或超过全精度网络的性能.

关 键 词:深度学习  网络量化  知识蒸馏  模型压缩
收稿时间:2021-04-20

Low-Bit Quantization of Neural Network Based on Exponential Moving Average Knowledge Distillation
LÜ,Junhuan,XU Ke,WANG Dong.Low-Bit Quantization of Neural Network Based on Exponential Moving Average Knowledge Distillation[J].Pattern Recognition and Artificial Intelligence,2021,34(12):1143-1151.
Authors:  Junhuan  XU Ke  WANG Dong
Affiliation:1. Institute of Information Science, Beijing Jiaotong University, Beijing 100044;
2. Beijing Key Laboratory of Advanced Information Science and Network Technology, Beijing Jiaotong University, Beijing 100044
Abstract:Now the memory and computational cost restrict the popularization of deep neural network application, whereas neural network quantization is an effective compression method. As the number of quantized bits is lower, the classification accuracy of neural networks becomes poorer in low-bit quantization of neural networks. To solve this problem, a low-bit quantization method of neural networks based on knowledge distillation is proposed. Firstly, a few images are exploited for adaptive initialization to train the quantization step of activation and weight to speed up the convergence of the quantization network. Then, the idea of exponential moving average knowledge distillation is introduced to normalize distillation loss and task loss and guide the training of quantization network. Experiments on ImageNet and CIFAR-10 datasets show that the performance of the proposed method is close to or better than that of the full precision network.
Keywords:Deep Learning  Network Quantization  Knowledge Distillation  Model Compression  
本文献已被 万方数据 等数据库收录!
点击此处可从《模式识别与人工智能》浏览原始摘要信息
点击此处可从《模式识别与人工智能》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号