首页 | 本学科首页   官方微博 | 高级检索  
     

基于关系型蒸馏的分步神经网络压缩方法
引用本文:刘昊,张晓滨.基于关系型蒸馏的分步神经网络压缩方法[J].计算机系统应用,2021,30(12):248-254.
作者姓名:刘昊  张晓滨
作者单位:西安工程大学 计算机学院, 西安 710048
基金项目:陕西省自然科学基金(2019JQ-849)
摘    要:针对关系型知识蒸馏方法中教师网络与学生网络的层数差距过大导致蒸馏效果下降的问题,提出一种基于关系型蒸馏的分步神经网络压缩方法.该方法的要点在于,在教师网络和学生网络之间增加一个中间网络分步进行关系型蒸馏,同时在每一次蒸馏过程中都增加额外的单体信息来进一步优化和增强学生模型的学习能力,实现神经网络压缩.实验结果表明,本文的方法在CIFAR-10和CIFAR-100图像分类数据集上的分类准确度相较于原始的关系型知识蒸馏方法均有0.2%左右的提升.

关 键 词:模型压缩  知识蒸馏  关系型知识蒸馏  神经网络  神经网络压缩
收稿时间:2021/2/24 0:00:00
修稿时间:2021/3/15 0:00:00

Compression Method for Stepwise Neural Network Based on Relational Distillation
LIU Hao,ZHANG Xiao-Bin.Compression Method for Stepwise Neural Network Based on Relational Distillation[J].Computer Systems& Applications,2021,30(12):248-254.
Authors:LIU Hao  ZHANG Xiao-Bin
Abstract:This study aims at the problem that the distillation effect decreases when the gap between the teacher network and the student network in relational knowledge distillation is too large. A stepwise neural network compression method based on relational distillation is proposed. The key point of this method is to add an intermediate network between the teacher network and the student network for relational distillation step by step. Moreover, in each distillation process, additional monomer information is added to further optimize and enhance the learning ability of the student model. The experimental results show that the classification accuracy of the proposed method on CIFAR-10 and CIFAR-100 image classification datasets is improved by about 0.2% compared with that of the original relational knowledge distillation method.
Keywords:model compression  knowledge distillation  relational knowledge distillation  neural network  neural network compression
本文献已被 万方数据 等数据库收录!
点击此处可从《计算机系统应用》浏览原始摘要信息
点击此处可从《计算机系统应用》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号