首页 | 本学科首页   官方微博 | 高级检索  
     

图卷积网络与自注意机制在文本分类任务上的对比分析
引用本文:蒋浩泉,张儒清,郭嘉丰,范意兴,程学旗.图卷积网络与自注意机制在文本分类任务上的对比分析[J].中文信息学报,2021,35(12):84-93.
作者姓名:蒋浩泉  张儒清  郭嘉丰  范意兴  程学旗
作者单位:1. 中国科学院 计算技术研究所 网络数据科学与技术重点实验室,北京 100190;
2. 中国科学院大学,北京 100049
基金项目:北京智源人工智能研究院项目(BAAI2019ZD0306); 国家自然科学基金(62006218,61902381,61773362,61872338); 中国科学院青年创新促进项目(20144310,2016102,2021100); 国家重点研发计划(2016QY02D0405); 联想 中科院联合实验室青年科学家项目; 王宽诚教育基金会项目; 重庆市基础科学与前沿技术研究专项项目(重点)(cstc2017jcjyBX0059)
摘    要:图卷积网络近年来受到大量关注,同时自注意机制作为Transformer结构及众多预训练模型的核心之一也得到广泛运用。该文从原理上分析发现,自注意机制可视为图卷积网络的一种泛化形式,其以所有输入样本为节点,构建有向全连接图进行卷积,且节点间连边权重可学。在多个文本分类数据集上的对比实验一致显示,使用自注意机制的模型较使用图卷积网络的对照模型分类效果更佳,甚至超过了目前图卷积网络用于文本分类任务的最先进水平,并且随着数据规模的增大,两者分类效果的差距也随之扩大。这些证据表明,自注意力机制更具表达能力,在文本分类任务上能够相对图卷积网络带来分类效果的提升。

关 键 词:图卷积网络  自注意机制  文本分类  
收稿时间:2020-11-09

A Comparative Study of Graph Convolutional Networks and Self-Attention Mechanism on Text Classification
JIANG Haoquan,ZHANG Ruqing,GUO Jiafeng,FAN Yixing,CHENG Xueqi.A Comparative Study of Graph Convolutional Networks and Self-Attention Mechanism on Text Classification[J].Journal of Chinese Information Processing,2021,35(12):84-93.
Authors:JIANG Haoquan  ZHANG Ruqing  GUO Jiafeng  FAN Yixing  CHENG Xueqi
Affiliation:1. Key Laboratory of Network Data Science and Technology, Institute of Computing Technology, Chinese Academy of Sciences, Beijing 100190, China;
2. University of Chinese Academy of Sciences, Beijing 100049, China
Abstract:Graph Convolutional Networks has drawn much attention recently, and the self-attention mechanism has been widely applied as the core of the Transformer and many pre-trained models. We disclose that the self attention mechanism can be seen as a generalization of Graph Convolutional Networks, in that it takes all input samples as nodes and then constructs a directed fully connected graph with learnable edge weights for convolution. Experiments show that the self attention mechanism achieves better text classification accuracy than many state of the art Graph Convolutional Networks. Meanwhile, the performance gap of classification widens as the data size increases. These show that the self-attention mechanism is more expressive, and may surpass Graph Convolutional Networks with potential performance improvements on the task of text classification.
Keywords:graph convolutional networks  self-attention mechanism  text classification  
点击此处可从《中文信息学报》浏览原始摘要信息
点击此处可从《中文信息学报》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号