首页 | 本学科首页   官方微博 | 高级检索  
     

融合BERT和卷积门控的生成式文本摘要方法
引用本文:邓维斌,李云波,张一明,王国胤,朱坤.融合BERT和卷积门控的生成式文本摘要方法[J].控制与决策,2023,38(1):152-160.
作者姓名:邓维斌  李云波  张一明  王国胤  朱坤
作者单位:重庆邮电大学 计算智能重庆市重点实验室,重庆 400065;中国人民解放军78111部队,成都 610031
基金项目:国家研发计划项目(2018YFC0832100,2018YFC0832102);国家自然科学基金重点项目(61936001);国家自然科学基金项目(61876027);重庆市自然科学基金创新群体科学基金项目(cstc2019jcyj-cxttX0002).
摘    要:结合注意力机制的循环神经网络(RNN)模型是目前主流的生成式文本摘要方法,采用基于深度学习的序列到序列框架,但存在并行能力不足或效率低的缺陷,并且在生成摘要的过程中存在准确率低和重复率高的问题.为解决上述问题,提出一种融合BERT预训练模型和卷积门控单元的生成式摘要方法.该方法基于改进Transformer模型,在编码器阶段充分利用BERT预先训练的大规模语料,代替RNN提取文本的上下文表征,结合卷积门控单元对编码器输出进行信息筛选,筛选出源文本的关键内容;在解码器阶段,设计3种不同的Transformer,旨在探讨BERT预训练模型和卷积门控单元更为有效的融合方式,以此提升文本摘要生成性能.实验采用ROUGE值作为评价指标,在LCSTS中文数据集和CNN/Daily Mail英文数据集上与目前主流的生成式摘要方法进行对比的实验,结果表明所提出方法能够提高摘要的准确性和可读性.

关 键 词:生成式文本摘要  序列到序列  预训练模型  卷积门控单元  信息筛选  Transformer模型

An abstractive text summarization method combining BERT and convolutional gating unit
DENG Wei-bin,LI Yun-bo,ZHANG Yi-ming,WANG Guo-yin,ZHU Kun.An abstractive text summarization method combining BERT and convolutional gating unit[J].Control and Decision,2023,38(1):152-160.
Authors:DENG Wei-bin  LI Yun-bo  ZHANG Yi-ming  WANG Guo-yin  ZHU Kun
Affiliation:Chongqing Key Laboratory of Computational Intelligence,Chongqing University of Posts and Telecommunications,Chongqing 400065,China;78111 Troops of People''s Liberation Army of China,Chengdu 610031,China
Abstract:The recurrent neural network(RNN) model combined with the attention mechanism is the current mainstream abstractive text summarization method, which uses a sequence-to-sequence framework based on deep learning. However, the abstractive summarization model based on the RNN has insufficient parallel ability or performance defects of long-term dependence, and the problem of low accuracy and high repetition rate in the process of generating summary. In order to overcome these problems, an abstractive summarization model method combining the BERT pre-training model and the convolutional gating unit is proposed based on the improved Transformer model. In the encoder stage, it makes full use of the large-scale corpus pre-trained by the BERT to replace the RNN to extract the contextual representation of the text, and then combines the convolutional gating unit to filter the output of the encoder to filter out the source text. In the decoder stage, three different Transformers are designed, for exploring a more effective fusion method of the BERT pre-training model and convolutional gating unit to improve the performance of text summarization. The ROUGE value is used as the evaluation index in the experiments. The experimental results on the LCSTS Chinese dataset and CNN/Daily Mail dataset show that the proposed method improves the accuracy and readability of the abstract.
Keywords:
点击此处可从《控制与决策》浏览原始摘要信息
点击此处可从《控制与决策》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号