首页 | 本学科首页   官方微博 | 高级检索  
     

融合上下文信息和关键信息的文本摘要
引用本文:李志欣,彭智,唐素勤,马慧芳. 融合上下文信息和关键信息的文本摘要[J]. 中文信息学报, 2022, 36(1): 83-91
作者姓名:李志欣  彭智  唐素勤  马慧芳
作者单位:1.广西师范大学 广西多源信息挖掘与安全重点实验室,广西 桂林 541004;
2.西北师范大学 计算机科学与工程学院,甘肃 兰州 730070
基金项目:国家自然科学基金(61966004,61663004,61967002,61866004,61762078);广西自然科学基金(2019GXNS-FDA245018,2018GXNSFDA281009);广西八桂学者创新科研团队项目
摘    要:文本摘要的一个迫切需要解决的问题是如何准确地概括文本的核心内容.目前文本摘要的主要方法是使用编码器-解码器架构,在解码过程中利用软注意力获取所需的上下文语义信息.但是,由于编码器有时候会编码过多的信息,所以生成的摘要不一定会概括源文本的核心内容.为此,该文提出一种基于双注意指针网络的文本摘要模型.首先,该模型使用了双注...

关 键 词:文本摘要  神经网络  注意力机制  指针网络

Fusing Context Information and Key Information for Text Summarization
LI Zhixin,PENG Zhi,TANG Suqin,MA Huifang. Fusing Context Information and Key Information for Text Summarization[J]. Journal of Chinese Information Processing, 2022, 36(1): 83-91
Authors:LI Zhixin  PENG Zhi  TANG Suqin  MA Huifang
Affiliation:1.Guangxi Key Lab of Multi-source Information Mining and Security, Guangxi Normal University, Guilin, Guangxi 541004, China;
2.School of Computer Science and Engineering, Northwest Normal University, Lanzhou, Gansu 730070, China
Abstract:In text summarization, the mainstream method is to use encoder-decoder architecture to obtain the required context semantic information by using soft attention in the decoding process. Since the encoder sometimes encodes too much information, the generated summary does not always summarize the core content of the source text. To address this issue, this paper proposes a text summarization model based on a dual-attention pointer network. Firstly, in the dual-attention pointer network, the self-attention mechanism collects key information from the encoder, while the soft attention and the pointer network generate more coherent core content through context information. The fusion of both will generate accurate and coherent summaries. Secondly, the improved coverage mechanism is applied to address the repetition problem and improve the quality of the generated summaries. Simultaneously, scheduled sampling and reinforcement learning are combined to generate new training methods to optimize the model. Experiments on the CNN/Daily Mail dataset and the LCSTS dataset show that the proposed model performs as well as many state-of-the-art models.
Keywords:text summarization    neural network    attention mechanism    pointer network  
点击此处可从《中文信息学报》浏览原始摘要信息
点击此处可从《中文信息学报》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号