首页 | 本学科首页   官方微博 | 高级检索  
     

基于双层解码的多轮情感对话生成模型
引用本文:罗红,陆海俊,陈娟娟,慎煜杰,王丹.基于双层解码的多轮情感对话生成模型[J].计算机应用研究,2024,41(6).
作者姓名:罗红  陆海俊  陈娟娟  慎煜杰  王丹
作者单位:中移(杭州)信息技术有限公司,中移(杭州)信息技术有限公司,西安电子科技大学,西安电子科技大学杭州研究院,西安电子科技大学
摘    要:情感对话系统的成功取决于语言理解、情感感知和表达能力,同时面部表情和个性等也能提供帮助。然而,尽管这些信息对于多轮情感对话至关重要,但是现有系统既未能够充分利用多模态信息的优势,又忽略了上下文相关性的重要性。为了解决这个问题,提出了一种基于双层解码的多轮情感对话生成模型(MEDG-DD)。该模型利用异构的图神经网络编码器将历史对话、面部表情、情感流和说话者信息进行融合,以获得更加全面的对话上下文。然后,使用基于注意力机制的双层解码器,以生成与对话上下文相关的富含情感的言辞。实验结果表明,该模型能够有效地整合多模态信息,实现更为准确、自然且连贯的情感话语。与传统的ReCoSa模型相比,该模型在各项评估指标上均有显著的提升。

关 键 词:图神经网络编码器    注意力机制    双层解码    对话生成
收稿时间:2023/10/24 0:00:00
修稿时间:2024/5/13 0:00:00

Multi-turn emotion dialogue generation model based on dual-decoder
luo hong,lu hai jun,chen juan juan,shen yu jie and wang dan.Multi-turn emotion dialogue generation model based on dual-decoder[J].Application Research of Computers,2024,41(6).
Authors:luo hong  lu hai jun  chen juan juan  shen yu jie and wang dan
Affiliation:China Mobile (Hangzhou) Information Technology Co., Ltd.,,,,
Abstract:The success of emotional dialogue systems relies on the ability to comprehend, perceive, and express emotions, while facial expressions and personality can also help. However, despite the crucial importance of this multi-modal information in multi-turn emotional dialogues, existing systems still need to be improved to leverage multi-modal information''s advantages and overlook the significance of contextual relevance. To address this issue, this paper proposed a multi-turn emotional dialogue generation model based on a dua-decoding method(MEDG-DD). The model utilized a heterogeneous graph neural network encoder to integrate historical dialogue, facial expressions, emotion flow, and speaker information, obtaining a more comprehensive dialogue context. Subsequently, it employed a dual-decoding mechanism based on attention to generate emotionally rich expressions relevant to the dialogue context. Experimental results demonstrate that the proposed model effectively integrates multi-modal information, achieving more accurate, natural, and coherent emotional expressions. Compared to the traditional ReCoSa model, this model exhibits significant improvements across various evaluation metrics.
Keywords:graph neural network encoder  attention mechanism  dual-decoder  dialogue generation
点击此处可从《计算机应用研究》浏览原始摘要信息
点击此处可从《计算机应用研究》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号