首页 | 本学科首页   官方微博 | 高级检索  
     

基于非文本模态强化和门控融合方法的多模态情感分析
引用本文:邵新慧,魏金龙.基于非文本模态强化和门控融合方法的多模态情感分析[J].计算机应用研究,2024,41(1).
作者姓名:邵新慧  魏金龙
作者单位:东北大学,东北大学
摘    要:针对各模态之间信息密度存在差距和融合过程中可能会丢失部分情感信息等问题,提出一种基于非文本模态强化和门控融合方法的多模态情感分析模型。该模型通过设计一个音频-视觉强化模块来实现音频和视觉模态的信息增强,从而减小与文本模态的信息差距。之后,通过跨模态注意力和门控融合方法,使得模型充分学习到多模态情感信息和原始情感信息,从而增强模型的表达能力。在对齐和非对齐的CMU-MOSEI数据集上的实验结果表明,所提模型是有效的,相比现有的一些模型取得了更好的性能。

关 键 词:多模态情感分析    多模态融合    模态强化    门控机制
收稿时间:2023/4/30 0:00:00
修稿时间:2023/12/15 0:00:00

Multimodal sentiment analysis based on non-text modality reinforcement and gating fusion method
Xin-Hui Shao and Jin-Long Wei.Multimodal sentiment analysis based on non-text modality reinforcement and gating fusion method[J].Application Research of Computers,2024,41(1).
Authors:Xin-Hui Shao and Jin-Long Wei
Affiliation:Northeastern University,
Abstract:To address the problems of information density gaps between modalities and the possibility of losing some sentiment information in the fusion process, this paper proposed a multimodal sentiment analysis model based on non-text modality reinforcement and gating fusion method. The model reduced the gap with text modality by designing an audio-visual reinforcement module to achieve information enhancement of audio and visual modalities. Then, the cross-modal attention and gating fusion method allowed the model to fully learn the multimodal sentiment information and the original sentiment information to enhance the representation of the model. Experimental results on the aligned and unaligned CMU-MOSEI datasets show that the proposed model is effective and achieves better performances than some existing models.
Keywords:multimodal sentiment analysis  multimodal fusion  modality reinforcement  gating mechanism
点击此处可从《计算机应用研究》浏览原始摘要信息
点击此处可从《计算机应用研究》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号